Theodore Vasiloudis
Theodore Vasiloudis
Hello Julaiti, Great talking to you at NeurIPS! I was wondering if you plan to make some of the experiment automation code available, i.e. for setting up a cluster and...
Got this when running with a custom DGL installation. Using a training command e.g. ```bash python3 -m graphstorm.run.gs_node_classification \ --extra-envs LD_LIBRARY_PATH="/opt/gs-venv/lib/python3.9/site-packages/dgl/:$LD_LIBRARY_PATH" \ --num-trainers 1 \ --num-servers 1 \ --num-samplers 0...
*Issue #, if available:* *Description of changes:* * We change the way we configure the Spark env, to only create our own config for SageMaker, as EMR/EMRS will have pre-configured...
Right now for the distributed pipeline we would first create raw id mappings using GSProcessing, which the second step of DistPart does not touch. When a user runs a training...
With the number of arguments that we parse, it's likely that an arg can be misspelled. e.g. I ran into the case where I was trying to pass `--save-embeds-path` instead...
…nversion ## Description ## Checklist Please feel free to remove inapplicable items for your PR. - [ ] The PR title starts with [$CATEGORY] (such as [NN], [Model], [Doc], [Feature]])...
## 🐛 Bug I've observed an error when trying to use GraphBolt with `--num-trainers >1`. In this case I'm using DistGB through GraphStorm, so not sure if it's GSF or...
*Issue #, if available:* *Description of changes:* * Adds support for range partitioning for local and SageMaker execution environment. * Depends on #880 do not merge before. By submitting this...
For example, if we try to run https://github.com/awslabs/graphstorm/tree/main/training_scripts/gsgnn_mt on the GraphStorm image, we'd run into the error ``` python3 tests/end2end-tests/data_gen/process_movielens.py Traceback (most recent call last): File "/root/graphstorm/tests/end2end-tests/data_gen/process_movielens.py", line 29, in...
https://github.com/awslabs/graphstorm/pull/902 will simplify our instructions for how to use GSProcessing on EMR-EC2. Advanced users might still want to launch and configure their own clusters though