HugeCTR icon indicating copy to clipboard operation
HugeCTR copied to clipboard

HugeCTR is a high efficiency GPU framework designed for Click-Through-Rate (CTR) estimating training

Results 88 HugeCTR issues
Sort by recently updated
recently updated
newest added

Hi HugeCTR team, I'm testing the HugeCTR performance with our data of real scenario, but the throughtput is unecpected low. I double checked our data and scrpits, but still no...

question
P1
TBD

The current [Movie-Lens example](https://github.com/NVIDIA-Merlin/HugeCTR/blob/master/notebooks/movie-lens-example.ipynb) uses KNN to make a inference demo but produces suspicious results which have a lot of queries returning the same recommendation. According to @vinhngx, there is...

P1

when i try Embedding Training Cache Example [https://nvidia-merlin.github.io/HugeCTR/master/notebooks/embedding_training_cache_example.html](url) if i use multi gpu by set vvgpu to [[1,2]], i wil get error at second pass : HugeCTR Version:3.6 =====================================================Model Fit=====================================================...

question

when i try to run a sample of wdl https://github.com/NVIDIA-Merlin/HugeCTR/tree/master/samples/wdl i got a error at step :Preprocess the Dataset Through NVTabular bash preprocess.sh 0 criteo_data nvt 1 0 1 #...

question

**Describe the bug** when i try notebook https://nvidia-merlin.github.io/HugeCTR/master/notebooks/hugectr_wdl_prediction.html i got error when run preprocess.py TypeError: Unsupported type: Cannot convert object of type to Node. **To Reproduce** Steps to reproduce the...

**Describe the bug** A clear and concise description of what the bug is. **To Reproduce** Steps to reproduce the behavior: 1. Use AMI https://aws.amazon.com/releasenotes/deep-learning-ami-gpu-tensorflow-2-9-ubuntu-20-04/ to spin a ec2 cluster instance...

**Describe the bug** When running the hps_demo.ipynb notebook, with the latest docker container 22.05, the third code cell where we train the model will crash with a core dump. I...

As the title, I run the script: python ../tools/embedding_plugin/performance_profile/preprocess.py --src_csv_path=train.txt --dst_csv_path=train.out.txt --normalize_dense=0 --feature_cross=0 in the official docker,but it core dumped with A process in the process pool was terminated abruptly...

bug
critical

I know the performance of HugeCTR + NVTabular is awesome, the Criteo 1TB Click Logs Dataset can be processed and a full model can be trained in only six minutes....

question