inference icon indicating copy to clipboard operation
inference copied to clipboard

Reference implementations of MLPerf™ inference benchmarks

Results 331 inference issues
Sort by recently updated
recently updated
newest added

Bumps [transformers](https://github.com/huggingface/transformers) from 4.33.2 to 4.36.0. Release notes Sourced from transformers's releases. v4.36: Mixtral, Llava/BakLlava, SeamlessM4T v2, AMD ROCm, F.sdpa wide-spread support New model additions Mixtral Mixtral is the new...

dependencies

I run and have an error: ``` time python3 main.py --dataset "coco-1024" --dataset-path coco2014_full --profile stable-diffusion-xl-pytorch --model-path model/stable_diffusion _fp16/ --dtype fp16 --device cuda --scenario SingleStream --model-name stable-diffusion-xl --qps 0.022 --output...

Stable Diffusion Reference support `qps` as input parameter in a command but don't support `target_latency` for SingleStream. @pgmpablo157321 could you please add this support ?

**Current progress:** * CM coverage to automate and reproduce MLPerf inference: [GitHub](https://github.com/mlcommons/ck/issues/1052) * All reference implementations are supported by CM including GPT-J and Stable Diffusion (though didn’t run LLAMA). *...

Does the tensorflow version of DLRM support cpus?

Please see below for the detailed output. The run is done on Nvidia RTX 4090 GPU. ``` CMD: /home/arjun/cm/bin/python3 main.py --scenario SingleStream --profile stable-diffusion-xl-pytorch --dataset coco-1024 --dataset-path /home/arjun/CM/repos/local/cache/03fbdcf95b3d4104/install --dtype fp16...

Hi, I was running dlrm pytorch with CPU docker by using fake data. seeing below error. /root/mlcommons/recommendation/dlrm/pytorch/python/dlrm_data_pytorch.py:328: UserWarning: Creating a tensor from a list of numpy.ndarrays is extremely slow. Please...

multicard support and data distribution across ranks missing in the reference for SDXL (https://github.com/mlcommons/inference/tree/master/text_to_image). @pgmpablo157321 can you add this support ?

[root@poc4 bert]# make setup make[1]: Entering directory '/home/sunyu/benchmark/MLPerf/inference/language/bert' You need to run this command from the toplevel of the working tree. make[1]: *** [init_submodule] Error 1 make[1]: Leaving directory '/home/sunyu/benchmark/MLPerf/inference/language/bert'...

Hello, I get the inference result v3.1 and analysis the performance, throughput (token per second). In the large language model task, the test results are measured by Queries/s and Samples/s....