inference icon indicating copy to clipboard operation
inference copied to clipboard

Reduce the dataset size for LLAMA2

Open arjunsuresh opened this issue 11 months ago • 2 comments

I believe there is no need to run LLAMA2-70B model for 24576 samples. GPT-J-6B is run for 13368 samples, Stable diffusion is run for 5000 samples, 3d-unet is run for 43 samples and all other MLPerf inference models are way faster. In fact, LLAMA2-70B is at least an order of magnitude slower than any other MLPerf inference model. For coming rounds, we should reduce the LLAMA2-70B dataset size to some 4-digit number.

arjunsuresh avatar Feb 26 '24 23:02 arjunsuresh

Another idea would be to create a representative subset for performance runs, i.e. with sample distribution similar to the original dataset, but only 1/10th the size.

psyhtest avatar Feb 27 '24 15:02 psyhtest

@psyhtest Thats a good option. But in LLAMA2 case the dataset is already a subset selection. So, it might be easier to reduce the dataset size itself though this means a new accuracy threshold. If we use a different dataset for performance runs, TEST01 might be complicated when it gets introduced for LLAMA2.

arjunsuresh avatar Feb 27 '24 15:02 arjunsuresh