lm-evaluation-harness
lm-evaluation-harness copied to clipboard
avoid timeout errors with high concurrency in api_model
Longer timeout will avoid errors when running tests with high concurrency
Thanks v. much for the PR. left a comment. Also cc: @artemorloff, as he had also made the change in #2249.
sorry @dtrawins . Forgot to commit the review message.
Are there any blockers here? I also had to apply this fix, and it seems to make sense to merge sth like that, no?
When doing long evals on big models such as llama 70b it fails even with low concurrency (e.g. 4). I'm using 8xA100-80GB and it would time out.
I had to fix it this way:
echo "import aiohttp.client
aiohttp.client.DEFAULT_TIMEOUT = aiohttp.ClientTimeout(total=14400)" >> ~/lm-evaluation-harness/lm_eval/__init__.py