Yuge Zhang
Yuge Zhang
## Summary - allow `training_rollout_batch`/`validation_rollout_batch` APIs to receive per-task resource lists - add `batch_size` to `AgentRunner` and propagate through `Trainer` for batched polling - test batch rollouts with resource lists...
## Summary - allow passing defaults for lightning_cli arguments - support programmatic defaults overriding required CLI args - test default overrides for required parameters ## Testing - `pytest` ------ https://chatgpt.com/codex/tasks/task_e_689590e72300832e95482ca7a6b7e4f8
We can't get token ids from LLM Proxy when stream is enabled. vLLM has token_ids returned in streaming. The problem is with LiteLLM and has three parts. 1. `.venv/lib/python3.12/site-packages/litellm/litellm_core_utils/streaming_handler.py` has...
1. `/agl/v1` to `/v1/agl`. 2. Reserved OTLP endpoints: https://opentelemetry.io/docs/specs/otlp/#otlphttp-request 3. Allow CORS 4. Use portpicker to fix flaky tests 5. Support resource filtering
(cherry picked from commit e0b55ab0571dc6a6b7244cd709a0ad7777b1689a)