Yuge Zhang

Results 63 issues of Yuge Zhang

## Summary - allow `training_rollout_batch`/`validation_rollout_batch` APIs to receive per-task resource lists - add `batch_size` to `AgentRunner` and propagate through `Trainer` for batched polling - test batch rollouts with resource lists...

codex

## Summary - allow passing defaults for lightning_cli arguments - support programmatic defaults overriding required CLI args - test default overrides for required parameters ## Testing - `pytest` ------ https://chatgpt.com/codex/tasks/task_e_689590e72300832e95482ca7a6b7e4f8

codex

We can't get token ids from LLM Proxy when stream is enabled. vLLM has token_ids returned in streaming. The problem is with LiteLLM and has three parts. 1. `.venv/lib/python3.12/site-packages/litellm/litellm_core_utils/streaming_handler.py` has...

blocked by upstream
proxy

1. `/agl/v1` to `/v1/agl`. 2. Reserved OTLP endpoints: https://opentelemetry.io/docs/specs/otlp/#otlphttp-request 3. Allow CORS 4. Use portpicker to fix flaky tests 5. Support resource filtering

ci-all

(cherry picked from commit e0b55ab0571dc6a6b7244cd709a0ad7777b1689a)

backport