generative-ai-python icon indicating copy to clipboard operation
generative-ai-python copied to clipboard

Gemini 1.5 Flash Supervised Fine Tuning Updates

Open bgrove-s7 opened this issue 1 year ago • 2 comments

Description of the feature request:

Increase character limitation in tuning jobs to take advantage of the Flash 1.5 1,000,000 token window.

What problem are you trying to solve with this feature?

Gemini 1.5 Flash has a very large token window, which potentially makes it ideal for extracting needles from haystacks of text. We would like to fine tune Gemini 1.5 to perform this task for us. Tuning is appropriate because the content we are typically examining is hundreds of thousands of tokens, leaving no room for multishot prompting techniques.

Any other information you'd like to share?

No response

bgrove-s7 avatar Aug 26 '24 19:08 bgrove-s7

What is the current limit?

MarkDaoust avatar Aug 26 '24 20:08 MarkDaoust

It appears to be 40,000 characters, per the following error:

CreateTunedModelRequest.tuned_model.tuning_task.training_data.examples.examples[9].text_input: text_input is too long. The maximum character count accepted is 40000.

bgrove-s7 avatar Aug 26 '24 20:08 bgrove-s7

Thanks for raising this issue. Have there been updates since the last comment, and is this request still active?

Annhiluc avatar Mar 24 '25 22:03 Annhiluc

We are no longer trying to train with Gemini 1.5 Flash, but thank you for following up.

bgrove-s7 avatar Mar 25 '25 12:03 bgrove-s7

Marking this issue as stale since it has been open for 14 days with no activity. This issue will be closed if no further activity occurs.

github-actions[bot] avatar Apr 09 '25 02:04 github-actions[bot]