Mark Ericksen
Mark Ericksen
@nileshtrivedi I setup a GoogleAI account and got an API key. There are multiple issues with the ChatGoogleAI implementation. I've fixed a couple locally (not pushed yet), but there's an...
@nileshtrivedi This is _hopefully_ fixed now! :crossed_fingers: Just merged PR #152 to `main`. If you test using main it should be working now. Please let me know!
@nileshtrivedi I'm pretty annoyed by the GoogleAI honestly. It's such an oddball compared to others. The API docs are sparse and difficult to use too. Ugh. One odd thing that...
Hi @Calamari, no, there is currently is not a way to surface the rate information or the current token usage. The token limit one is a general think that applies...
Here's what I mean by the model limits varying. Heads up: I'm mostly talking out loud here as I think through it too. Here's the details on ChatGPT's 3.5 models:...
I looked into the JS version and they don't have anything documented at least. The Python version docs are much more complete here and I like their approach. https://python.langchain.com/docs/modules/model_io/models/llms/token_usage_tracking The...
But this still doesn't tell me what I want to know, which is, "given the model I'm using, how many tokens do I have left?" That becomes left up to...
There are two different types of limits being talked about here. 1. The max tokens allowed for the conversation (what the callback talks about) 2. The ratelimit tokens (what your...
@Calamari Just a quick follow-up. I'm not sure how to best support this feature. I'm also thinking of Bumblebee based LLMs. I've been in talks with that team about getting...
@jadengis, are you seeing this too? Are you able to weigh in?