gpt-tokenizer
gpt-tokenizer copied to clipboard
Support function calls?
As you may know, function calls are now supported by OpenAI, and the function call tokens will be taken into account. With some googling, I found the calculation here.
It will be great if the library could also calculate the required tokens when using with function calls. 🙏🏼
One related issue is the type signature of ChatMessage. Depending on whether there is functional calling, the ChatMessage may have either the content field or the function_call field, but not both. The current typing in gpt-tokenizer/src/GptEncoding.ts will need an update.
I saw that https://github.com/hmarr/openai-chat-tokens already does it - would it be fine to import that package here to provide that functionality?
PRs welcome! I would prefer not to pull in an external dependency, and it looks like the code isn't too complex.