mediapipe
mediapipe copied to clipboard
Fixed issue with `examples/llm_inference/js` failing to load larger models by implementing buffered loading system allowing for the loading of larger models like Gemma 3n E2B & E4B
Description
Please include a summary of the change and which issue is fixed. Please also include relevant motivation and context.
Fixes # (issue)
Checklist
Please ensure the following items are complete before submitting a pull request:
- [x] My code follows the code style of the project.
- [x] I have updated the documentation (if applicable).
- [x] I have added tests to cover my changes.
Type of Change
Please check the relevant option below:
- [x] Bug fix (non-breaking change which fixes an issue)
- [ ] Documentation update (non-breaking change which updates documentation)
- [x] New feature (non-breaking change which adds functionality)
- [ ] Breaking change (fix or feature that would cause existing functionality to not work as expected)
Additional Notes
Now capable of loading larger models (i.e. Google's E2B & E4B) without just completely crashing.
@tyrmullen Apologies beforehand for the ping, but it seems like a lot of people struggled with the same issue, not being able to get past it, and you're the most active contributor on the repository (which I appreciate, I love Google's work on AI efficiency, and especially edge AI :D).