gemma topic
LocalAI
:robot: The free, Open Source OpenAI alternative. Self-hosted, community-driven and local-first. Drop-in replacement for OpenAI running on consumer-grade hardware. No GPU required. Runs gguf, transfor...
inference
Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with any...
generative-ai-docs
Documentation for Google's Gen AI site - including the Gemini API and Gemma
LlamaIndex-RAG-WSL-CUDA
Examples of RAG using Llamaindex with local LLMs - Gemma, Mixtral 8x7B, Llama 2, Mistral 7B, Orca 2, Phi-2, Neural 7B
aikit
🏗️ Fine-tune, build, and deploy open-source LLMs easily!
Upsonic
Self-Driven Autonomous Python Libraries
nextjs-ollama-llm-ui
Fully-featured, beautiful web interface for Ollama LLMs - built with NextJS. Deploy with a single click.
JetStream
JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs welcome).
InfiniTransformer
Unofficial PyTorch/🤗Transformers(Gemma/Llama3) implementation of Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention
project-zephyrine
Introducing Project Zephyrine: Elevating Your Interaction Plug and Play, and Employing GPU Acceleration within a Modernized Automata Local Graphical User Interface.