gemma
Here are 153 public repositories matching this topic...
🤖 The free, Open Source alternative to OpenAI, Claude and others. Self-hosted and local-first. Drop-in replacement for OpenAI, running on consumer-grade hardware. No GPU required. Runs gguf, transformers, diffusers and many more models architectures. Features: Generate Text, Audio, Video, Images, Voice Cloning, Distributed, P2P inference
-
Updated
Nov 17, 2024 - Go
Firefly: 大模型训练工具,支持训练Qwen2.5、Qwen2、Yi1.5、Phi-3、Llama3、Gemma、MiniCPM、Yi、Deepseek、Orion、Xverse、Mixtral-8x7B、Zephyr、Mistral、Baichuan2、Llma2、Llama、Qwen、Baichuan、ChatGLM2、InternLM、Ziya2、Vicuna、Bloom等大模型
-
Updated
Oct 24, 2024 - Python
Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with any open-source language models, speech recognition models, and multimodal models, whether in the cloud, on-premises, or even on your laptop.
-
Updated
Nov 15, 2024 - Python
A snappy, keyboard-centric terminal user interface for interacting with large language models. Chat with ChatGPT, Claude, Llama 3, Phi 3, Mistral, Gemma and more.
-
Updated
Oct 10, 2024 - Python
Documentation for Google's Gen AI site - including the Gemini API and Gemma
-
Updated
Oct 29, 2024 - Jupyter Notebook
Fully-featured, beautiful web interface for Ollama LLMs - built with NextJS. Deploy with a single click.
-
Updated
Oct 31, 2024 - TypeScript
A collection of guides and examples for the Gemma open models from Google.
-
Updated
Nov 8, 2024 - Jupyter Notebook
Official repository for "Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing". Your efficient and high-quality synthetic data generation pipeline!
-
Updated
Nov 5, 2024 - Python
Chat with AI large language models running natively in your browser. Enjoy private, server-free, seamless AI conversations.
-
Updated
Nov 13, 2024 - TypeScript
🏗️ Fine-tune, build, and deploy open-source LLMs easily!
-
Updated
Nov 11, 2024 - Go
Unofficial PyTorch/🤗Transformers(Gemma/Llama3) implementation of Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention
-
Updated
Apr 23, 2024 - Python
InternEvo is an open-sourced lightweight training framework aims to support model pre-training without the need for extensive dependencies.
-
Updated
Nov 7, 2024 - Python
Inferflow is an efficient and highly configurable inference engine for large language models (LLMs).
-
Updated
Mar 15, 2024 - C++
JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs welcome).
-
Updated
Nov 8, 2024 - Python
On-device LLM Inference Powered by X-Bit Quantization
-
Updated
Nov 14, 2024 - Python
Improve this page
Add a description, image, and links to the gemma topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the gemma topic, visit your repo's landing page and select "manage topics."