PyTorch/XLA integration with JetStream (https://github.com/google/JetStream) for LLM inference"
-
Updated
May 18, 2024 - Python
PyTorch/XLA integration with JetStream (https://github.com/google/JetStream) for LLM inference"
🤖 The free, Open Source OpenAI alternative. Self-hosted, community-driven and local-first. Drop-in replacement for OpenAI running on consumer-grade hardware. No GPU required. Runs gguf, transformers, diffusers and many more models architectures. It allows to generate Text, Audio, Video, Images. Also with voice cloning capabilities.
JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs welcome).
Finetune Llama 3, Mistral & Gemma LLMs 2-5x faster with 80% less memory
Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with any open-source language models, speech recognition models, and multimodal models, whether in the cloud, on-premises, or even on your laptop.
An R wrapper for the Gemma RESTful API
🏗️ Fine-tune, build, and deploy open-source LLMs easily!
Documentation for Google's Gen AI site - including the Gemini API and Gemma
This repository highlights the LLMs reasoning capabilities of ✨ Mistral / LLaMA-3 / Phi-3 / Gemma / Flan-T5 / GPT-4o ✨ in Targeted Sentiment Analysis in Russian / Translated to English mass-media 📊
PaliGemma Inference and Fine Tuning
Various LMs/LLMs below 3B parameters (for now) trained using SFT (Supervised Fine Tuning) for several downstream tasks
Explore the power of Gemma model with GemGPT, a project leveraging AI for innovative solutions. Join us in shaping the future of AI!
A snappy, keyboard-centric terminal user interface for interacting with large language models. Chat with ChatGPT, Claude, Llama 3, Phi 3, Mistral, Gemma and more.
RAG-nificent is a state-of-the-art framework leveraging Retrieval-Augmented Generation (RAG) to provide instant answers and references from a curated directory of PDFs containing information on any given topic. Supports Llama3 and OpenAI Models via the Groq API.
Add a description, image, and links to the gemma topic page so that developers can more easily learn about it.
To associate your repository with the gemma topic, visit your repo's landing page and select "manage topics."