Run LLMs on weak devices or make powerful devices even more powerful by distributing the workload and dividing the RAM usage.
-
Updated
May 19, 2024 - C++
Run LLMs on weak devices or make powerful devices even more powerful by distributing the workload and dividing the RAM usage.
Private chat with local GPT with document, images, video, etc. 100% private, Apache 2.0. Supports oLLaMa, Mixtral, llama.cpp, and more. Demo: https://gpt.h2o.ai/ https://codellama.h2o.ai/
Design, conduct and analyze results of AI-powered surveys and experiments. Simulate social science and market research with large numbers of AI agents and LLMs.
🤖 Collect practical AI repos, tools, websites, papers and tutorials on AI. 实用的AI百宝箱 💎
memory-efficient fine-tuning; support 24G GPU memory fine-tuning 7B
Sample to envision intelligent apps with Microsoft's Copilot stack for AI-infused product experiences.
Empower Your Productivity with Local AI Assistants
Finetune Llama 3, Mistral & Gemma LLMs 2-5x faster with 80% less memory
🤯 Lobe Chat - an open-source, modern-design LLMs/AI chat framework. Supports Multi AI Providers( OpenAI / Claude 3 / Gemini / Ollama / Bedrock / Azure / Mistral / Perplexity ), Multi-Modals (Vision/TTS) and plugin system. One-click FREE deployment of your private ChatGPT chat application.
Foundation model benchmarking tool. Run any model on Amazon SageMaker and benchmark for performance across instance type and serving stack options.
Dify is an open-source LLM app development platform. Dify's intuitive interface combines AI workflow, RAG pipeline, agent capabilities, model management, observability features and more, letting you quickly go from prototype to production.
Local character AI chatbot with chroma vector store memory and some scripts to process documents for Chroma
End-to-end platform for building voice first multimodal agents
An Easy-to-use Knowledge Editing Framework for LLMs.
WebUI for Fine-Tuning and Self-hosting of Open-Source Large Language Models for Coding
This project offers a versatile platform for hand-related tasks, including dataset generation and custom hand gesture detection using Google's MediaPipe library and accelerated real-time sign language translation with LLMs on edge devices.
Add a description, image, and links to the llama2 topic page so that developers can more easily learn about it.
To associate your repository with the llama2 topic, visit your repo's landing page and select "manage topics."