Tensor parallelism is all you need. Run LLMs on weak devices or make powerful devices even more powerful by distributing the workload and dividing the RAM usage.
-
Updated
Jun 1, 2024 - C++
Tensor parallelism is all you need. Run LLMs on weak devices or make powerful devices even more powerful by distributing the workload and dividing the RAM usage.
🤖 Cybersecurity Automation & Investigation Assistant
LLM Orchestrator powered by langchain and Bot Framework V4 & several features including Whatsapp.
Burn is a new comprehensive dynamic Deep Learning Framework built using Rust with extreme flexibility, compute efficiency and portability as its primary goals.
🤖 Bring the magic of ChatGPT to Google Search (powered by GPT-4!)
A cli chatroom for AutoGen agents
Retrieval-Augmented (RAG) based pretrained GPT model that predicts and analyses the November 2024 US General Elections using news sources (CNN, FoxNews, Politico, and NPR) as context
Generative AI Scripting for VSCode
Elevate user interactions with ChatFAQ: your open-source chatbot solution, offering the full spectrum of ChatGPT capabilities. AI + LLM + CMS
AI powered tool to help software teams with Quality Assurance
🤖 Neovim code suggestion and completion (just like GitHub Copilot, but locally using Ollama)
SiLLM simplifies the process of training and running Large Language Models (LLMs) on Apple Silicon by leveraging the MLX framework.
DrakeLLM is developed to help students to solve the issue of making notes from videos, books and others. Utilising RAG, Drake helps in making quick notes along with a Q&A bot. Books, YouTube tutorials or Videos, Drake supports all your means.
Add a description, image, and links to the llm topic page so that developers can more easily learn about it.
To associate your repository with the llm topic, visit your repo's landing page and select "manage topics."