The easiest way to serve AI/ML models in production - Build Model Inference Service, LLM APIs, Multi-model Inference Graph/Pipelines, LLM/RAG apps, and more!
-
Updated
Jun 11, 2024 - Python
The easiest way to serve AI/ML models in production - Build Model Inference Service, LLM APIs, Multi-model Inference Graph/Pipelines, LLM/RAG apps, and more!
Repository for NLP usecases. MLOPS applied to NLP
Production Grade Nifi & Nifi Registry. Deploy for VM (Virtual Machine) with Terraform + Ansible, Helm & Helmfile for Kubernetes (EKS)
RAG (Retrieval Augmented Generation) Framework for building modular, open source applications for production by TrueFoundry
MLOps End-to-End Example using Amazon SageMaker Pipeline, AWS CodePipeline and AWS CDK
A high-throughput and memory-efficient inference and serving engine for LLMs
The official Python library for Openlayer, the Continuous Model Improvement Platform for AI. 📈
Workflow Engine for Kubernetes
Apache Airflow - A platform to programmatically author, schedule, and monitor workflows
Weaviate is an open-source vector database that stores both objects and vectors, allowing for the combination of vector search with structured filtering with the fault tolerance and scalability of a cloud-native database.
This is an AI-Gemini Chatbot LLM And Large Image Model Application. You can use this project run into local and ask you images like your talking with in realtime
AI Observability & Evaluation
Checklists and LLM prompts for efficient and effective test creation in data analysis
JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs welcome).
An orchestration platform for the development, production, and observation of data assets.
Homeworks and capstone project of MLOps Zoomcamp 2024 from DataTalksClub
Database for AI. Store Vectors, Images, Texts, Videos, etc. Use with LLMs/LangChain. Store, query, version, & visualize any AI data. Stream data in real-time to PyTorch/TensorFlow. https://activeloop.ai
Open source AI platform for rapid development of advanced AI and AGI pipelines.
Standardized Serverless ML Inference Platform on Kubernetes
Add a description, image, and links to the mlops topic page so that developers can more easily learn about it.
To associate your repository with the mlops topic, visit your repo's landing page and select "manage topics."