The production agentic context stack. Data streaming + knowledge graphs + vector search + LLM orchestration. All in a single deployment for self hosting, BYOC, or cloud.
-
Updated
Nov 1, 2025 - Python
The production agentic context stack. Data streaming + knowledge graphs + vector search + LLM orchestration. All in a single deployment for self hosting, BYOC, or cloud.
《大模型项目实战:多领域智能应用开发》配套资源
A curated collection of open-source Large Language Model (LLM) projects that are production-ready and can be used for solving real-world problems. This repository focuses on high-performance, scalable LLM solutions across various industries and applications.
🧠 A comprehensive toolkit for benchmarking, optimizing, and deploying local Large Language Models. Includes performance testing tools, optimized configurations for CPU/GPU/hybrid setups, and detailed guides to maximize LLM performance on your hardware.
API to efficiently deploy Language Model (LLM) applications using Flask API
AWS EKS + IRSA, Volumes, ISTIO & KServe+ NextJS App + Fastapi Serve + kubernetes + Helm charts + Multimodel or LLM-Deployment The School of AI EMLO-V4 course assignment https://theschoolof.ai/#programs
LLM App for summarization of Terms and Conditions agreements available on the internet.
Python codes generation from latex expressions. Using synthetic dataset and CodeT5-base model.
Add a description, image, and links to the llm-deployment topic page so that developers can more easily learn about it.
To associate your repository with the llm-deployment topic, visit your repo's landing page and select "manage topics."