optillm
Optimizing inference proxy for LLMs
Maintained by codelion
Project Information
- GitHub Stars
- 2,162
- Language
- Python
- Last Updated
- April 19, 2025 at 07:53 AM
Topics
agent
agentic-ai
agentic-workflow
agents
api-gateway
genai
large-language-models
llm
llm-inference
llmapi
mixture-of-experts
moa
openai
openai-api
optimization
proxy-server
agentic-framework
chain-of-thought
monte-carlo-tree-search
prompt-engineering