optillm logo

optillm

Optimizing inference proxy for LLMs

Maintained by codelion

Project Information

GitHub Stars
2,162
Language
Python
Last Updated
April 19, 2025 at 07:53 AM

Topics

agent
agentic-ai
agentic-workflow
agents
api-gateway
genai
large-language-models
llm
llm-inference
llmapi
mixture-of-experts
moa
openai
openai-api
optimization
proxy-server
agentic-framework
chain-of-thought
monte-carlo-tree-search
prompt-engineering

Explore More

Discover similar projects or browse the full catalog.