OpenLLM
🦾 OpenLLM lets developers run any open-source LLMs as OpenAI-compatible API endpoints with a single command.
- 🔬 Build for fast and production usages
 - 🚂 Support llama3, qwen2, gemma, etc, and many quantized versions full list
 - ⛓️ OpenAI-compatible API
 - 💬 Built-in ChatGPT like UI
 - 🔥 Accelerated LLM decoding with state-of-the-art inference backends
 - 🌥️ Ready for enterprise-grade cloud deployment (Kubernetes, Docker and BentoCloud)
 
Installation
Install openllm through PyPI
%pip install --upgrade --quiet  openllm
Launch OpenLLM server locally
To start an LLM server, use openllm hello command:
openllm hello
Wrapper
from langchain_community.llms import OpenLLM
server_url = "http://localhost:3000"  # Replace with remote host if you are running on a remote server
llm = OpenLLM(base_url=server_url, api_key="na")
llm("To build a LLM from scratch, the following are the steps:")
Related
- LLM conceptual guide
 - LLM how-to guides