Konko AI provides a fully managed API to help application developers
- Select the right open source or proprietary LLMs for their application
- Build applications faster with integrations to leading application frameworks and fully managed APIs
- Fine tune smaller open-source LLMs to achieve industry-leading performance at a fraction of the cost
- Deploy production-scale APIs that meet security, privacy, throughput, and latency SLAs without infrastructure set-up or administration using Konko AI’s SOC 2 compliant, multi-cloud infrastructure
Installation and Setup
- Sign in to our web app to create an API key to access models via our endpoints for chat completions and completions.
- Enable a Python3.8+ environment
- Install the SDK
- Set API Keys as environment variables(
KONKO_API_KEY
,OPENAI_API_KEY
)
LLM
Explore Available Models: Start by browsing through the available models on Konko. Each model caters to different use cases and capabilities. Another way to find the list of models running on the Konko instance is through this endpoint. See a usage example.Examples of Endpoint Usage
-
Completion with mistralai/Mistral-7B-v0.1:
Chat Models
See a usage example.-
ChatCompletion with Mistral-7B: