Seekr provides AI-powered solutions for structured, explainable, and transparent AI interactions.This guide provides a quick overview for getting started with Seekr chat models. For detailed documentation of all
ChatSeekrFlow
features and configurations, head to the API reference.
Overview
ChatSeekrFlow
class wraps a chat model endpoint hosted on SeekrFlow, enabling seamless integration with LangChain applications.
Integration Details
Class | Package | Local | Serializable | Downloads | Version |
---|---|---|---|---|---|
ChatSeekrFlow | seekrai | ❌ | beta |
Model Features
Tool calling | Structured output | JSON mode | Image input | Audio input | Video input | Token-level streaming | Native async | Token usage | Logprobs |
---|---|---|---|---|---|---|---|---|---|
✅ | ✅ | ✅ | ❌ | ❌ | ❌ | ✅ | ❌ | ✅ | ❌ |
Supported Methods
ChatSeekrFlow
supports all methods of ChatModel
, except async APIs.
Endpoint Requirements
The serving endpointChatSeekrFlow
wraps must have OpenAI-compatible chat input/output format. It can be used for:
- Fine-tuned Seekr models
- Custom SeekrFlow models
- RAG-enabled models using Seekr’s retrieval system
AsyncChatSeekrFlow
(coming soon).
Getting Started with ChatSeekrFlow in LangChain
This notebook covers how to use SeekrFlow as a chat model in LangChain.Setup
Ensure you have the necessary dependencies installed:API Key Setup
You’ll need to set your API key as an environment variable to authenticate requests. Run the below cell. Or manually assign it before running queries:Instantiation
Invocation
Chaining
Error Handling & Debugging
API reference
ChatSeekrFlow
class:langchain_seekrflow.ChatSeekrFlow
- PyPI package:
langchain-seekrflow