Observability for TrueFoundry with Opik
TrueFoundry is an enterprise MLOps platform that provides a unified interface for deploying and managing ML models, including LLMs. It offers features like model deployment, monitoring, A/B testing, and cost optimization.
Gateway Overview
TrueFoundry provides enterprise-grade features for managing ML and LLM deployments, including:
- Unified OpenAI-Compatible Endpoint: Routes through TrueFoundry to any supported model (OpenAI, Anthropic, self-hosted, etc.)
- End-to-End Tracing: Full request/response logs with system messages, token breakdowns (prompt, completion, total), latency per call, and cost analytics per model and environment
- Production-Grade Controls: Rate limiting, quotas by user/team, budget alerts and spend caps, scoped API keys with RBAC
- Data Sovereignty: VPC and on-premises deployment options for compliance and data privacy
- Multi-Cloud Support: Deploy across AWS, Azure, GCP, and on-premise infrastructure
Account Setup
Comet provides a hosted version of the Opik platform. Simply create an account and grab your API Key.
You can also run the Opik platform locally, see the installation guide for more information.
Getting Started
Installation
First, ensure you have both opik and openai packages installed:
Configuring Opik
Configure the Opik Python SDK for your deployment type. See the Python SDK Configuration guide for detailed instructions on:
- CLI configuration:
opik configure - Code configuration:
opik.configure() - Self-hosted vs Cloud vs Enterprise setup
- Configuration files and environment variables
Configuring TrueFoundry
You’ll need your TrueFoundry API endpoint and credentials. You can get these from your TrueFoundry dashboard.
Set your configuration as environment variables:
Or set them programmatically:
Logging LLM Calls
Since TrueFoundry provides an OpenAI-compatible API for LLM deployments, we can use the Opik OpenAI SDK wrapper to automatically log TrueFoundry calls as generations in Opik.
Simple LLM Call
Advanced Usage
Using with the @track decorator
If you have multiple steps in your LLM pipeline, you can use the @track decorator to log the traces for each step. If TrueFoundry is called within one of these steps, the LLM call will be associated with that corresponding step:
The trace will show nested LLM calls with hierarchical spans.
Further Improvements
If you have suggestions for improving the TrueFoundry integration, please let us know by opening an issue on GitHub.