Overview
Laminar automatically instruments the official Cohere Python SDK with a single line of code, allowing you to trace and monitor all your Cohere API calls without modifying your existing code. This provides complete visibility into your AI application’s performance, costs, and behavior.Getting Started (Python)
1. Install Laminar and Cohere
2. Set up your environment variables
Store your API keys in a.env
file:
3. Initialize Laminar and Cohere client
Just add a single line at the start of your application or file to instrument Cohere with Laminar.Use Cohere as usual
After initialization, make API calls to Cohere exactly as you normally would. Laminar will automatically capture traces for Chat, Embed, and Rerank endpoints.Chat (Command family)
Streaming Chat
RAG with Documents (Observed Pipeline)
Rerank
Semantic Search (Embeddings)
Monitoring Your Cohere Usage
After instrumenting your Cohere calls with Laminar, you’ll be able to:- View detailed traces of each Cohere API call, including request and response
- Track token usage and cost across different models
- Monitor latency and performance metrics
- Open LLM span in Playground for prompt engineering
- Debug issues with failed API calls or unexpected model outputs
Advanced Features
- Sessions - Learn how to add session structure to your traces
- Metadata - Discover how to add additional context to your LLM spans
- Trace structure - Explore creating custom spans and more advanced tracing
- Realtime Monitoring - See how to monitor your Cohere calls in real-time