Get started with Observyze.
Learn how to instrument your AI agents, enforce security policies (PII, Toxicity), and monitor cost analytics in real-time.
Python Quickstart
Get up and running with our Python SDK in under 5 minutes.
Node.js Quickstart
Instrument your TypeScript/JavaScript applications with ease.
AI Proxy (No-Code)
Connect by just changing your OpenAI base URL. Zero SDK required.
Webhook Ingest
Send traces from Zapier, Make, or any HTTP client with a simple POST.
Universal Agent Instrumentation
Our universal SDKs allow you to monitor any LLM provider or specialized agent framework. We support OpenAI, Anthropic, LangChain, and any custom agents.
import Observyze as nw
from openai import OpenAI
# 1. Initialize with your API Key
nw.init(api_key="nw_prod_...")
# 2. Wrap your client or agent
client = nw.wrap(OpenAI())
# 3. Use as normal - traces and evaluations are automatic
response = client.chat.completions.create(
model="gpt-4o",
messages=[{"role": "user", "content": "Hello!"}]
)Async Handling
Non-blocking IO ensure zero latency impact.
Streaming
Full support for chunked LLM responses.
Auto-Masking
Securely redact PII from spans automatically.
Low Payload
Compressed data transfer for high performance.
Need more help?
Our engineering team is available for technical support.