Getting Started

Changelog


Opik Dashboard:

  • Improved the UX when navigating between the project list page and the traces page

Python SDK:

  • Make the logging of spans and traces optional when using Opik LLM metrics

JS / TS SDK:

  • Added logs and better error handling

Opik Dashboard:

  • Added support for local models in the Opik playground

Python SDK:

  • Improved the @track decorator to better support nested generators.
  • Added a new Opik.copy_traces(project_name, destination_project_name) method to copy traces from one project to another.
  • Added support for searching for traces that have feedback scores with spaces in their name.
  • Improved the LangChain and LangGraph integrations

JS / TS SDK:

  • Released the Vercel AI integration
  • Added support for logging feedback scores

Opik Dashboard:

  • You can now view feedback scores for your projects in the Opik home page
  • Added line highlights in the quickstart page
  • Allow users to download experiments as CSV and JSON files for further analysis

Python SDK:

  • Update the evaluate_* methods so feedback scores are logged after they computed rather than at the end of an experiment as previously
  • Released a new usefulness metric
  • Do not display warning messages about missing API key when Opik logging is disabled
  • Add method to list datasets in a workspace
  • Add method to list experiments linked to a dataset

JS / TS SDK:

  • Official release of the first version of the SDK - Learn more here
  • Support logging traces using the low-level Opik client and an experimental decorator.

Opik Dashboard:

  • Performance improvements for workspaces with 100th of millions of traces
  • Added support for cost tracking when using Gemini models
  • Allow users to diff prompt

SDK:

  • Fixed the evaluate and evaluate_* functions to better support event loops, particularly useful when using Ragas metrics
  • Added support for Bedrock invoke_agent API

Opik Dashboard:

  • Added logs for online evaluation rules so that you can more easily ensure your online evaluation metrics are working as expected
  • Added auto-complete support in the variable mapping section of the online evaluation rules modal
  • Added support for Anthropic models in the playground
  • Experiments are now created when using datasets in the playground
  • Improved the Opik home page
  • Updated the code snippets in the quickstart to make them easier to understand

SDK:

  • Improved support for litellm completion kwargs
  • LiteLLM required version is now relaxed to avoid conflicts with other Python packages

Opik Dashboard:

  • Datasets are now supported in the playground allowing you to quickly evaluate prompts on multiple samples
  • Updated the models supported in the playground
  • Updated the quickstart guides to include all the supported integrations
  • Fix issue that means traces with text inputs can’t be added to datasets
  • Add the ability to edit dataset descriptions in the UI
  • Released online evaluation rules - You can now define LLM as a Judge metrics that will automatically score all, or a subset, of your production traces.

Online evaluation

SDK:

  • New integration with CrewAI
  • Released a new evaluate_prompt method that simplifies the evaluation of simple prompts templates
  • Added Sentry to the Python SDK so we can more easily

Opik Dashboard:

  • Fixed an issue with the trace viewer in Safari

SDK:

  • Added a new py.typed file to the SDK to make it compatible with mypy

Opik Dashboard:

  • Added duration chart to the project dashboard
  • Prompt metadata can now be set and viewed in the UI, this can be used to store any additional information about the prompt
  • Playground prompts and settings are now cached when you navigate away from the page

SDK:

  • Introduced a new OPIK_TRACK_DISABLE environment variable to disable the tracking of traces and spans
  • We now log usage information for traces logged using the LlamaIndex integration

SDK:

  • Improved error messages when getting a rate limit when using the evaluate method
  • Added support for a new metadata field in the Prompt object, this field is used to store any additional information about the prompt.
  • Updated the library used to create uuidv7 IDs
  • New Guardrails integration
  • New DSPY integration

Opik Dashboard:

  • The Opik playground is now in public preview
  • You can now view the prompt diff when updating a prompt from the UI
  • Errors in traces and spans are now displayed in the UI
  • Display agent graphs in the traces sidebar
  • Released a new plugin for the Kong AI Gateway

SDK:

  • Added support for serializing Pydantic models passed to decorated functions
  • Implemented get_experiment_by_id and get_experiment_by_name methods
  • Scoring metrics are now logged to the traces when using the evaluate method
  • New integration with aisuite
  • New integration with Haystack
Built with