Maxim AI release notes
Maxim AI release notes
www.getmaxim.ai

⚙️ Flexi evals

 

New

  

We’ve made evaluations on Maxim logs fully configurable. Instead of being limited to predefined parameters like input, output, retrieval, etc, you can now decide exactly which value in your trace or session should serve as the “input,” “output,” or any other field for your evaluators.

Key highlights of Flexi evals:

  • Custom mapping: Configure any element of a trace/session to serve as evaluator fields, such as inputs, outputs, etc.
  • Programmatic flexibility: Create custom code blocks (in JS) to extract or combine fields and map them to any evaluator parameter. You can pull values from JSON, perform string manipulations, or apply validations to shape evaluations however you need.

This gives teams greater control over how evaluations are run on Maxim – allowing them to focus on specific areas of LLM interactions, eliminate noise from evaluation parameters, and generate more precise, actionable insights.

🚀 Google Cloud Marketplace x Maxim AI

 

New

  

We’re excited to share that Maxim is now available on the Google Cloud Marketplace. This makes it even easier for our customers, especially enterprises already on Google Cloud, to integrate Maxim into their AI development workflows.

Through the Marketplace, customers gain access to Maxim’s powerful simulation, evaluation, and observability infrastructure to ship reliable AI applications with the speed and quality required for real-world use -- while benefitting from centralized GCP billing. For customers who prefer full control over their data, the Maxim platform is also available as a self-hosted deployment within their own Google Cloud environment.

🌐 Checkout Maxim on Marketplace

🛡️ Read more about self-hosting/ in-VPC deployments

🗂️ Workspace duplication

 

New

  

Teams can now duplicate an entire Maxim workspace, making it easier to set up new workspaces by reusing the workflows and assets of an existing one.

Key highlights:

  • What’s duplicated: Prompts, agents (via HTTP endpoint), voice agents, and no-code agents are duplicated along with session and version history. In addition to these, prompt tools, datasets, prompt partials, and evaluators are also eligible.
  • What’s not duplicated: Log repositories, context sources, evaluation runs, and dashboards.
  • Access control: You can decide whether the users of the original workspace should also gain access to the duplicated one.

This gives you full flexibility to select which components you want to carry over into the new workspace.

📈 Custom metric support

 

New

  

We’ve introduced custom metric support, giving teams full flexibility to log and track the KPIs that matter most beyond the default metrics that are already logged. You can now push any metric as part of your traces, generations, or retrievals via the Maxim SDK.

These metrics can be plotted on Maxim’s built-in or custom dashboards, used in evaluators, and even tied to alerts – providing instant visibility into the signals that matter.

👨‍💻 SAML-based Single Sign-On (SSO)

 

New

  

We’ve added support for SAML-based Single Sign-On (SSO) in Maxim, starting with integrations for Okta and Google Workspace. This enables teams to connect Maxim to their Identity Provider (IdP) and manage access centrally. Users who are assigned permission to Maxim within Okta or Google Workspace can log in seamlessly through SSO, ensuring secure and simplified onboarding to the platform.

Set up SSO with Okta: https://www.getmaxim.ai/docs/settings/setup-sso-with-okta

Set up SSO with Google: https://www.getmaxim.ai/docs/settings/setup-sso-with-google

image.png

🚀 New providers: OpenRouter and Cerebras

 

New

  

Maxim now supports two more providers – OpenRouter and Cerebras. OpenRouter gives you the flexibility to connect with a wide range of popular open-source and hosted models, and

Cerebras enables running large-scale models with low latency and efficient compute.

🚀 OpenAI's GPT-5 model is live on Maxim

 

New

  

OpenAI's GPT-5 is now available on Maxim. Use the latest GPT-5 model, offering stronger reasoning, enhanced multi-turn dialogue, expanded context, and multimodal support to power your experimentation and evaluation workflows.

Start using this model via the OpenAI or Azure provider: ✅ Go to Settings > Models > Select OpenAI or Azure provider > Add GPT-5

Human annotation on logs: Revamped

 

New

  

We’ve simplified the experience for human evaluation of logs. You can now add annotations and scores for each human evaluator directly from the main logs table, eliminating the need to create separate annotation queues.

With this update, you can evaluate response quality more efficiently – either by adding annotations for individual evaluators directly in the table, or by switching to the detailed trace/session view to annotate for all human evaluators at once.

📂 Datasets now support file attachments

 

New

  

You can now attach image, audio, and PDF files to your test datasets in Maxim and use them for your evaluation workflows. This enhancement allows you to prototype complex document/file processing flows and experiment with a wider variety of use cases directly on Maxim.

File types supported:

  • Images: JPEG, PNG, TIFF, SVG, WebP, BMP
  • Audio: MP3, WAV, FLAC, M4A, OGG, AAC

This gives you greater flexibility when building high-quality multimodal applications.

image.png

🚀 xAI's Grok 4 model is live on Maxim!

 

New

  

Grok 4, xAI’s latest flagship LLM, is now available on Maxim. Access powerful capabilities like PhD‑level reasoning, a 256k token context window, and advanced math performance to supercharge your experimentation and evaluation workflows.

✅ Start using these models via the xAI provider on Maxim: Go to Settings > Models > Select xAI provider > Add Grok 4 to start using.