Skip to content
This repository was archived by the owner on Mar 25, 2026. It is now read-only.

Commit c91eca7

Browse files
author
MrSnapsClaws[bot]
committed
fix: remove outdated references from multimodal eval docs
1 parent 3d0917a commit c91eca7

1 file changed

Lines changed: 2 additions & 2 deletions

File tree

evaluations/experiments/multimodal-evaluation.mdx

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,13 +1,13 @@
11
---
22
title: Multimodal Evaluation — Images, PDFs, and Vision
33
sidebarTitle: Multimodal Evaluation
4-
description: Evaluate image generation, virtual try-on, document parsing, and other multimodal AI pipelines with LLM-as-a-Judge vision models.
4+
description: Evaluate image generation, document parsing, and other multimodal AI pipelines with LLM-as-a-Judge vision models.
55
---
66

77
LangWatch supports multimodal evaluation out of the box. You can evaluate image inputs and outputs using any vision-capable model (GPT-4o, GPT-5.2, Claude Sonnet, Gemini, etc.) through the built-in LLM-as-a-Judge evaluators — no custom code required.
88

99
This covers common multimodal use cases:
10-
- **Image generation quality** — score outputs of diffusion models, GANs, or other image generators
10+
- **Image generation quality** — score outputs of image generation models
1111
- **Document parsing** — evaluate extracted metadata from PDFs and scanned documents
1212
- **Content moderation** — detect NSFW or low-quality uploaded images
1313
- **Visual QA** — evaluate answers to questions about images

0 commit comments

Comments
 (0)