Virgo

Foundation models for endoscopy

Virgo builds frontier AI models to solve colorectal cancer.

Press & podcast coverage

Podcast

Episode 39 — Interview with Matt Schwartz

The Interventional Endoscopist Dec 2, 2025
Press

Virgo & Rajpurkar Lab partner on next-gen endoscopy AI foundation model

EIN Presswire Apr 29, 2026
Podcast

Latest conversation with Matt Schwartz

Scope Forward Mar 21, 2026
Press

Foundation Model Platforms & APIs Market Map

Elion Health Jan 20, 2026
Podcast

Building foundation models for endoscopy

Nebius for Startups Apr 2, 2026
Press

AI in Endoscopy — Market Report

Navistrat Analytics 2026
Podcast

Inside Virgo's research roadmap

Scope Forward Feb 21, 2026
Press

Virgo launches EndoML at DDW 2025

PR Newswire May 2, 2025
Podcast

AI in Endoscopy — turning lost video into insight

HealthTech Remedy Nov 12, 2025
Press

Virgo launches AI-powered EndoML platform

Practical Patient Care May 5, 2025
Podcast

Countdown to AI super-intelligence in GI

Scope Forward May 28, 2025
Press

Meet EndoDINO — a SOTA foundation model for endoscopy

Cerebral Valley Jan 10, 2025
Podcast

Foundation Model Series — Advancing Endoscopy

Impact AI Feb 24, 2025
Press

Virgo launches EndoML powered by EndoDINO

PR Newswire Jan 14, 2025
Podcast

AGA Innovation conversation

AGA Innovation Talk
Press

EndoDINO — paper

arXiv Jan 8, 2025
Podcast

Founding Virgo

GI Startup Podcast Nov 20, 2022
Podcast

The future of endoscopy data

HealthBiz Podcast 2023

EndoDINO

The most powerful AI for endoscopy.
Trained on the largest dataset.

Pre-training scale · videos

Largest endoscopy video dataset in the literature

EndoDINO (Virgo) 130,037 videos

Dermyer et al. 2025

EndoFM 33,000 videos

Wang et al. 2023

Etro (Roche) 5,145 videos

Yao et al. 2023

ArgesFM (J&J) 3,927 videos

Chaitanya et al. 2024

DovaVision 845 videos

Byrne et al. 2025

Pre-training scale · images

57× more frames than the next largest dataset

EndoDINO (Virgo)

Dermyer et al. 2025

3.5B frames

ArgesFM (J&J)

Chaitanya et al. 2024

61M frames

GastroNet-5M

Jong et al. 2026

4.8M frames

Etro (Roche)

Yao et al. 2023

526K frames
EndoDINO 3.5B frames

Validation

State-of-the-art on every benchmark.
Validated across sites, scopes, and patient populations.

HyperKvasir · 3-class Mayo Endoscopic Scoring

State-of-the-art with frozen features

Etrolizumab (DINOv1) 0.706

Schwartz et al., 2023

EndoFM (DINOv1) 0.699

Wang et al., MICCAI 2023

DenseNet (Supervised) 0.729

Huang et al., CVPR 2017

DINOv2 ViT-g/14 (LVD-142M) 0.735

Oquab et al., Meta AI 2024

EndoDINO ViT-g/14 0.748

Virgo, 2025

Macro F1, linear probe on frozen backbone. Comparator values from each model's original publication; see chart for citations.

UNIFI Phase 3 · Ustekinumab in UC

Predicting 8-week endoscopic healing from baseline video

Placebo arm

EndoDINO features 0.78
Standard UC covariates 0.70

Treatment arm

EndoDINO features 0.75
Standard UC covariates 0.72

AUROC for 8-week endoscopic healing (MES ≤ 1).

AUROC, 5-fold CV. EndoDINO video embeddings vs. 21 standard UC clinical covariates. Data presented at UEGW 2025.

Demographic diversity · procedure-weighted

The most demographically representative endoscopy dataset

Virgo (EndoDINO training) index 0.713

169 centers · multi-continental

PolypGen index 0.460

6 centers · Europe + Africa

Hyper-Kvasir index 0.280

1 center · Norway

LDPolypVideo index 0.095

1 center · China

White
Black
Asian
Hispanic
Other

Virgo: 1,053,880 US procedures across 148 centers (Sept 2025 audit). Public dataset demographics from each dataset's published documentation. Diversity index = Shannon entropy across White / Black / Asian / Hispanic / Other; higher is more balanced.

Out-of-distribution validation · external datasets

Validated across new sites, scopes, and disease areas

HyperKvasir

Colonoscopy

Mayo Endoscopic Scoring

Bærum Hospital, Norway

Kvasir-Capsule

Capsule endoscopy

Lesion / anatomy classification

Bærum Hospital, Norway

CholecT50

Laparoscopy

Surgical action triplets

IHU Strasbourg, France

SUN

Colonoscopy

Polyp detection

Showa University, Japan

UNIFI Phase 3

Colonoscopy

Endoscopic healing prediction

Janssen multi-site trial

YODA / external UC cohort

Colonoscopy

MES (QWK 0.83)

Independent academic centers

Competitors (e.g. DovaVision UC, Iterative Health) typically train and evaluate on a single internal data source. EndoDINO is pre-trained on Virgo's corpus and validated on independent public benchmarks and external clinical trial cohorts.

148

US medical centers

vs. 1–6 in public datasets

1.05M

US procedures with demographics

procedure-weighted, not patient-weighted

46.6%

non-White representation

12.9% Black · 15.1% Asian · 13.9% Hispanic

0.713

Shannon diversity index

Hyper-Kvasir 0.28 · LDPolypVideo 0.10

HyperKvasir · 4-class Mayo Endoscopic Scoring

State-of-the-art with frozen features.

EndoDINO ViT-g/14 delivers leading performance on Mayo endoscopic scoring with a frozen backbone.

Model
EndoDINO ViT-g/14
Data
130K+ procedures
Endoscopic lumen view representing Mayo 0: Normal or inactive disease

Mayo 0

Normal or inactive disease

Conf 0.97
Endoscopic lumen view representing Mayo 1: Mild disease

Mayo 1

Mild disease

Conf 0.92
Endoscopic lumen view representing Mayo 2: Moderate disease

Mayo 2

Moderate disease

Conf 0.90
Endoscopic lumen view representing Mayo 3: Severe disease

Mayo 3

Severe disease

Conf 0.91

Frame-level predictions aggregated per procedure.

Scored by EndoDINO • Inference latency 14.8 ms.

Macro F1 0.748 · Linear probe on frozen backbone

How EndoDINO learns

From raw procedure video to a shared representation layer for GI.

One model. Every downstream task: scoring, detection, prediction, biomarker discovery.

01

Capture

Raw endoscopy video from the procedure stream.

02

Structure

Frames organized, deduplicated, temporally aligned.

03

Pretrain

Self-supervised learning at population scale.

04

Represent

A reusable embedding for any downstream task.

Capabilities

One representation layer. Many downstream tasks.

Proof point · UNIFI, Phase 3 UC

7 mo

Saved

$38M

Avoided

Validated on Stelara Phase 3 UC trial data. UNIFI could have reached the same readout faster and at lower cost using EndoDINO as a covariate.

  • 01

    Placebo response

    Covariate models that reduce trial size and accelerate enrollment.

  • 02

    Subgroup response

    Precision enrichment: identify likely responders before randomization.

  • 03

    Continuous AI scores

    UC and CD efficacy assessment beyond Mayo and SES-CD categories.

  • 04

    Bayesian priors

    Real-world evidence priors from EndoDINO at population scale.

The data moat

Endoscopy video is the substrate. We capture more of it, from more procedures, than anyone else. The archive grows every day.

Capture is the foundation of everything downstream. Real-world endoscopy video (at population scale, longitudinal, and continuously growing) is what makes a foundation model for GI possible. Models built on smaller datasets plateau. Ours compound.

  • 3M+ Procedures recorded across partner sites through 2025
  • 1M+ New procedures captured each year, and growing
  • 3.5B Video frames already in the EndoDINO training set
  • 24/7 Live capture pipeline across institutional partners

The platform

One model base. Built for the full procedure.

Foundation model

01

EndoDINO

Virgo's foundation model for endoscopy. One model base for scoring, prediction, detection, and biomarker work, trained on the full procedure, not just the frame.

Build environment

02

EndoML

The environment for building on top of EndoDINO. A GI-specific model layer for clinical and research workflows.

Request access

See the full evidence package.

Manuscript, UEGW 2025 poster, benchmark results, and partnership models. Sent directly to qualified researchers and partners.

Prefer email? research@virgosvs.com

Infrastructure for the future of endoscopy AI.