Representation Learning

Medical AI has long depended on handcrafted features and narrow datasets built for single tasks. Ataraxis takes a different approach. We are building foundation models that learn effective representations from raw complexity of medical data, uncovering hidden biological structure within histopathology without manual annotation.

State-of-the-art foundation models for digital pathology

Utilizing novel hyperparameter optimization strategies and a growing dataset of 260,000 whole slide images (2 billion image tiles) across 25 organs, Ataraxis™ Falcon foundation model outperformed the strongest public foundation models for histopathology, including Prov-GigaPath and UNI-2, on a suite of 24 pan-cancer tasks.

Bar chart showing mean performance (AUC) over 24 pan-cancer tasks for models Kestrel 2023 (0.620), GigaPath 2024 (0.633), UNI 2 2025 (0.645), and Falcon Q1 2026 (0.652), with Falcon noted as 3 times larger than Kestrel and trained on 5 times as much data.
“Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam.”
—Joe Cappadona
AI Research Scientist

Built to generalize across diseases and clinical contexts

Our models learn rich data representations by combining standard H&E pathology with routine clinical variables, discovering the most predictive signals on their own rather than being told what to look for. That shared foundation adapts to new clinical questions over time, enabling a growing portfolio of tests that deliver calibrated, patient-level predictions for real clinical decisions: recurrence risk, treatment benefit and biomarker discovery.

From routine pathology to scalable clinical representations

Four-step diagram showing breast tumor biopsy, digitization of H&E pathology slide, division of the slide into small tissue patches of 128 by 128 micrometers, and analysis of morphologic features by AI models.
1
Core needle biopsy of resection
Illustration of breast tissue undergoing a biopsy procedure with surgical tools extracting a tissue sample.
2
H&E pathology slide with breast tumor is digitized
Mobile phone screen showing a microscopic image of purple-stained cells, resembling a pathology slide.
3
Whole slide image divided into small tissue patches
Microscopic view of tissue cells stained pink with a 128μm scale bar and a zoomed-in inset showing detailed cell structure.
4
Morphologic features from each area are analyzed by AI models, and the most important features are identified
Yellow triangle with three upward-pointing chevrons inside on a white rounded square background.

Built differently from the ground up

search icon
Pan-Cancer Foundation Models

Our models are pretrained on unlabeled histopathology images spanning multiple cancer types. Exposure to a broader range of cellular and tissue patterns improves generalization, and cross-cancer pretraining enables to transfer knowledge more effectively to downstream clinical tasks than disease-specific training.

Abstract target icon
Pathology-Native Augmentation

Staining protocols, scanners, and tissue preparation vary across hospitals. We developed augmentation pipelines that simulate this variability during training, so our models learn features that are robust to site-specific artifacts and perform consistently across institutions.

node icon
Gene Expression Prediction

Our models are trained to predict gene expression directly from standard H&E-stained slides without requiring a separate genomic assay. This allows Ataraxis tests to incorporate molecular signals that genomic-only tests depend on expensive sequencing to access, making our platform more comprehensive and adaptable across clinical questions.