diff --git a/docs/images/Fig1a.png b/docs/assets/Fig1a.png similarity index 100% rename from docs/images/Fig1a.png rename to docs/assets/Fig1a.png diff --git a/docs/assets/Fig1c.png b/docs/assets/Fig1c.png new file mode 100644 index 0000000..3cbeabc Binary files /dev/null and b/docs/assets/Fig1c.png differ diff --git a/docs/assets/Fig2b.png b/docs/assets/Fig2b.png new file mode 100644 index 0000000..f6e79a1 Binary files /dev/null and b/docs/assets/Fig2b.png differ diff --git a/docs/assets/Patch_Encoder.png b/docs/assets/Patch_Encoder.png new file mode 100644 index 0000000..ea912e7 Binary files /dev/null and b/docs/assets/Patch_Encoder.png differ diff --git a/docs/assets/Patch_Writting.png b/docs/assets/Patch_Writting.png new file mode 100644 index 0000000..5fa2a7b Binary files /dev/null and b/docs/assets/Patch_Writting.png differ diff --git a/docs/assets/Tissue.png b/docs/assets/Tissue.png new file mode 100644 index 0000000..188ae6d Binary files /dev/null and b/docs/assets/Tissue.png differ diff --git a/docs/assets/Tissue_Grid.png b/docs/assets/Tissue_Grid.png new file mode 100644 index 0000000..0a12c2b Binary files /dev/null and b/docs/assets/Tissue_Grid.png differ diff --git a/docs/assets/Tissue_Mask.png b/docs/assets/Tissue_Mask.png new file mode 100644 index 0000000..740ad0b Binary files /dev/null and b/docs/assets/Tissue_Mask.png differ diff --git a/docs/index.html b/docs/index.html index 855b33b..03f7a5b 100644 --- a/docs/index.html +++ b/docs/index.html @@ -9,13 +9,19 @@
+ +Patch extraction from gigapixel whole-slide images, typically guided by tissue detection methods, forms the backbone of computational pathology workflows, - and remains a major computational bottleneck. Here we present AtlasPatch, an efficient and scalable slide preprocessing framework designed to enable accurate tissue - detection and high-throughput patch extraction with minimal computational overhead. To ensure robust and generalizable slide processing, we curated and - semi-manually annotated a diverse dataset of approximately 35,000 whole-slide image thumbnails spanning multiple cohorts, tissue types, and organ systems. Using this - corpus, we implement selective finetuning of the normalization layers of the Segment-Anything2 model for efficient thumbnail-level segmentation. This strategy ensures - fast processing while achieving segmentation accuracy comparable to—or exceeding—that of existing methods. From the thumbnail masks, tissue coordinates are efficiently - extrapolated to full-resolution slides for precise patch extraction. The entire pipeline is parallelized for both CPU and GPU execution to maximize throughput. We assess - AtlasPatch across segmentation accuracy, computational complexity, and downstream multiple-instance learning performance, showing consistent predictive power with - state-of-the-art methods while operating at a fraction of their computational cost.
- -
+ Whole-slide image (WSI) preprocessing, typically comprising tissue detection followed by patch extraction, is foundational to AI-driven + and image-based computational pathology workflows. This remains a major computational bottleneck as existing tools either rely on inacurate + heuristic thresholding for tissue detection, or adopt AI-based approaches trained on limited-diversity data that operate at the patch level, + incurring substantial computational complexity. We present AtlasPatch, an efficient and scalable slide preprocessing framework for accurate + tissue detection and high-throughput patch extraction with minimal computational overhead. AtlasPatch’s tissue detection module is trained + on a heterogeneous and semi-manually annotated dataset of $\sim$35,000 WSI thumbnails, using efficient fine-tuning of the Segment Anything2 + model. The tool extrapolates tissue masks from thumbnails to full-resolution slides to extract patch coordinates at user-specified magnifications, + with options to stream patches directly into commonly used image encoders for embedding generation or export patch images for storage, all efficiently + parallelized across CPUs and GPUs to maximize throughput. We assess AtlasPatch across segmentation accuracy, computational complexity, and downstream + multiple-instance learning, matching state-of-the-art performance while operating at a fraction of their computational cost.
+ +
+ AtlasPatch efficiently segments tissue regions from whole-slide images using a fine-tuned Segment-Anything2 (SAM2) model.
+
+
+ AtlasPatch efficiently extracts patch coordinates from the generated SAM2 tissue masks.
+
+
+ Can perform feature embedding with numerous feature encoders available, with the use of custom encoders also possible.
+
+ Can save and export tissue patch images for patch visualization or for downstream task use.
+
+ Our high quality tissue detector generates masks using Segment Anything Model 2 (SAM2), finetuned using a large + and diverse annotated dataset. This dataset, comprised of over 35,000 whole-slide image (WSI) thumbnails, was curated + to cover several organs and tissue types, institutions, scanner vendors, acquisition protocols, and covering variations + in illumination, tissue fragment number and size, tissue boundary definition, and histologic heterogeneity. We finetuned + the SAM2 model by freezing the backbone and training only the normalization layers for the tissue-versus-background task.
+ +
+ AtlasPatch
+CLAM
+Trident-GrandQC
+Trident-Hest
+All runs shown compare the speed of image segmentation and patch extraction of 100 whole-slide images run on the + same computer hardware (time sped up 10x).
+
+ @software{atlaspatch,
+ title = {AtlasPatch},
+ author = {Ahmed Alagha, Christopher Leclerc, Yousef Hassan, Omar Abdelwahed, Calvin Moras, Peter Rentopoulos, Rose Rostami,
+ Bich Ngoc Nguyen, Jumanah Baig, Abdelhakim Khellaf, Vincent Quoc-Huy Trinh, Rabeb Mizouni, Hadi Otrok, Jamal Bentahar,
+ Mahdi S. Hosseini},
+ year = {2025},
+ url = {https://github.com/AtlasAnalyticsLab/AtlasPatch},
+ }
+
+