Skip to content

theveryhim/Polyp-detection

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

18 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Polyp detection

In this repo, we are going to employ Machine learning methods to detect colon tissues(polyp).

Data

PolypDB

  • WLI (White Light Imaging): Standard endoscopic view
  • NBI (Narrow Band Imaging): Enhanced vascular pattern visualization
  • LCI (Linked Color Imaging): Improved color contrast
  • FICE (Flexible Spectral Imaging Color Enhancement): Spectral enhancement
  • BLI (Blue Laser Imaging): Surface structure enhancement

REAL-Colon

REAL-Colon provides 60 full-resolution, real-world colonoscopy videos (2.7M frames) from multiple centers, with 350k expert-annotated polyp bounding boxes. Includes clinical metadata, acquisition details, and histopathology. Designed for robust CADe/CADx development and benchmarking. Released for non-commercial research. See the paper for details.

YOLOv9-FineTune:

  • Train
YOLOv9m summary (fused): 151 layers, 20,013,715 parameters, 0 gradients, 76.5 GFLOPs
                 Class     Images  Instances      Box(P          R      mAP50  mAP50-95)
                 all        359        382      0.933      0.912      0.963      0.816
  • detect polyps(bounding box)

Descriptive Alt Text Descriptive Alt Text

MambaYOLO-Train

Mamba-YOLO merges the state-space modeling efficiency of Mamba with the real-time detection strength of YOLOv8.
The architecture replaces the CSP backbone with a Selective Scan (Mamba) block, enabling long-range spatial dependency modeling at reduced computational cost.

This implementation targets medical image analysis, specifically polyp detection from multimodal colonoscopy datasets.


Architecture

Component Description
Backbone Mamba-based state-space selective scan layers replacing CSP blocks
Neck PANet-style feature pyramid
Head YOLOv8 detection head (multi-scale anchors)
Losses CIoU + BCE + objectness loss
Training Framework Ultralytics YOLO API
Hardware NVIDIA T4 (16 GB) × 2
Software Stack PyTorch 2.3.1 + CUDA 12.1, Python 3.11

Training Methodology

Parameter Value
Modality WLI
Epochs 300
Batch size 16
Optimizer AdamW
Image size 640×640
Scheduler Cosine annealing
Mixed precision AMP enabled

Descriptive Alt Text

Train metrix

Inference: WLI(Training modality)

Image 1 Image 2

Image 2 Image 2

Metrics of prediction on WLI

Image 1 Image 2

GroundTruth/Prediction on WLI

=== Test Set Metrics on trained modality(WLI)===
mAP50: 0.9138
mAP50-95: 0.7264
Precision: 0.8923
Recall: 0.8278

Inference: NBI-LCI-FICE-BLI

Image 1 Image 2

Image 2 Image 2

Metrics of prediction on NBI-LCI-FICE-BLI

Image 1 Image 2

GroundTruth/Prediction on NBI-LCI-FICE-BLI

=== NBI-LCI-FICE-BLI Metrics on unseen modalities===
mAP50: 0.6947
mAP50-95: 0.4972
Precision: 0.8402
Recall: 0.5672

Inference: REAL-Colon

Image 1 Image 2

Image 2 Image 2

Metrics of prediction on REAL-Colon

Image 1 Image 2

GroundTruth/Prediction on REAL-Colon

YOLOv11-Train

This project implements a robust polyp detection system using YOLO11 (You Only Look Once version 11) for medical image analysis. The model is trained on WLI modality.

Training Configuration

Model Architecture

  • Base Model: YOLO11 from Ultralytics
  • Input Resolution: 640×640 pixels
  • Backbone: CSPDarkNet
  • Neck: PANet
  • Head: Multi-scale detection

Hyperparameters

  • Epochs: 50
  • Batch Size: 16
  • Initial Learning Rate: 0.001
  • Optimizer: Auto-selected
  • Early Stopping Patience: 10 epochs

Data Augmentation

  • Mosaic: 0.8 probability
  • MixUp: 0.1 probability
  • Copy-Paste: 0.1 probability
  • Horizontal Flip: 0.5 probability
  • Color Augmentation: HSV adjustments
  • Spatial Transformations: Rotation, translation, scaling, shearing

Detection Parameters

  • Training Confidence Threshold: 0.1
  • IoU Threshold: 0.4
  • Augmentation Focus: Small object detection

Inference: WLI

=== Test Set Metrics on trained modality(WLI)===
mAP50: 0.9282
mAP50-95: 0.7067
Precision: 0.8799
Recall: 0.8638

Descriptive Alt Text Descriptive Alt Text Descriptive Alt Text

Inference: NBI-LCI-FICE-BLI

=== NBI-LCI-FICE-BLI Metrics on unseen modalities===
mAP50: 0.7619
mAP50-95: 0.5383
Precision: 0.8357
Recall: 0.6675

Descriptive Alt Text Descriptive Alt Text Descriptive Alt Text

Inference: REAL-Colon

Descriptive Alt Text Descriptive Alt Text Descriptive Alt Text

=== Metrics on unseen Dataset: REAL-Colon ===
mAP50: 0.3415
mAP50-95: 0.1922
Precision: 0.4986
Recall: 0.3413

About

Detecting colonoscopy polyps with aid of Deep networks

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published