Skip to content

pipiwang/ProFound

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

23 Commits
 
 
 
 
 
 

Repository files navigation

ProFound Logo

Python 3.10 PyTorch 2.6 MONAI 1.5 CUDA 11.8 License: Apache 2.0 Hugging Face Demo

ProFound: Vision Foundation Models for Prostate Multiparametric MR Images

ProFound is a suite of vision foundation models, pre-trained on multiparametric 3D magnetic resonance (MR) images from large collections of prostate cancer patients.

We aim to open-source all code for pre-training, fine-tuning, and evaluation, together with weights of pre-trained and fine-tuned ProFound models. This is an ongoing effort, so please check back later for updates.

🤗 Try ProFound Online

Interact with ProFound directly in your browser via our Hugging Face Space.

🐣 Downstream Clinical Tasks

Profound can be fine-tuned for a wide range of prostate imaging tasks. Switch to the demo branch for examples:

git checkout demo
  • Download weights and example images here.

  • Decompress (if needed) and place the downloaded folders, checkpoints and demo, at the repository root directory.

  • Configure dependencies.

    • Install PyTorch version specified in requirements-pytorch.txt.
    • Install MONAI and other packages:
    pip install -r requirements.txt
  • Run the following tasks:

Radiological cancer classification

  • Run:
    sh demo_run_classification.sh

Lesion segmentation

  • Run:
    sh demo_run_lesion_segmentation.sh

Anatomy segmentation

  • Run:
    sh demo_run_anatomy_segmentation.sh

More tasks are on the way...

🥚 Pre-trained Models

Available models

Pre-trained on approximately 5,000 international, cross-institute, multiparametric prostate MRI studies, each of which includes T2w, ADC and high-b DWI volumes

More models coming soon!

🤝 Contact

Open an issue for questions and feedback.

🌞 Acknowledgement

This work is supported by the International Alliance for Cancer Early Detection, an alliance between Cancer Research UK, Canary Center at Stanford University, the University of Cambridge, OHSU Knight Cancer Institute, University College London and the University of Manchester.

This work is also supported by the National Institute for Health Research University College London Hospitals Biomedical Research Centre.

The authors acknowledge the use of resources provided by the Isambard-AI National AI Research Resource (AIRR). Isambard-AI is operated by the University of Bristol and is funded by the UK Government’s Department for Science, Innovation and Technology (DSIT) via UK Research and Innovation; and the Science and Technology Facilities Council [ST/AIRR/I-A-I/1023].


About

Vision Foundation Models for Prostate Multiparametric MR Images

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published