4 research outputs found
Annotating 8,000 Abdominal CT Volumes for Multi-Organ Segmentation in Three Weeks
Annotating medical images, particularly for organ segmentation, is laborious
and time-consuming. For example, annotating an abdominal organ requires an
estimated rate of 30-60 minutes per CT volume based on the expertise of an
annotator and the size, visibility, and complexity of the organ. Therefore,
publicly available datasets for multi-organ segmentation are often limited in
data size and organ diversity. This paper proposes a systematic and efficient
method to expedite the annotation process for organ segmentation. We have
created the largest multi-organ dataset (by far) with the spleen, liver,
kidneys, stomach, gallbladder, pancreas, aorta, and IVC annotated in 8,448 CT
volumes, equating to 3.2 million slices. The conventional annotation methods
would take an experienced annotator up to 1,600 weeks (or roughly 30.8 years)
to complete this task. In contrast, our annotation method has accomplished this
task in three weeks (based on an 8-hour workday, five days a week) while
maintaining a similar or even better annotation quality. This achievement is
attributed to three unique properties of our method: (1) label bias reduction
using multiple pre-trained segmentation models, (2) effective error detection
in the model predictions, and (3) attention guidance for annotators to make
corrections on the most salient errors. Furthermore, we summarize the taxonomy
of common errors made by AI algorithms and annotators. This allows for
continuous refinement of both AI and annotations and significantly reduces the
annotation costs required to create large-scale datasets for a wider variety of
medical imaging tasks
MedShapeNet -- A Large-Scale Dataset of 3D Medical Shapes for Computer Vision
16 pagesPrior to the deep learning era, shape was commonly used to describe the objects. Nowadays, state-of-the-art (SOTA) algorithms in medical imaging are predominantly diverging from computer vision, where voxel grids, meshes, point clouds, and implicit surface models are used. This is seen from numerous shape-related publications in premier vision conferences as well as the growing popularity of ShapeNet (about 51,300 models) and Princeton ModelNet (127,915 models). For the medical domain, we present a large collection of anatomical shapes (e.g., bones, organs, vessels) and 3D models of surgical instrument, called MedShapeNet, created to facilitate the translation of data-driven vision algorithms to medical applications and to adapt SOTA vision algorithms to medical problems. As a unique feature, we directly model the majority of shapes on the imaging data of real patients. As of today, MedShapeNet includes 23 dataset with more than 100,000 shapes that are paired with annotations (ground truth). Our data is freely accessible via a web interface and a Python application programming interface (API) and can be used for discriminative, reconstructive, and variational benchmarks as well as various applications in virtual, augmented, or mixed reality, and 3D printing. Exemplary, we present use cases in the fields of classification of brain tumors, facial and skull reconstructions, multi-class anatomy completion, education, and 3D printing. In future, we will extend the data and improve the interfaces. The project pages are: https://medshapenet.ikim.nrw/ and https://github.com/Jianningli/medshapenet-feedbac
MedShapeNet -- A Large-Scale Dataset of 3D Medical Shapes for Computer Vision
16 pagesPrior to the deep learning era, shape was commonly used to describe the objects. Nowadays, state-of-the-art (SOTA) algorithms in medical imaging are predominantly diverging from computer vision, where voxel grids, meshes, point clouds, and implicit surface models are used. This is seen from numerous shape-related publications in premier vision conferences as well as the growing popularity of ShapeNet (about 51,300 models) and Princeton ModelNet (127,915 models). For the medical domain, we present a large collection of anatomical shapes (e.g., bones, organs, vessels) and 3D models of surgical instrument, called MedShapeNet, created to facilitate the translation of data-driven vision algorithms to medical applications and to adapt SOTA vision algorithms to medical problems. As a unique feature, we directly model the majority of shapes on the imaging data of real patients. As of today, MedShapeNet includes 23 dataset with more than 100,000 shapes that are paired with annotations (ground truth). Our data is freely accessible via a web interface and a Python application programming interface (API) and can be used for discriminative, reconstructive, and variational benchmarks as well as various applications in virtual, augmented, or mixed reality, and 3D printing. Exemplary, we present use cases in the fields of classification of brain tumors, facial and skull reconstructions, multi-class anatomy completion, education, and 3D printing. In future, we will extend the data and improve the interfaces. The project pages are: https://medshapenet.ikim.nrw/ and https://github.com/Jianningli/medshapenet-feedbac
MedShapeNet -- A Large-Scale Dataset of 3D Medical Shapes for Computer Vision
16 pagesPrior to the deep learning era, shape was commonly used to describe the objects. Nowadays, state-of-the-art (SOTA) algorithms in medical imaging are predominantly diverging from computer vision, where voxel grids, meshes, point clouds, and implicit surface models are used. This is seen from numerous shape-related publications in premier vision conferences as well as the growing popularity of ShapeNet (about 51,300 models) and Princeton ModelNet (127,915 models). For the medical domain, we present a large collection of anatomical shapes (e.g., bones, organs, vessels) and 3D models of surgical instrument, called MedShapeNet, created to facilitate the translation of data-driven vision algorithms to medical applications and to adapt SOTA vision algorithms to medical problems. As a unique feature, we directly model the majority of shapes on the imaging data of real patients. As of today, MedShapeNet includes 23 dataset with more than 100,000 shapes that are paired with annotations (ground truth). Our data is freely accessible via a web interface and a Python application programming interface (API) and can be used for discriminative, reconstructive, and variational benchmarks as well as various applications in virtual, augmented, or mixed reality, and 3D printing. Exemplary, we present use cases in the fields of classification of brain tumors, facial and skull reconstructions, multi-class anatomy completion, education, and 3D printing. In future, we will extend the data and improve the interfaces. The project pages are: https://medshapenet.ikim.nrw/ and https://github.com/Jianningli/medshapenet-feedbac