Department of Radiology, Uppsala University Hospital

MIT researchers have developed an artificial intelligence system that streamlines medical image analysis by enabling rapid annotation of regions of interest, potentially transforming how scientists conduct clinical studies and treatment research.

The system called MultiverSeg addresses the time-consuming process of medical image segmentation, where researchers manually outline structures like brain regions in scans before conducting studies on disease progression or treatment efficacy, reports MIT News.

The technology allows users to segment biomedical imaging datasets through simple interactions such as clicking, drawing boxes, and scribbling on images. As researchers process additional images, the required user input decreases progressively until the model can perform accurate segmentations independently.

MultiverSeg’s architecture distinguishes it from existing medical segmentation tools by maintaining a context set of previously segmented images, which informs future predictions. This approach eliminates the need to retrain models for new tasks whilst avoiding repetitive work across image datasets.

The system requires no presegmented training data or extensive machine learning expertise, making it accessible to clinical researchers without computational backgrounds. Users can apply MultiverSeg to new segmentation tasks immediately without model retraining.

“Many scientists might only have time to segment a few images per day for their research because manual image segmentation is so time-consuming. Our hope is that this system will enable new science by allowing clinical researchers to conduct studies they were prohibited from doing before because of the lack of an efficient tool,” said Hallee Wong, the study’s lead author and electrical engineering and computer science graduate student.

Performance testing revealed MultiverSeg outperformed existing segmentation tools, requiring only two user clicks by the ninth image to generate segmentations more accurate than task-specific models. For certain image types like X-rays, manual segmentation of just one or two images suffices before the model achieves independent accuracy.

The research team, including Jose Javier Gonzalez Ortiz, Professor John Guttag, and senior author Adrian Dalca from Harvard Medical School and MIT’s Computer Science and Artificial Intelligence Laboratory, will present their findings at the International Conference on Computer Vision.

Leave a Reply

Your email address will not be published. Required fields are marked *

You May Also Like

Meta launches ad-free subscriptions after ICO forces compliance changes

Meta will offer UK users paid subscriptions to use Facebook and Instagram…

World nears quarter million crypto millionaires in historic wealth boom

Global cryptocurrency millionaires have reached 241,700 individuals, marking a 40 per cent…

Wong warns AI nuclear weapons threaten future of humanity at UN

Australia’s Foreign Minister Penny Wong has warned that artificial intelligence’s potential use…

Legal scholar warns AI could devalue humanity without urgent regulatory action

Artificial intelligence systems pose worldwide threats to human dignity by potentially reducing…

AI creates living viruses for first time as scientists make artificial “life”

Stanford University researchers have achieved a scientific milestone by creating the world’s…

MIT accelerator shows AI enhances startup building without replacing core principles

Entrepreneurs participating in MIT’s flagship summer programme are integrating artificial intelligence tools…

Engineers create smarter artificial intelligence for power grids and autonomous vehicles

Researchers have developed an artificial intelligence system that manages complex networks where…

Artificial intelligence threatens subtitle writers despite creative demands of accessibility work

Professional subtitle creators face declining wages and job insecurity as artificial intelligence…