Department of Radiology, Uppsala University Hospital

MIT researchers have developed an artificial intelligence system that streamlines medical image analysis by enabling rapid annotation of regions of interest, potentially transforming how scientists conduct clinical studies and treatment research.

The system called MultiverSeg addresses the time-consuming process of medical image segmentation, where researchers manually outline structures like brain regions in scans before conducting studies on disease progression or treatment efficacy, reports MIT News.

The technology allows users to segment biomedical imaging datasets through simple interactions such as clicking, drawing boxes, and scribbling on images. As researchers process additional images, the required user input decreases progressively until the model can perform accurate segmentations independently.

MultiverSeg’s architecture distinguishes it from existing medical segmentation tools by maintaining a context set of previously segmented images, which informs future predictions. This approach eliminates the need to retrain models for new tasks whilst avoiding repetitive work across image datasets.

The system requires no presegmented training data or extensive machine learning expertise, making it accessible to clinical researchers without computational backgrounds. Users can apply MultiverSeg to new segmentation tasks immediately without model retraining.

“Many scientists might only have time to segment a few images per day for their research because manual image segmentation is so time-consuming. Our hope is that this system will enable new science by allowing clinical researchers to conduct studies they were prohibited from doing before because of the lack of an efficient tool,” said Hallee Wong, the study’s lead author and electrical engineering and computer science graduate student.

Performance testing revealed MultiverSeg outperformed existing segmentation tools, requiring only two user clicks by the ninth image to generate segmentations more accurate than task-specific models. For certain image types like X-rays, manual segmentation of just one or two images suffices before the model achieves independent accuracy.

The research team, including Jose Javier Gonzalez Ortiz, Professor John Guttag, and senior author Adrian Dalca from Harvard Medical School and MIT’s Computer Science and Artificial Intelligence Laboratory, will present their findings at the International Conference on Computer Vision.

Leave a Reply

Your email address will not be published. Required fields are marked *

You May Also Like

Scientists find ‘brake’ in the brain that stops us starting stressful tasks

We all know the feeling: staring at a tax return or a…

Bosses should fund your knitting: Hobbies can boost workplace creativity

New Year’s resolutions to take up painting, coding or gardening might do…

‘Super agers’ win the genetic lottery twice to keep their memories young

People in their 80s who retain the sharp memories of those decades…

World’s first graviton detector hunts ‘impossible’ ghost particle of gravity

Physicists are building a machine to solve the biggest problem in science…