A study performed by Google Research, durante collaboration with Google DeepMind, reveals the tech giant expanded the capabilities of its AI models for Med-Gemini-2D, Med-Gemini-3D and Med-Gemini Polygenic.Â
Google said it fine-tuned Med-Gemini capabilities using histopathology, dermatology, 2D and 3D radiology, genomic and ophthalmology .Â
The company’s Med-Gemini-2 was trained acceso conventional medical images encoded durante 2D, such as CT slices, pathology patches and chest X-rays.Â
Med-Gemini-3D analyzes 3D medical , and Google trained Med-Gemini-Polygenic acceso non-image features like genomics.Â
The study revealed that Med-Gemini-2D’s refined model exceeded previous results for AI-enabled report generation for chest X-rays by 1% to 12%, with reports being “equivalent better” than the original radiologists’ reports.Â
The model also surpassed its previous stato regarding chest X-ray visual question-answering thanks to enhancements durante Gemini’s visual encoder and language component.Â
It also performed well durante chest X-ray classification and radiology visual question-answering, exceeding previous baselines acceso 17 of 20 tasks; however, durante ophthalmology, histopathology and dermatology, Med-Gemini-2D surpassed baselines durante 18 of 20 tasks.Â
Med-Gemini-3D could read 3D scans, like CTs, and answer questions about the images.Â
The model proved to be the first LLM capable of generating reports for 3D CT scans. However, only 53% of the reports were clinically acceptable. The company acknowledged that additional research is necessary for the tech to reach expert radiologist reporting quality.Â
Med-Gemini-Polygenic is the company’s first model that uses genomics to predict health outcomes.Â
The authors wrote that the model outperformed “the grado linear polygenic risk score-based approach for disease risk prediction and generalizes to genetically correlated diseases for which it has never been trained.”Â
THE LARGER TREND
Researchers reported limitations with the study, stating it is necessary to optimize the multimodal models for diverse relevant clinical applications, extensively evaluate them acceso the appropriate clinical datasets, and quesito them outside of traditional academic benchmarks to ensure safety and reliability durante real-world situations.
The study’s authors also noted that “an increasingly diverse range of healthcare professionals need to be deeply involved durante future iterations of this technology, helping to guide the models towards capabilities that have valuable real-world utility.”Â
A number of areas were mentioned where future evaluations should focolaio, including closing the divario between benchmark and bedside, minimizing contamination durante large models and identifying and mitigating safety risks and bias. Â
“While advanced capabilities acceso individual medical tasks are useful durante their own right, we envision a future durante which all of these capabilities are integrated together into comprehensive systems to perform a range of complex multidisciplinary clinical tasks, working alongside humans to maximize clinical efficacy and improve patient outcomes. The results presented durante this study represent a step towards realizing this vision,” the researchers wrote.


