Project Details
Abstract
The proposed project aims to advance medical artificial intelligence (AI) by developing an interactive multimodal AI system capable of performing high-precision diagnostic tasks guided by natural language instructions. This research addresses a critical challenge in AI-assisted medicine: the need for systems that not only process multimodal data (e.g., vision and text) but also interact with clinicians to perform complex visual reasoning tasks, such as identifying, segmenting, and counting medical structures from clinical images. Modern diagnostic workflows increasingly depend on image-intensive procedures, ranging from histopathological slides and radiology scans to microscopic cell images. However, current AI systems are typically static, task specific, non-interactive and heavily dependent on large amounts of manually labeled datasets. Moreover, they often lack transparency and adaptability to specific diagnostic questions posed by physicians. To this end, this project addresses these limitations by creating a general-purpose interactive AI model that interprets language-based instructions and performs corresponding visual analysis tasks in a clinical setting.
Submitting Institute Name
Hamad Bin Khalifa University (HBKU)
| Sponsor's Award Number | HBKU-INT-VPR-TG-03-07 |
|---|---|
| Proposal ID | HBKU-OVPR-TG-03-12 |
| Status | Active |
| Effective start/end date | 1/11/25 → 31/10/27 |
Collaborative partners
- Hamad Bin Khalifa University (lead)
- Sidra Medicine
Primary Theme
- Artificial Intelligence
Primary Subtheme
- AI - Healthcare
Secondary Theme
- Precision Health
Secondary Subtheme
- PH - Diagnosis Treatment
Keywords
- Multimodal AI
- Interactive Operation
- High-precision Diagnosis
Fingerprint
Explore the research topics touched on by this project. These labels are generated based on the underlying awards/grants. Together they form a unique fingerprint.