We propose a scalable computerized approach for large-scale inference of Liver Imaging Reporting and Data System (LI-RADS) final assessment categories in narrative ultrasound (US) reports. Although our model was trained on reports created using a LI-RADS template, it was also able to infer LI-RADS scoring for unstructured reports that were created before the LI-RADS guidelines were established. No human-labelled data was required in any step of this study; for training, LI-RADS scores were automatically extracted from those reports that contained structured LI-RADS scores, and it translated the derived knowledge to reasoning on unstructured radiology reports. By providing automated LI-RADS categorization, our approach may enable standardizing screening recommendations and treatment planning of patients at risk for hepatocellular carcinoma, and it may facilitate AI-based healthcare research with US images by offering large scale text mining and data gathering opportunities from standard hospital clinical data repositories.

Learning Objective 1: Develop a computerized approach to follow a HCC screening patient longitudinally (throughout multiple years) whose ultrasound screening exam was reported before US-LI-RADS was implemented.


Imon Banerjee (Presenter)
Stanford University

Hailey Choi, Stanford University School of Medicine
Terry Desser, Stanford University School of Medicine
Daniel Rubin, Stanford University

Presentation Materials: