Browsing Anatomical Image Databases: A Case Study of the Visible Human

Chris North, Flip Korn
Human-Computer Interaction Laboratory and
Department of Computer Science
University of Maryland, College Park, MD 20742 USA
+1-301-405-2725
{north, flip}@cs.umd.edu

To appear in Video Proc. ACM CHI '96 Conf, ACM Press, 1996.
(Video also available in HCIL Video Reports 1995.)

Copyright on this material is held by the author(s).


ABSTRACT

This video demonstrates two user interface prototypes for browsing the National Library of Medicine Visible Human dataset on the internet. The first uses a graphical approach and demonstrates a general interface for exploring volumetric data. The second uses a textual approach for exploring hierarchical information containing inter-relationships.

Keywords

User interface, information exploration, digital library, medical imaging, volume visualization, hierarchical information, network access

INTRODUCTION

The U.S. National Library of Medicine (NLM), for its Visible Human Project [1], is in the process of creating a large digital library of anatomical images of both a male and female subject. It contains MRI and CT scans as well as cryosection images (digital color photos of axial cross- sections). It is available via the internet to a large community of users with varying backgrounds and expertise. However, since the dataset is extremely large (15 Gigabytes for the male) and could take weeks to download, it is important for users to be able to retrieve only their desired subset of images. We propose two user interface prototypes to assist users in browsing the dataset and downloading desired images.

A GRAPHICAL APPROACH

The first prototype [3] utilizes users' knowledge of the location and visual appearance of anatomical structures within the body. It allows the user to browse a miniature version of the Visible Human digital body in order to select and retrieve images. The direct manipulation interface (Figure 1) presents the user with a coordinated pair of orthogonal 2D cross-section views of the body. The left view-window displays a coronal section (a front-view longitudinal cut of the body), which gives an overview of the dataset. The right view-window displays an axial section (a cross-section perpendicular to the longitudinal axis of the body), which acts as a preview of the higher resolution cryosection images in the dataset.

A horizontal indicator on each view indicates the position, on the body, of the cut shown in the other view. The user can vertically drag each indicator to sweep the cut, shown in the other view, through the body. The views provide smooth, rapid feedback (approximately 20 fps on a SUN SparcStation 1+) reflecting the cross-section at the sliding cut plane, resulting in a dynamic animated effect of motion through the body [4]. This allows the user to easily explore the contents of the entire dataset. When the user locates an axial section for which high resolution data is desired, the system downloads the corresponding full-size cryosection image from the NLM archive over the internet and displays it in a detail view.

[Screen Shot of Interface]
Figure 1: The Graphical Prototype

A TEXTUAL APPROACH

The second prototype utilizes users' knowledge of medical terminology of human anatomy. Users can interact with portions of the Medical Subject Headings (MeSH) Tree [2] in order to select and retrieve Visible Human images. MeSH, published by NLM, is a hierarchy of medical terms pertaining to anatomy, histology, pathology, etc. Dynamic interaction with the MeSH reveals inter-relationships between terms, such as spatial proximity of anatomical structures within the body.

The browser (Figure 2) displays the MeSH tree on the left side of the screen as a table of contents that can be dynamically expanded and contracted, level by level, via a slider widget. The children of any term can be viewed separately in subtree windows which appear in the middle of the screen. A multi-selection widget filters these terms by body region. Inter-relationships can be visualized by clicking on a term, causing related terms in other windows to be highlighted. Double-clicking on any term shows a range of thumbnail images described by that term. Clicking on a thumbnail will retrieve the corresponding full-size cryosection image from the Visible Human dataset.

[Screen Shot of Interface]
Figure 2: The Textual Prototype

For future work, we believe that combining these graphical and textual approaches would provide a rich comprehensive browsing environment.

ACKNOWLEDGMENTS

This research was supported by the National Library of Medicine Fellowship Program. Support was administered by the Oak Ridge Institute for Science and Education. We greatly appreciate the guidance and encouragement received from Catherine Plaisant and Ben Shneiderman.

The prototype interface software for the graphical approach is fully functional and freely available. For information, see http://www.cs.umd.edu/projects/hcil/Research/1995/vhe.html or anonymous ftp.cs.umd.edu in /pub/hcil/Demos/VHP.

REFERENCES

  1. National Library of Medicine Long Range Plan: Electronic Imaging. NIH Publication No. 90-2197, US Dept. of Health and Human Services (April 1990).
  2. National Library of Medicine, Bethesda, MD. MeSH -- Tree Structures, 1995.
  3. North, C., Shneiderman, B., Plaisant, C. User Controlled Overviews of an Image Library: A Case Study of the Visible Human. Proc. ACM Digital Libraries '96 Conf, ACM Press, 1996.
  4. Shneiderman, B. Dynamic Queries for Visual Information Seeking. IEEE Software, Nov 1994, 70-77. (Also available as CS-TR-3022, HCIL-TR-93-01.)


Browsing Anatomical Image Databases: A Case Study of the Visible Human / north@cs.umd.edu

Web Accessibility