For years, a central focus of the Center for Integrative Biomedical Computing has been software dissemination. Over the Center's twelve-year history, the infrastructure of CIBC's software dissemination efforts have undergone a radical evolution, with many lessons learned for both users and developers.
Software Dissemination as a Form of Science and Technology Dissemination
The SCIRun/BioPSE Problem-Solving Environment. This system was designed as a 'computational workbench' and represented a new approach to bringing high-end computational tools to the biomedical researcher.
From the outset, the Center's leadership believed that along with the typical avenues of dissemination, publications, seminars, workshops, conference presentations, etc.‚ software dissemination held a particularly high potential as a means to disseminate the knowledge and advances of the Center and its collaborators. This fundamental belief led to experimentation in a variety of topics, such as software licensing, open source repositories, make systems, operating systems, source code/binary releases, research code versus releasable code, software support, and release schedules.
The origins of our success in developing widely used software tools lie in a set of strategies for algorithm research and software development. One such strategy is the production of software tools with low barriers to entry. This entails the release of documented, tested, complete applications that do not require learning new programming languages or complex, architecture-specific build environments. We also continue to follow an initiative to create a suite of lightweight, stand-alone applications, directed at specific tasks of common interest across a wide set of disciplines. The result is a set of programs, such as Seg3D, with large and growing user bases.
Today's Tools - Tomorrow's Scientists: An ImageVis3D Project
The SCI Institute holds a strong belief that providing research opportunities to undergraduate and even high school students will not only encourage them to pursue studies in the sciences, but also give them a head start in their future academic lives. Being allowed to work side by side with PhD-level scientists within a real research institute moves science from something that happens in a text book or highly structured laboratory to the dynamic work environment shared by scientists around the world.
In this year's high school summer intern program, the SCI Institute invited four students, one each from Juan Diego Catholic High School, The Waterford School, and two from West High School. These students were given the opportunity to work with a lead software developer from the National Institutes of Health (NIH) sponsored Center for Integrative Biomedical Computing (CIBC). Their task seemed simple: take Seg3D and ImageVis3D (two advanced software tools developed by the CIBC), find a dataset of interest to the student, load that data, and experiment with the software on both desktop and iPad versions. And then, present your results to your high school peers. In the end, the students learned that research is a full-contact sport, not just a homework assignment. They had to 'dig-in', expand their knowledge, and learn about their subjects of interest, their data, their software, even their computers. In the end, the students translated this process and knowledge to science classes at their school. And, the top question after the presentations? Oddly enough, "how do I get an internship like yours?" Kids excited about a science internship! Mission Accomplished.
Solving Mysteries of Autism via The Power of Collaboration
Dr. Guido Gerig Early-Brain Development Research Reveals Vibrant Clues
By Peta Owens-Liston
Dr. Guido Gerig
The glossy whiteboards that line the walls in offices, lounge areas, and conference rooms are one of the first things Guido Gerig, PhD, noted when the University of Utah's Scientific Computing and Imaging (SCI) Institute first began courting him to join their team in 2007. For someone so prominent worldwide for his image analysis expertise and seminal research, whiteboards seemed the simplest of visual tools. Yet, what these signaled to Gerig was that this place fostered collaboration among students, postdocs, and faculty; these ubiquitous boards were an immediate means to visually improve understanding and share knowledge.
Pencil in hand, Gerig fills three pages with a whirl of sketches as he explains how his imaging work illuminates clinical findings in his research involving early brain development, and more specifically autism. The sketches fade to stick figure-status as Gerig jumps back and forth between the paper and the color-exploding images on his computer screen. Vivid and seemingly pulsating with life, the brain-development images are a result of thousands of highly precise, quantifiable measurements never before captured visually.
Mobile Mayhem: Researchers Harness Kraken to Model Explosions via Transport
by Gregory Scott Jones - NICS
The crater resulting from the Spanish Fork Detonation.
First, the bad news: all across America, trucks and tractor-trailers are transporting industrial explosives on nearly every artery of the country's interstate and highway system. That's right, volatile explosives, including munitions, rocket motors, and dynamite, are moving at a high rate of speed down a roadway not too far from you.
Now, the good news: America's track record in transporting these materials is about as safe as they come. Very rarely, almost never in fact, are the potential dangers of these transports realized, largely due to instituted safeguards that seem to work very well.
However, accidents can happen. Take the August 2005 incident in Spanish Fork Canyon, Utah, for instance. A truck carrying 35,500 pounds of explosives—specifically small boosters used in seismic testing—overturned and exploded, creating a crater in the highway estimated to be between 20 to 35 feet deep and 70 feet wide according to the Utah Department of Transportation. But the damage wasn't solely financial. Four people, including the truck driver and a passenger, were hospitalized.
Meshing for Multimaterial Biological Volumes: BioMesh3D
Figure from R.S. MacLeod, et al., Subjectspecific, multiscale simulation of electrophysiology: a software pipeline for image-based models and application examples. Example A particle system provides adaptive sampling the various material boundaries of a segmented CT volume from a human torso.
Over the past year the CIBC, in partnership with our collaborators, has begun to introduce a generalized image-processing pipeline and associated software to the biomedical community.
With the widespread use of medical imaging, there is a growing need for better analysis of datasets. One method for improving analysis is to simulate biological processes and medical interventions in silico, in order to render better predictions. For example, the CIBC center is currently collaborating with Dr. Triedman at Children's Hospital in Boston to develop a computer model that will help guide the implantation of Implantable Cardiac Defibrillators (ICDs). This model uses pediatric imaging to select placement of electrode leads to generate the optimal field for defibrillation. One of the critical pieces in the development of the model is the generation of quality meshes for electric field simulation. Because the project is entering the validation phase where many cases need to be reviewed, a robust and automated Meshing Pipeline is required.
Atrial fibrillation (AF) is an electrophysiological condition that represents an increasing problem in the aging populations of the world; AF doubles the risk of stroke and mortality and diminishes quality of life. The best current method to evaluate the progression of AF and monitor the success of interventions is via an invasive intra-cardiac catheter-based electrical mapping procedure. A noninvasive means to evaluate characteristics of AF prior to treatment and to track the effect of interventions over time would be extremely valuable, and magnetic resonance imaging (MRI) offers such an opportunity. Before MRI can achieve its potential, there are challenging technical problems to overcome, such as the high spatial resolution required to image the thin atrial wall and the temporal resolution and gating necessary to compensate for the distorting effects of respiratory and cardiac motion. The Comprehensive Arrhythmia Research and Management (CARMA) Center has become a world leader in the use of MRI in AF and has overcome many of the image acquisition hurdles to make MRI a standard component of AF patient management at our institution. These improvements in image acquisition have opened up significant opportunities and new questions for the understanding and clinical management of AF.
An Isosurface visualization of a magnetic resonance imaging data set (in orange) surrounded by a volume rendered region of low opacity (in green) to indicate uncertainty in surface position.
The estimation and visualization of uncertainty information is an important research problem in both simulation and visualization. Uncertainty is a term used to describe the error, confidence, and variation of a dataset in order to allow a scientist to understand the accuracy not only of the data but also of the processes used to explore the data. One such technique, sensitivity analysis, helps the scientist to understand the effects of perturbing parameters of a function. Small perturbations of the input parameters that create large perturbations in the output results can indicate areas of the function that are highly dependent on the input parameters and may be interpreted as unstable or possibly wrong. Sensitivity analysis techniques can be used not only to explore the mathematical models used to generate uncertainty data but also to better understand the effects of input parameters of visualization techniques. Uncertainty data generated from the analysis of a mathematical model reconstructing biological experiment have been a focus of the CIBC team.
Atrial fibrillation (AF) is the most common—and perhaps most insidious—form of heart rhythm disturbance and treating it has become the focus of a group of bioengineers, imaging physicists, and physicians at the University of Utah.
In atrial fibrillation, the upper two chambers (the left and right atria) of the heart lose their synchronization and beat erratically and inefficiently. The same condition in the lower chambers (ventricles) of the heart is fatal within minutes and defibrillators are necessary to restore coordination. In the atria, death is by stealth and occurs over years, which is both good news and bad.
Because it is not immediately fatal, there is time to treat atrial fibrillation–but also time to ignore it. While it is not immediately life-threatening, AF does immediately reduce the pumping capacity of the heart and elevates the heart rate of the entire organ. Patients cannot be as physically active as they often wish but many adjust to the symptoms and live with the disease untreated for many years.
Figure from T. Fogal and J. Krüger, a Clearview rendering of the visible human male dataset
Attempting to display the entirety of a large volumetric dataset at one time would result in an overwhelming amount of information. Furthermore, visualization tools based on volume rendering present the user with a host of confusing options. We present ClearView, which provides a simplified volume visualization tool with a focus on doing what matters most: looking at your data. Users frequently want to direct the viewer's attention to a particular region of their volumes. With many volume rendering tools, this means setting up complex transfer functions to highlight the region of interest, with the unfortunate side effect of potentially affecting the larger image. ClearView allows the user to focus their visualization efforts on the area of their choice, while separating parameters for visualizing of surrounding data. This provides not only a simplified user interface, but finer-grained control over the final publication-quality visualization. Through advanced GPU rendering techniques, ClearView presents all of this to the user at highly interactive frame rates.
Simulation of Electric Stimulation for Bone Growth
Figure from B.M. Issacson, et al., A unilateral hierarchical model was assembled as a representative image consisting of skin (purple) adipose tissue (yellow), musculature (pink), bone (blue), bone marrow (orange), and internal organs (green) (a). Each tissue type was assigned a specific conductivity using SCIRun. A large serpentine-like mass of HO was identified in the medial aspect of the residual limb, and was demonstrated in more detail in an axial cross section of the affected limb (b).
Osseointegration is a surgical procedure that provides direct skeletal attachment between an implant and host tissue with proven success in dental, auricle, and transfemoral implants. However, one challenge with using natural biological fixation is attaining a strong skeletal interlock at the implant interface, a prerequisite for long-term implant function. Utilizing metallic implants as a means of biological fixation has been the objective of orthopedic surgeons over the past two centuries. However, controlling osteogenesis at the implant interface, which is essential for providing strong skeletal fixation, remains challenging. Regulated electrical stimulation has proven effective in fracture healing and non-traumatized bone models, but has not been investigated in a percutaneous osseointegrated implant system. One advantage of the veteran patient population is that an orthopedic implant protrudes from the residual limb functioning as an exoprosthesis attachment and may operate as a potential cathode for an external electrical stimulation device.
Subject Specific, Multiscale Simulation of Electrophysiology
A "typical" workflow that applies to many problems in biomedical simulation contains the following elements:
(i) Image acquisition and processing for a tissue, organ or region of interest (imaging and image processing),
(ii) Identification of structures, tissues, cells or organelles within the images(image processing and segmentation),
(iii) Fitting of geometric surfaces to the boundaries between structures and regions (geometric modelling),
(iv) Generation of three-dimensional volume mesh from hexahedra or tetrahedra (meshing), and
(v) Application of tissue parameters and boundary conditions and computation of spatial distribution of scalar, vector or tensor quantities of interest (simulation).
Over the past year the CIBC, in partnership with our collaborators, has begun to introduce a generalized processing pipeline and associated software to the biomedical community. This work has been largely influenced by DBP collaborators such as those collaborating to develop optimization strategies for ICD placement in children; Dr. John Triedman at the Department of Cardiology, Children's Hospital Boston, Dr. Matthew Jolley, Stanford University Medical Center. and Drs. Elizabeth Saarel, Tom Pilcher, and Michael Puchalski, all from the Department of Cardiology at Primary Childrens' Hospital in Salt Lake City. Additionally, collaborative work with the goal of the making osseointegrated amputee implants part of an electrical system to accelerate skeletal attachment also influenced the creation of the pipeline described below; collaborators are Brad Isaacson, Dr. Joseph Webster, Dr. James Beck, and Dr. Roy Bloebaum from the Department of Veteran Affairs and University of Utah.
Despite great advances in neuroscience and medical technology in recent decades, nearly ten million Americans still suffer blindness due to retinal degenerative diseases such as retinitis pigmentosa (RP), age-related macular degeneration (AMD), diabetic retinopathy, and glaucoma. Unfortunately, current treatments available for these conditions are still quite limited. A primary challenge to developing effective treatments is the need for a complete understanding of the highly complex and delicate systems that compose the retina and how those systems change in response to degenerative disorders.
Remodeling processes that occur in the neuronal pathways within the retina during the course of retinal deterioration are of particular importance to the development of treatments for these conditions. Researchers at the Robert E. Marc Laboratory at the Moran Eye Center are collaborating with the SCI Institute on a project supported by the NIH-NIBIB (grant number 5R01EB005832) to develop high-throughput techniques for reconstructing and visualizing the neural structures that compose the retina in order to meet these challenges.
Developing the Next Generation Tools for Preoperative Planning for Implantable Cardiac Defibrillators
BioPSE visualizing the electrical field generated by an ICD device.
One of the exciting projects in which the SCI Institute is collaborating is the development of new tools and techniques to assist doctors optimize the placement of Implantable Cardiac Defibrillators (ICDs) prior to surgery.
The use of ICDs has greatly increased over the last few years due to their efficacy in preventing sudden cardiac death (SCD) in patients with congenital heart defects or heart disease. These devices work by continually monitoring the rhythm of the patient's heart and immediately delivering a corrective electric shock if a life-threatening tachycardia is detected. Through this innovation, thousands of lives are saved each year. Surprisingly, these devices are sometimes implanted in newborns and older children with congenital heart defects. Pediatric patients present a particular challenge to the surgeons planning an implantation due to the wide variety of shapes and sizes of torsos. It often has proven difficult for physicians to determine the ideal placement and orientation of the electrodes prior to surgery. Accurate placement of the electrodes is crucial to ensure successful defibrillation with a minimum amount of electric current and to minimize potential damage to the heart and the surrounding tissues.
In 1909 the Broadman areas map was published which localized various functions of cortex.
Since the beginning of the 20th century, scientists have used brain models to study the anatomy and function of the human brain. It was discovered that each part of the brain's anatomy has specific functions for the mind and body. Over the years increasingly sophisticated atlases, or maps of brain anatomy and function, have been developed to help build our understanding of the human brain and guide scientists and physicians studying individual brains.
Although we've identified many structures and characteristics that are common in all human brains, in reality every brain is different and we need to improve our understanding of how brains vary between individuals. One problem that persists is that most current atlases have been based on arbitrarily chosen individuals. Even today, when intense research has been directed toward the development of digital three-dimensional atlases of the brain, most digital brain atlases so far have been based on a single subject's anatomy. This introduces a bias into the analysis when comparing individual brains to the atlas and does not provide a meaningful baseline with which to measure individual anatomical variation.
Modern problem solving environments such as SCIRun have provided scientists with the essential tools for composing complex simulations and visualizations of large scale data. The dataflow programming paradigm has made what was once a rather daunting programming endeavor into a relatively simple point and click process. By condensing the programs into nice little modules that could be visually strung together with pipes, scientists no longer had to worry about the computer programming under the hood. As the new paradigm opened the door to new possibilities and allowed them to explore higher levels of complexity and larger datasets, however, the dataflow networks themselves became rather large and complicated. Scientists now require even greater levels flexibility and organization in dataflow management. For example, it is often necessary for them to construct multiple simulation-visualization scenarios to compare the possibilities and develop new insights.
Figure 1: The VisTrails Visualization Spreadsheet. Surface salinity variation at the mouth of the Columbia River over the period of a day. The green regions represent the fresh-water discharge of the river into the ocean. A single vistrail specification is used to construct this ensemble. Each cell corresponds to a single visualization pipeline specification executed with a different timestamp value.
Angiogenesis, or the formation of new blood vessels, is a critical part of tissue growth and healing processes. Understanding the underlying mechanisms of this process and how it affects the perfused tissues is fundamental to many issues in medicine. For example, control of angiogenesis could help bones and tissues to heal faster or with better results. In contrast, the inhibition of angiogenesis in some cases could stop the development unwanted tissues such as a tumor, or slow the healing process for better results. Controlled angiogenesis is also important for engineering of new tissues or possibly whole organs needed to replace damaged ones. Researchers in the Musculoskeletal Research Laboratories (MRL) at the SCI Institute are investigating the angiogenesis process and how it effects tissue development. They are also developing computer models to accurately simulate this process and predict the effects of angiogenesis on the mechanical properties of tissues.
New Method Can Identify What Genes Do, Test Drugs' Safety
May 4, 2006 -- Utah and Texas researchers combined miniature medical CT scans with high-tech computer methods to produce detailed three-dimensional images of mouse embryos – an efficient new method to test the safety of medicines and learn how mutant genes cause birth defects or cancer.
"Our method provides a fast, high-quality and inexpensive way to visually explore the 3-D internal structure of mouse embryos so scientists can more easily and quickly see the effects of a genetic defect or chemical damage,” says Chris Johnson, a distinguished professor of computer science at the University of Utah.
A study reporting development of the new method – known as “microCT-based virtual histology” – was published recently in PLoS Genetics, an online journal of the Public Library of Science.
The study was led by Charles Keller, a pediatric cancer specialist who formerly worked as a postdoctoral fellow in the laboratory of University of Utah geneticist Mario Capecchi. Keller now is an assistant professor at the Children's Cancer Research Institute at the University of Texas Health Science Center in San Antonio.
The Center for Interactive Ray-Tracing and Photo Realistic Visualization
The SCI Institute is honored to host a new Center of Excellence for Interactive Ray-Tracing and Photo Realistic Visualization dedicated to the development of tools for interactively visualizing large-scale datasets on the fly with advanced lighting and material models that help users understand the subtle detail in high-fidelity datasets. The Center of Excellence program is funded by the state of Utah to accelerate the commercialization of promising technologies developed at the state's Universities.
Almost every modern computer comes with a graphics processing unit (GPU) that implements an object-based graphics algorithm for fast 3-D graphics. The object-based algorithm in these chips was developed at the University of Utah in the 1970s. While these chips are extremely effective for video games and the visualization of moderately sized models, they cannot interactively display many of the large models that arise in computer-aided design, film animation, and scientific visualization. Researchers at the University of Utah have demonstrated that image-based ray tracing algorithms are more suited for such large-scale applications. A substantial code base has been developed in the form of two ray tracing programs. The new Center aims to improve and integrate these programs to make them appropriate for commercial use.
Scientific visualization as currently understood and practiced is still a relatively new discipline. As a result, we visualization researchers are not necessarily accustomed to undertaking the sorts of self-examinations that other scientists routinely undergo in relation to their work. Yet if we are to create a disciplinary culture focused on matters of real scientific importance and committed to real progress, it is essential that we ask ourselves hard questions on an ongoing basis. What are the most important research issues facing us? What underlying assumptions need to be challenged and perhaps abandoned? What practices need to be reviewed? In this article, I attempt to start a discussion of these issues by proposing a list of top research problems and issues in scientific visualization. [PDF version]
CT Segmentation and 3-D Reconstruction with Morphometric Analysis for Evaluation of Occipital-Cervical Instability in Children
Lindsey J. Healy
Down syndrome is a common chromosomal disorder that affects 0.15 percent of the total population. Individuals with Down syndrome are prone to spinal instability due to congenital abnormalities in the occipital-cervical (O-C1) joint near the base of the skull. To determine the possible abnormality causing this instability, an image-processing pipeline was created by combining several available software packages and custom made software. Patients with Down syndrome and spinal instability at the O-C1 joint were age-matched with controls. The subject data was assessed and a congenital abnormality was defined in the superior articular facets of C1 in the Down syndrome patients. The software developed helped to visualize the abnormality and could be used in a clinical setting to help aide in the diagnosis and screening for spinal instability in Down syndrome patients.