- Research
- Open Access

# ImageParser: a tool for finite element generation from three-dimensional medical images

- HM Yin
^{1, 3}, - LZ Sun
^{1}Email author, - G Wang
^{2}, - T Yamada
^{2, 4}, - J Wang
^{2, 5}and - MW Vannier
^{2, 6}

**3**:31

https://doi.org/10.1186/1475-925X-3-31

© Yin et al; licensee BioMed Central Ltd. 2004

**Received:**07 July 2004**Accepted:**01 October 2004**Published:**01 October 2004

## Abstract

### Background

The finite element method (FEM) is a powerful mathematical tool to simulate and visualize the mechanical deformation of tissues and organs during medical examinations or interventions. It is yet a challenge to build up an FEM mesh directly from a volumetric image partially because the regions (or structures) of interest (ROIs) may be irregular and fuzzy.

### Methods

A software package, ImageParser, is developed to generate an FEM mesh from 3-D tomographic medical images. This software uses a semi-automatic method to detect ROIs from the context of image including neighboring tissues and organs, completes segmentation of different tissues, and meshes the organ into elements.

### Results

The ImageParser is shown to build up an FEM model for simulating the mechanical responses of the breast based on 3-D CT images. The breast is compressed by two plate paddles under an overall displacement as large as 20% of the initial distance between the paddles. The strain and tangential Young's modulus distributions are specified for the biomechanical analysis of breast tissues.

### Conclusion

The ImageParser can successfully exact the geometry of ROIs from a complex medical image and generate the FEM mesh with customer-defined segmentation information.

## Keywords

- Breast imaging
- image segmentation
- biomechanical analysis
- meshing
- finite element method (FEM)

## Background

Diagnostic imaging devices such as CT, MRI and PET scanners are able to produce three-dimensional (3-D) descriptions of various features such as tissues and organs. In a computer, these images are some data to describe the intensity at each spatial point of a volume. The interpretation of the dataset requires special training and depends on the experience. Researchers have introduced a variety of algorithms to visualize 3-D medical images, and to extract the geometric information of objects from volumetric image data [1–3].

In recent years, the finite element method (FEM) has widely been used to simulate the mechanical deformation of tissues and organs during examinations or interventions [4–6]. To build up an FEM mesh from a medical image, the contour information of segmented regions of interest (ROIs) need to be first extracted from a volume of data [7, 8]. Then, the volume is meshed into nodes and elements, and material properties are endowed to each element in accordance with the segmentation information [9]. By further applying the boundary conditions and mechanical loadings on the corresponding nodes or elements, commercial FEM software packages such as ANSYS and ABAQUS may calculate the mechanical stress and strain, and predict the deformation and motion in the field of view.

In this paper, a software package called the ImageParser is developed to generate an FEM model from 3-D medical images. While aiming at the imaging segmentation, mesh generation, and deformation simulation of heterogeneous breast tissues, the method is applicable to the many biomedical imaging and biomechanical analysis of soft/hard tissues such as mammography and cardiovascular imaging. This software uses a semi-automatic method to detect the objective constituents from the context of an image including neighboring tissues and organs. It segments an image based on customer-defined grayscale ranges, and meshes tissues into elements with a customer-defined size. Inputting the generated FEM mesh into an FEM program, we can calculate the mechanical deformation under specific boundary conditions and mechanical loadings. The ImageParser is written in Microsoft Visual C# .NET (Microsoft Development Environment 2003 Version 7.1), and can be integrated into a high-level image analysis environment with a good extendibility and scalability.

## Description

### Overview

After the mesh of the breast is generated, we can implement it into an FEM software package to simulate the mechanical deformation of the breast with given material constants of all the tissues and appropriate boundary conditions.

### Borderline Detection

- 1.
We first focus on one slice. When the first point (

*x*_{0},*y*_{0}) is selected close to the borderline, a function is then used to search the most possible border point in the square region with the left-top point (*x*_{0}-*s*,*y*_{0}-*s*) and the right-bottom point (*x*_{0}+*s*,*y*_{0}+*s*). Here*s*is a customer-defined parameter with a default value as 3 pixels. In this region, the gradient of each point Δ(*x*,*y*) is defined by a Laplace operator [12]

so that

*f*(

*x*,

*y*) denotes the intensity at (

*x*,

*y*). The detected point is the one with the different color from the background and with the maximum value of Δ(

*x*,

*y*)/[(

*x*-

*x*

_{0})

^{2}+ (

*y*-

*y*

_{0})

^{2}+

*ε*] where

*ε*is a customer-defined parameter with a default value as 0.1 to prevent the singularity on the point (

*x*

_{0},

*y*

_{0}). The detected point is denoted as (

*x*

_{1},

*y*

_{1}). It is noted that the Laplacian normalized by the distance of the point to the selected seed point is to make the neighboring points have a higher priority to be detected. At certain regions, the borderline may not be clear or two borderlines are close enough, in which cases the program will not get lost.

- 2.
We start to select the next point of the borderline. After we manually select one point visually close to the borderline following the method of detecting (

*x*_{1},*y*_{1}), the software can adjust the location of the point and detect the second point (*x*_{2},*y*_{2}) on the borderline. As seen in Figure 6, a function detects the borderline between (*x*_{1},*y*_{1}) and (*x*_{2},*y*_{2}). Two squares with edge length 2*s*are marked by the dark color in a big square having a diagonal line from (*x*_{1},*y*_{1}) to (*x*_{2},*y*_{2}). We find the most possible border point in the two dark squares by using the same method in the first step. If this new border point is in the right-top square, we replace (*x*_{2},*y*_{2}) by this point. Otherwise, we replace (*x*_{1},*y*_{1}) by the new border point. Once (*x*_{1},*y*_{1}) or (*x*_{2},*y*_{2}) is updated, we continue to find the next border point in the same way. Repeat this procedure until the distance between the two points is less than*2s*. Connecting all points in such an orderly way, we obtain the borderline. It is noted that this method is convergent because the distance between two working points becomes smaller and smaller in this procedure.

- 3.
We repeat step 2 until the borderline is closed. We thus obtain the whole closed borderline in the slice.

- 4.
For a 3-D medical image, due to the similarity of neighboring slices, the proposed software can map the selected key border points of the slice onto the neighboring slice and use the method in step 1 to find the corresponding border points in the new slice. After that we adopt step 2 to detect the borderline between the border points. In this way, we are able to detect the borderlines in all the slices. From these borderlines we can finally construct the surface of the selected ROI.

Because the borderlines of other slices are detected on the basis of the first slice, selection of this slice greatly affects the quality of results. We suggest that this slice need contain the most representative information. If the change of two neighboring slices is large, we can optionally reselect the border point in the new slice instead of detecting the borderline by the computer. It is further noted that because the borderlines detected by the computer may be very irregular, we can use a cubic Bezier curve fitting technique to smooth the borderlines.

### FEM mesh Generation

Among several methods to automating mesh generation [9], the mesh with cuboidal elements is the fastest and most stable method to mesh an organ with irregular shape even though it may require more elements at the boundary. We therefore apply the cuboidal-element mesh to make this software applicable for complex cases. For instance, in Figure 3 the cloud-like parenchyma is dispersed in the fatty tissue. It is almost impossible to extract the exact geometry of parenchyma. In this case, most geometry-based methods are invalid. In the cuboidal mesh, elements are generated layer by layer and are automatically connected through the overlaid nodes. Given an element size, we can calculate how many slices each layer of elements spans. For simplicity, we assume the borderline of the central slice to be the borderline of that layer. Since the borderline consists of many points, we first build up a grid using the element size, move each border point to the closest cross-point of the grid, and remove the repeated points. We thus obtain the borderline denoted by the cross-points of the grid. It is noted the borderline may be entangled somewhere due to the numerical truncation. We need normalize the borderline so that it encloses a single-connected region and the distance between two neighboring point is equal to the size of the elements.

When we scan the single connected region, there exist two types of points on the borderline: jumping points and inertial points. If the left and right sides of a point are in the different states; i.e. one side is in the inside of the objective region and the other side is in the outside, then the point is called a jumping point. Otherwise, this point is called an inertial point. For instance, in an upstanding rectangle, all points on the left and right sides are jumping points, whereas the rest points on the top and bottom sides are inertial ones. On a closed borderline, each point is connected to two points. For a jumping point, the two neighboring points apparently have different values of *y* coordinate, whereas those for an inertial point do not. From this criterion we can identify the jumping points on the borderline. Once the points of the borderline are given, we can sort the points from top to bottom by *y* coordinate and from left to right by *x* coordinate. Then, for any *y* coordinate we can obtain a list of points with increasing *x* coordinates. During a horizontal scan for a fixed *y* coordinate, the number of jumping points in this list must be even, with which we obtain the pair-wise jumping points and find all the internal points between each pair of jumping points. Scanning the points from top to bottom, we can obtain all the internal points for the connected region. Then we can obtain the cuboidal-element mesh for this layer by mapping one point onto one element.

Because different elements may have different material properties, we need find the segmentation information for each element. From the grayscale histogram, we have defined the grayscale ranges corresponding to the tissues. Typically, an element may contain many voxels that belong to different tissues, whereas the FEM requires the element to be homogeneous. We count the number of voxels for each tissue in the element and assume the maximum one to be the material of the element. Thus, we can map the elements onto the different tissues as shown in Figure 5. We can thus mesh the object layer by layer and finally obtain the total FEM mesh, from which we can further calculate the volume of each tissue.

The surface information of the object is important for applying boundary conditions and mechanical loadings. This software uses the 2-D rectangular elements to describe the surface. Each 3-D cuboidal element has six rectangular faces. We collect the faces from all cuboidal elements. Thus, for an object containing *N* cuboidal elements, we can obtain 6*N* 2-D rectangular elements. Obviously not all the rectangular elements are on the surface of the object. If an element is not on the surface, from the connectivity, another 2-D element containing the same nodes must exist which belongs to the neighboring cuboidal element. Eliminating each pair of these inside elements, we are able to obtain surface elements. We further input the mesh with segmentation information into an FEM program based on the required data format, assign material properties to tissues, and apply boundary conditions on the surface nodes. We can eventually calculate the mechanical deformation, internal stress and strain by the FEM software.

## Results and Discussion

### 3-D FEM Mesh and Material Properties

*mm*

^{3}. As an example, the 148

^{th}slice is shown in Figure 1. The breast includes three kinds of tissues: fat, parenchyma, and tumor, which are represented by three grayscales as dark, gray, and light, respectively. Using the ImageParser package, we are able to mesh the breast by cuboidal elements with a size of 2.8125 × 2.8125 × 3

*mm*

^{3}. The breast is meshed into 14,902 elements with 18,486 nodes as shown in Figure 7. The tumor, parenchyma, and fatty tissue consist of 154, 5783, and 8965 elements, respectively. The surface of the breast includes 6,900 rectangular 2-D surface elements. The region of breast is defined as follows: 0 <

*x*< 84.375

*mm*, 0 <

*y*< 87.1875

*mm*and 0 <

*z*< 135

*mm*. Here

*x*is from left to right in a slice of the image,

*y*is from the top to bottom, and

*z*is from the first slice to the last slice. Corresponding to the human body,

*y*represents the normal direction of the coronal plane, while

*z*signifies the normal direction of the axial (transverse) plane (Figure 7).

Based on Krouskop et al. [13], the initial elastic moduli of three tissues are taken as 20 KPa for fat, 35 KPa for parenchyma, and 100 KPa for tumor. Because these tissues may undergo large (finite) deformation, we apply the Mooney-Rivlin nonlinear elastic (hyperelastic) model to describe the constitutive law for the finite deformation. Using the initial elastic moduli we can calculate the Mooney-Rivlin material constants as: *C*
_{01} = 1,333 Pa, *C*
_{10} = 2,000 Pa for fat; *C*
_{01} = 2333.3 Pa, *C*
_{10} = 3500 Pa for parenchyma; and *C*
_{01} = 6,667 Pa, *C*
_{10} = 10,000 Pa for tumor. It is noted that, due to the nonlinear characteristics, the elastic modulus for each tissue change as a function of deformation.

### FEM Modeling by ANSYS

ANSYS 7.0 [14] is the commercial nonlinear FEM software. We input the nodes and elements into ANSYS, and define the material models for three tissues. The applied compression with two flat-paddles is designed to simulate the clinical mammography examination. The ANSYS elastic contact model is adopted for the interaction between the breast tissue and the much more rigid paddle. The paddle's Young's modulus and Poisson's ratio are taken as 210 GPa and 0.3, respectively. During the compression process, the breast deforms. The contact area between the breast surface and the paddle increases automatically. The friction coefficient between the breast and the paddle is assumed to be 0.2. The boundary conditions are assumed that all nodes attached to thorax are constraint as *U*
_{
x
}= *U*
_{
y
}= 0, so that they can only move in the *z* direction for computational convenience. The two paddles move toward each other with a quasi-static strain rate. The maximum paddle movement is limited to be 13.5 *mm* (20% deformation) in the *z* direction.

### FEM Results

where *σ*
_{
ε
}and *ε*
_{
e
}are the von Mises stress and strain, respectively. It is noted that, for linear elastic material, *E* is always a material constant. Because the material properties of skin, normal entity and tumor are all nonlinear, *E* should change during the process of compression. Figure 8(a) shows the von Mises strain for the sections at *z* = 69,78,87 *mm*. The strain around the thorax is quite significant due to the boundary constraint, whereas the strain close to skin is small because of the free boundary condition. In the region of tumor, it is shown that the strain is much smaller than that in the neighboring region because the tumor is much harder than other tissues. Figure 8(b) demonstrates that the tangential Young's modulus is no longer uniform even in the same tissue because its strain field is not uniform.

## Conclusions

The ImageParser system has been developed to create FEM mesh models from 3-D medical images. A semi-automatic method has been proposed to detect the ROIs from the context of complex image structures. The ROIs can be meshed into cuboidal elements and segmented based on the grayscale of the voxels. It has been demonstrated that, through a 3-D CT image volume of the woman breast, the ImageParser can effectively mesh the breast into cuboidal elements, and simulate the realistic nonlinear deformation responses of the breast tissues upon compression.

## Declarations

### Acknowledgements

This work is sponsored by the University of Iowa's Iowa Informatics Initiative.

## Authors’ Affiliations

## References

- Lorensen WE, Cline HE:
**Marching cubes: a high resolution 3D surface construction algorithm.***Comput Graph*1987,**21:**163–169.View ArticleGoogle Scholar - Gueziec A, Hummel R:
**Exploiting triangulated surface extraction using tetrahedral decomposition.***IEEE Trans Vis Comput Graph*1995,**1:**328–342. 10.1109/2945.485620View ArticleGoogle Scholar - Wu Z, Sullivan JM:
**Multiple material marching cubes algorithm.***Inter J Numer Meth Eng*2003,**58:**189–207. 10.1002/nme.775View ArticleGoogle Scholar - Maurel W, Wu Y, Magnenat Thalmann N, Thalmann D:
*Biomechanical Models for Soft Tissue Simulation*. Berlin, Springer-Verlag; 1998.View ArticleGoogle Scholar - Samani A, Bishop J, Yaffe MJ, Plewes DB:
**Biomechanical 3-D finite element modelling of the human breast using MRI data.***IEEE Trans Med Imaging*2001,**20:**877–885. 10.1109/42.952726View ArticleGoogle Scholar - Azar FS, Metaxas DN, Schnall MD:
**Methods for modelling predicting mechanical deformations of the breast under external perturbations.***Med Image Anal*2002,**6:**1–27. 10.1016/S1361-8415(01)00053-6View ArticleGoogle Scholar - Kang Y, Engelke K, Kalender WA:
**Interactive 2D editing tools for image segmentation.***Med Image Anal*2004,**8:**35–46. 10.1016/j.media.2003.07.002View ArticleGoogle Scholar - Viceconti M, Zannoni C, Pierotti L:
**TRI2SOLID: an application of reverse engineering methods to the creation of CAD models of bone segments.***Comput Meth Prog Bio*1998,**56:**211–220. 10.1016/S0169-2607(98)00011-XView ArticleGoogle Scholar - Viceconti M, Bellingeri L, Cristofolini L, Toni A:
**A comparative study on different methods of automatic mesh generation of human femurs.***Med Eng Phys*1998,**20:**1–10. 10.1016/S1350-4533(97)00049-0View ArticleGoogle Scholar - Ballard DH, Brown CM:
*Computer Vision*. Englewood Cliffs: Prentice-Hall; 1982.Google Scholar - Sonka M, Hlavac V, Boyle R:
*Image Processing, Analysis and Machine Vision*. London: Chapman & Hall; 1993.View ArticleGoogle Scholar - Jähne B:
*Digital Image Processing*. Berlin: Springer-Verlag; 1991.Google Scholar - Krouskop TA, Wheeler TM, Kallel F, Garra BS, Hall T:
**Elastic moduli of breast and prostate tissues under compression.***Ultrasonic Imaging*1998,**20:**260–274.View ArticleGoogle Scholar - ANSYS is a commercial FEM software package developed by ANSYS, Inc Google Scholar

## Copyright

This article is published under license to BioMed Central Ltd. This is an open-access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.