Abstract
Bone age is one of the most important indicators for assessing bone’s maturity, which can help to interpret human’s growth development level and potential progress. In the clinical practice, bone age assessment (BAA) of X-ray images requires the joint consideration of the appearance and location information of hand bones. These kinds of information can be effectively captured by the relation of different anatomical parts of hand bone. Recently developed methods differ mostly in how they model the part relation and choose useful parts for BAA. However, these methods neglect the mining of relationship among different parts, which can help to improve the assessment accuracy. In this paper, we propose a novel part relation module, which accurately discovers the underlying concurrency of parts by using multi-scale context information of deep learning feature representation. Furthermore, based on the part relation, we explore a new part selection module, which comprehensively measures the importance of parts and select the top ranking parts for assisting BAA. We jointly train our part relation and selection modules in an end-to-end way, achieving state-of-the-art performance on the public RSNA 2017 Pediatric Bone Age benchmark dataset and outperforming other competitive methods by a significant margin.
You have full access to this open access chapter, Download conference paper PDF
Similar content being viewed by others
1 Introduction
Bone age assessment (BAA) requires to interpret the maturation of bone, playing an important role in understanding the growth of human. It is utilized in an array of scenarios, such as the diagnosis and treatment of disorder of body. Generally, in clinical diagnosis, radiologists estimate the bone age by using the Graulich-Pyle (G-P) [5] and Tanner-Whitehouse (T-W) methods [12], which rely on expertise at the cost of tremendous time for observing each sample. However, due to the complex pattern of bones, different experts may provide various observations, easily leading to problematic judgements for the down-stream diagnosis and treatment. Thus, recent methods [3, 13] incorporate more effective computer-aided system to assist BAA. For example, the BoneXpert [13] diagnosis system has been approved and applied in various countries. Yet, these systems require too expensive process to generate high-quality images for judging bone ages.
The latest methods [1, 6, 8, 11, 14] borrow the success of deep networks in medical image analysis [2, 4], and apply deep learning framework in BAA. To address the large variation of bones with respect to positions, sizes and shapes, Pan et al. [6] and Iglovikov et al. [8] partition a bone structure into parts, and select a single part as input to train network individually. It leads to a more focused learning of invariant features, which are discriminative to bone ages. Wu et al. [14] uses the attention module to detect key parts of the hand bone. Bae et al. [1], Larson et al. [9] and Spampinato et al. [11] further extract features on multiple parts to yield richer information. However, the previous methods neglect the relationship between parts of a bone structure, which is important to select useful parts for BAA.
In this paper, we advocate the idea of building relationship between parts of the hand bone and selecting discriminative parts for BAA. Given the hand bone in X-ray images, we propose a Part Relation Module which connects strong-correlated parts. It enables the direct communication between parts, embedding more effective context information in part representations. Based on the part relationship, we employ a Part Selection Module to harness useful parts for the final estimation of bone age. Note that the part selection is done by self-supervision, without the requirement of heavy labelling effort. More importantly, the relation and selection modules form an end-to-end framework, jointly distilling the features for BAA. This enables our approach to provide more details, i.e., part relation map and part selection results, which are critical to medical analysis. Our method achieves state-of-the-art result on the public benchmark, i.e., RNSA pediatric bone age dataset. It demonstrates the effectiveness of our approach.
2 Part Relation Module
Compared to individual parts of the hand bone, part relation provides richer information for understanding the importance of different parts, and yields more accurate information for BAA. But recent methods use independent parts to learn assessment models, which are oblivious to part relations. The latest works compute global representation on the whole image, and implicitly models part relation. However, the whole image is insensitive to positions, sizes and shapes of parts, making it difficult to construct effective representations.
In this paper, we propose the part relation module to discover the useful relationship between parts for BAA. As illustrated in Fig. 1(a), given an X-ray image, we use a CNN to compute different levels of convolutional feature maps (see the blue blocks). Each feature map is used to compute a corresponding relation map for modelling part relation. In the relation map, we produce high responses for strongly-correlated parts of the corresponding hand bone, while suppressing irrelevant regions. At each level, we use the convolutional feature map and the associated relation map to produce the context representation of correlated parts. We use all levels of context representations to provide more detailed information, for precisely scoring parts that have variant spatial and appearance properties.
The architecture of the part relation module. At each level, the convolutional feature map of backbone network undergoes different convolutional operations to yield relation map, trunk map and residual map. The new feature maps are used to compute the context representation, which is fed to estimate anchor scores in the part selection module (see Fig. 3).
More formally, we use a backbone ResNet-34 [7] to produce a set of convolutional feature maps \(\{X_i\}\). The feature map \(X_i\) is output by the residual block at the \(i^{th}\) level. As shown in Fig. 2, we apply different \(1 \times 1\) convolutional operations on the feature map \(X_i\) to produce the relation map \(R_i\), trunk map \(T_i\) and residual map \(D_i\), respectively. Then we compute the context representation of correlated parts as:
where \(\sigma \) is sigmoid activation function. In Eq. (1), we apply the sigmoid activation function to the relation map \(R_i\), yielding an activation map. The new activation map plays as an information adaptor. It scores the part information contained in the trunk map \(T_i\), by respecting the importance of parts. To avoid missing part information during the scoring process, we combine the residual map \(D_i\) with the scored part information to form the context representation \(F_i\). As shown in Fig. 2, \(F_i\) is fed to the next level for producing the context representation \(F_{i+1}\). Using Eq. 1, we compute context representations \(\{F_i\}\) at all levels. Each level of context representation is used for scoring parts for the further selection process, as we will describe below.
3 Part Selection Module
The discriminative parts of hand bone are of importance to accurate estimation of bone age. Generally, recent methods use single or all parts for BAA, without understanding the part relation. As the relation model provide the importance of parts, the existing methods are incapable of selecting useful parts, which limits the performance of BAA. In this section, we advocate the idea of using the part relation to select useful parts.
We propose a part selection module to select the important parts, which are used to construct the representation for BAA. Given an X-ray image, our part selection module employs a set of anchors \(\{A_n\}\) to represent the part candidates for selection. Note that anchors have different sizes and ratios, and span over the entire image to cover potential parts, as illustrated in Fig. 3.
Next, we conduct a scoring process on each anchor, as illustrated in Fig. 1(b). This is done by using different levels of context representations, which are provided by the part relation module, to regress a set of scores \(\{S_n\}\) for all anchors. Here, we resort to the convention of object detection [10] and use lower/higher levels of representation to compute scores for smaller/larger anchors. Given the scores for all anchors, we sort the set of scores at all levels, then we employ the non-maximum suppression to eliminating overlapping anchors. and select top-M anchors having highest scores in the image. By following the top-M order of the selected anchors, we crop the corresponding image regions from the X-ray image. These image regions are resized uniformly, and each region is fed to another ResNet-34 for producing a feature vector. We sum all regions’ feature vectors as a local feature. Together, we concatenate the local feature \(F_{local}\), global feature \(F_{global}\) (i.e., the highest level of context representation) and gender feature \(F_{gender}\) (i.e., a one-hot vector) for predicting the bone age (see Fig. 1(c)).
To jointly train the part relation and selection modules, we employ the ranking loss [15] that measures the quality of part relation and selection. As illustrated in Fig. 4, given the selected parts, we associate them with anchor scores. We use each selected part to estimate a bone age, which is compared to the ground-truth age for computing a confidence. Here, the intuition is that the useful part, which has higher anchor score, contributes to the prediction of bone age that is similar to the ground-truth. Thus, the ranking loss for part relation and selection modules is computed as:
where we denote \(S_i\) as the anchor score for the \(i^{th}\) selected part. Given the \(i^{th}\) selected part, \(C_i\) is the confidence formulated as:
where \(\sigma \) also means the sigmoid activation function. \(y_i\) means the predicted bone age by using the feature vector computed on \(i^{th}\) selected part, and \(y^*\) is the ground-truth bone age. The confidence measures the difference between the predicted age and the ground-truth age. By using Eq. (2), we compare each pair of parts. In the case where a part has higher anchor score but leads to lower confidence, by comparing to another part, the ranking loss proposes a penalty to guide the training of network.
4 Network Training
For network training, we use an objective function, which includes the ranking loss for relation and section module and L2-loss for the final prediction of bone age. The objective function is formulated as:
We construct PRSNet with the open-source Pytorch toolkit. We use different ResNet-34 models in the part relation and selection modules, respectively. In the part relation module, we employ the layers res3, res4 and res5 of the ResNet-34 model to compute anchor scores and select top-3 anchors over all levels. In the selection module, each anchor is used to extract a feature vector on the corresponding layer (res3, res4 or res5) of another ResNet-34 model. During the network training, we augment X-ray images with conventional strategies, i.e., horizontal flipping, rotation, shifting and scaling. We use SGD to optimize PRSNet, where each mini-batch consists of 16 \(512 \times 512\) images. We set the initial learning rate to 1e−3, and decay the learning rate by 10 after every 25 epochs. We train the network on 2 TITAN XP, for 100 epochs that need 5 hours totally. The BAA process on a testing image requires 20 ms. On average, our PRSNet increases the training/testing time by 10%, in comparison with the baseline network.
5 Experiments
5.1 Dataset and Pre-processing
We evaluate PRSNet on the dataset of RSNA 2017 Pediatric Bone Age Challenge. Totally, this dataset contains 12611 X-ray images for training and 200 images for testing. To reduce noise from background in X-ray images, we conduct a lightweight annotation on 200 images to train a foreground segmentation model. Below, we use the trained segmentation model to remove background and select regions of hand bones for training and testing PRSNet. We report all results on the test set, in terms of Mean Absolute Error (MAE). Smaller score of MAE means better performance.
5.2 Ablation Studies
First, we exam the effect on the BBA task by removing the key components of PRSNet, i.e., relation and selection modules. Without these modules, the whole model degrades to the baseline ResNet-34, which yields the score of 6.52 MAE. It lags far behind our full model in terms of BBA accuracy. In Table 1, we test the network without using part relation maps. Here, the part relation module degrades to a basic ResNet-34 model, which yields the score of 5.05 MAE. Comparably, our full model achieves a better score of 4.49 MAE. This is because the relation module provides relevant part information, which is useful for constructing context representation. Next, we disable the part selection module. In this case, we omit the local feature vector and only concatenate the global feature and gender feature for BAA. This model produces a score of 5.20 MAE, which is significantly lower than the result of our full model. Note that hand bone contains useless parts, which embed redundant information to the final feature vector. It demonstrates the effectiveness of our selection module in terms of choosing useful part information (Fig. 5).
5.3 Comparsion with State-of-the-Arts
In Fig. 6, we compare our PRSNet with state-of-the-art methods in terms of the accuracy of BAA. We divide the compared methods into two groups. In the first group, the methods (marked in yellow) select a single part for BAA. Compared to this kind of methods, our PRSNet yields better performance, since our approach provides richer context information of different parts. In the second group, similar to our approach, the methods (marked in blue) also choose multiple parts for BAA. However, this kind of methods neither model part relation nor select parts according to their importance. Thus, our approach outperforms these methods. For a fair comparison with methods (see [6] in Fig. 6) that ensemble several deep models, we train several PRSNets with different random initializations. We achieve a better result than other methods, demonstrating the effectiveness of our joint consideration of part relation and selection.
6 Conclusion
Recent progress on bone age assessment benefits from context information of different parts. In this paper, we have proposed a novel scheme for modeling the part relation. Our method uses relation maps to activate concurrent parts, which form useful context information. Furthermore, the part relation improves the selection of parts for BAA, and the part selection module provides supervision for updating the relation model. We have demonstrated that our approach is effective and outperforms the state-of-the-art on the public benchmarks.
References
Bae, B.U., Bae, W., Jung, K.H.: Improved deep learning model for bone age assessment using triplet ranking loss (2018)
Bejnordi, B.E., et al.: Diagnostic assessment of deep learning algorithms for detection of lymph node metastases in women with breast cancer. JAMA 318(22), 2199–2210 (2017)
Davis, L.M., Theobald, B.-J., Bagnall, A.: Automated bone age assessment using feature extraction. In: Yin, H., Costa, J.A.F., Barreto, G. (eds.) IDEAL 2012. LNCS, vol. 7435, pp. 43–51. Springer, Heidelberg (2012). https://doi.org/10.1007/978-3-642-32639-4_6
Dou, Q., et al.: Automatic detection of cerebral microbleeds from MR images via 3D convolutional neural networks. IEEE Trans. Med. Imaging 35(5), 1182–1195 (2016)
Greulich, W.W., Pyle, S.I., Todd, T.W.: Radiographic Atlas of Skeletal Development of the Hand and Wrist, vol. 2. Stanford University Press, Stanford (1959)
Halabi, S.S., et al.: The RSNA pediatric bone age machine learning challenge. Radiology 290(2), 498–503 (2018)
He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770–778 (2016)
Iglovikov, V.I., Rakhlin, A., Kalinin, A.A., Shvets, A.A.: Paediatric bone age assessment using deep convolutional neural networks. In: Stoyanov, D., et al. (eds.) DLMIA/ML-CDS -2018. LNCS, vol. 11045, pp. 300–308. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-00889-5_34
Larson, D.B., Chen, M.C., Lungren, M.P., Halabi, S.S., Stence, N.V., Langlotz, C.P.: Performance of a deep-learning neural network model in assessing skeletal maturity on pediatric hand radiographs. Radiology 287(1), 313–322 (2017)
Lin, T.Y., Dollár, P., Girshick, R., He, K., Hariharan, B., Belongie, S.: Feature pyramid networks for object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2117–2125 (2017)
Spampinato, C., Palazzo, S., Giordano, D., Aldinucci, M., Leonardi, R.: Deep learning for automated skeletal bone age assessment in X-ray images. Med. Image Anal. 36, 41–51 (2017)
Tanner, J.M., Whitehouse, R., Cameron, N., Marshall, W., Healy, M., Goldstein, H., et al.: Assessment of Skeletal Maturity and Prediction of Adult Height (TW2 Method), vol. 16. Academic press, London (1975)
Thodberg, H.H., Kreiborg, S., Juul, A., Pedersen, K.D.: The BoneXpert method for automated determination of skeletal maturity. IEEE Trans. Med. Imaging 28(1), 52–66 (2009)
Wu, E., et al.: Residual attention based network for hand bone age assessment. arXiv preprint arXiv:1901.05876 (2018)
Yang, Z., Luo, T., Wang, D., Hu, Z., Gao, J., Wang, L.: Learning to navigate for fine-grained classification, pp. 420–435 (2018)
Acknowledgments
We thank the anonymous reviewers for their constructive comments. This work was supported in part by NSFC (61702338) and Shenzhen Science and Technology Program (No. JCYJ20180507182410327).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Ji, Y., Chen, H., Lin, D., Wu, X., Lin, D. (2019). PRSNet: Part Relation and Selection Network for Bone Age Assessment. In: Shen, D., et al. Medical Image Computing and Computer Assisted Intervention – MICCAI 2019. MICCAI 2019. Lecture Notes in Computer Science(), vol 11769. Springer, Cham. https://doi.org/10.1007/978-3-030-32226-7_46
Download citation
DOI: https://doi.org/10.1007/978-3-030-32226-7_46
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-32225-0
Online ISBN: 978-3-030-32226-7
eBook Packages: Computer ScienceComputer Science (R0)