Journal article
Accurate volume alignment of arbitrarily oriented tibiae based on a mutual attention network for osteoarthritis analysis
- Abstract:
-
Damage to cartilage is an important indicator of osteoarthritis progression, but manual extraction of cartilage morphology is time-consuming and prone to error. To address this, we hypothesize that automatic labeling of cartilage can be achieved through the comparison of contrasted and non-contrasted Computer Tomography (CT). However, this is non-trivial as the pre-clinical volumes are at arbitrary starting poses due to the lack of standardized acquisition protocols. Thus, we propose an annotation-free deep learning method, D-net, for accurate and automatic alignment of pre- and post-contrasted cartilage CT volumes. D-Net is based on a novel mutual attention network structure to capture large-range translation and full-range rotation without the need for a prior pose template. CT volumes of mice tibiae are used for validation, with synthetic transformation for training and tested with real pre- and post-contrasted CT volumes. Analysis of Variance (ANOVA) was used to compare the different network structures. Our proposed method, D-net, achieves a Dice coefficient of 0.87, and significantly outperforms other state-of-the-art deep learning models, in the real-world alignment of 50 pairs of pre- and post-contrasted CT volumes when cascaded as a multi-stage network.
- Publication status:
- Published
- Peer review status:
- Peer reviewed
Actions
Access Document
- Files:
-
-
(Preview, Version of record, pdf, 8.1MB, Terms of use)
-
- Publisher copy:
- 10.1016/j.compmedimag.2023.102204
Authors
- Publisher:
- Elsevier
- Journal:
- Computerized Medical Imaging and Graphics More from this journal
- Volume:
- 106
- Article number:
- 102204
- Publication date:
- 2023-02-24
- Acceptance date:
- 2023-02-14
- DOI:
- EISSN:
-
1879-0771
- ISSN:
-
0895-6111
- Language:
-
English
- Keywords:
- Pubs id:
-
1330643
- Local pid:
-
pubs:1330643
- Deposit date:
-
2023-02-27
Terms of use
- Copyright holder:
- Zheng et al.
- Copyright date:
- 2023
- Rights statement:
- © 2023 The Author(s). Published by Elsevier Ltd. This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/).
- Licence:
- CC Attribution (CC BY)
If you are the owner of this record, you can report an update to it here: Report update to this record