Home > Publications Database > AutoSNAP: Automatically Learning Neural Architectures for Instrument Pose Estimation |
Contribution to a conference proceedings/Contribution to a book | DZNE-2025-01010 |
; ; ; ; ; ; ; ; ; ; ;
2020
Springer International Publishing
Cham
ISBN: 978-3-030-59715-3 (print), 978-3-030-59716-0 (electronic)
This record in other databases:
Please use a persistent id in citations: doi:10.1007/978-3-030-59716-0_36
Abstract: Despite recent successes, the advances in Deep Learning have not yet been fully translated to Computer Assisted Intervention (CAI) problems such as pose estimation of surgical instruments. Currently, neural architectures for classification and segmentation tasks are adopted ignoring significant discrepancies between CAI and these tasks. We propose an automatic framework (AutoSNAP) for instrument pose estimation problems, which discovers and learns architectures for neural networks. We introduce 1) an efficient testing environment for pose estimation, 2) a powerful architecture representation based on novel Symbolic Neural Architecture Patterns (SNAPs), and 3) an optimization of the architecture using an efficient search scheme. Using AutoSNAP, we discover an improved architecture (SNAPNet) which outperforms both the hand-engineered i3PosNet and the state-of-the-art architecture search method DARTS.
![]() |
The record appears in these collections: |