Core Concepts
The author proposes a two-shot training paradigm for breast ultrasound video segmentation to address the challenges of dense annotation requirements and lack of space-time awareness in existing methods.
Abstract
The content discusses a novel approach for segmenting breast lesions in ultrasound videos using a label-efficient two-shot training paradigm. By leveraging semi-supervised learning and source-dependent augmentation, the proposed method achieves comparable performance with minimal training labels. The study highlights the importance of accurate lesion delineation in early breast cancer diagnosis and treatment. Computer-aided diagnosis tools are essential due to the complexity of interpreting ultrasound images, emphasizing the need for automated segmentation methods. The research introduces a space-time consistency supervision module to enhance feature alignment across video frames, improving segmentation accuracy. Experimental results demonstrate that the proposed method outperforms fully-supervised models, showcasing its potential for efficient and accurate breast ultrasound video segmentation.
Stats
Results showed that it gained comparable performance to the fully annotated ones given only 1.9% training labels.
STCN achieved 72.1% J & F score, 73.1% J score, 71.1% F score, 80.4% DSC, and 7.77 HD.
XMem reached 73.4% J & F score, 74.6% J score, 72.3% F score, 82.6% DSC, and 7.82 HD.
STCN-vanilla achieved 68.1% J & F score, 69.1% J score, 67.0% F score, 77.3% DSC, and 8.17 HD.
STCN w/ Ours obtained a J & F score of 72.3+4%, J score of 73.1+4%, F score of 71.5+4%, DSC of 81.2+3%, and reduced HD to 7..87.
Quotes
"The proposed method prevents model degradation by discarding inaccurate pseudo-labels during training."
"Our results show that the two-shot annotation strategy can generate satisfactory BUS segmentation masks with proper design."
"The addition of space-time consistency supervision elevated baseline performance by enhancing temporal dependency."