253 related articles for article (PubMed ID: 37103233)
1. A 3DCNN-Based Knowledge Distillation Framework for Human Activity Recognition.
Ullah H; Munir A
J Imaging; 2023 Apr; 9(4):. PubMed ID: 37103233
[TBL] [Abstract][Full Text] [Related]
2. Self-knowledge distillation for surgical phase recognition.
Zhang J; Barbarisi S; Kadkhodamohammadi A; Stoyanov D; Luengo I
Int J Comput Assist Radiol Surg; 2024 Jan; 19(1):61-68. PubMed ID: 37340283
[TBL] [Abstract][Full Text] [Related]
3. Human Activity Recognition Using Cascaded Dual Attention CNN and Bi-Directional GRU Framework.
Ullah H; Munir A
J Imaging; 2023 Jun; 9(7):. PubMed ID: 37504807
[TBL] [Abstract][Full Text] [Related]
4. Cross-View Gait Recognition Method Based on Multi-Teacher Joint Knowledge Distillation.
Li R; Yun L; Zhang M; Yang Y; Cheng F
Sensors (Basel); 2023 Nov; 23(22):. PubMed ID: 38005675
[TBL] [Abstract][Full Text] [Related]
5. DCCD: Reducing Neural Network Redundancy via Distillation.
Liu Y; Chen J; Liu Y
IEEE Trans Neural Netw Learn Syst; 2023 Jan; PP():. PubMed ID: 37022254
[TBL] [Abstract][Full Text] [Related]
6. A General Dynamic Knowledge Distillation Method for Visual Analytics.
Tu Z; Liu X; Xiao X
IEEE Trans Image Process; 2022 Oct; PP():. PubMed ID: 36227819
[TBL] [Abstract][Full Text] [Related]
7. LHAR: Lightweight Human Activity Recognition on Knowledge Distillation.
Deng S; Chen J; Teng D; Yang C; Chen D; Jia T; Wang H
IEEE J Biomed Health Inform; 2023 Jul; PP():. PubMed ID: 37494155
[TBL] [Abstract][Full Text] [Related]
8. Light-M: An efficient lightweight medical image segmentation framework for resource-constrained IoMT.
Zhang Y; Chen Z; Yang X
Comput Biol Med; 2024 Mar; 170():108088. PubMed ID: 38320339
[TBL] [Abstract][Full Text] [Related]
9. Multi-view Teacher-Student Network.
Tian Y; Sun S; Tang J
Neural Netw; 2022 Feb; 146():69-84. PubMed ID: 34839092
[TBL] [Abstract][Full Text] [Related]
10. SSD-KD: A self-supervised diverse knowledge distillation method for lightweight skin lesion classification using dermoscopic images.
Wang Y; Wang Y; Cai J; Lee TK; Miao C; Wang ZJ
Med Image Anal; 2023 Feb; 84():102693. PubMed ID: 36462373
[TBL] [Abstract][Full Text] [Related]
11. Multistage feature fusion knowledge distillation.
Li G; Wang K; Lv P; He P; Zhou Z; Xu C
Sci Rep; 2024 Jun; 14(1):13373. PubMed ID: 38862547
[TBL] [Abstract][Full Text] [Related]
12. MSKD: Structured knowledge distillation for efficient medical image segmentation.
Zhao L; Qian X; Guo Y; Song J; Hou J; Gong J
Comput Biol Med; 2023 Sep; 164():107284. PubMed ID: 37572439
[TBL] [Abstract][Full Text] [Related]
13. Real-Time Correlation Tracking via Joint Model Compression and Transfer.
Wang N; Zhou W; Song Y; Ma C; Li H
IEEE Trans Image Process; 2020 Apr; ():. PubMed ID: 32356748
[TBL] [Abstract][Full Text] [Related]
14. An ultra-fast deep-learning-based dose engine for prostate VMAT via knowledge distillation framework with limited patient data.
Tseng W; Liu H; Yang Y; Liu C; Lu B
Phys Med Biol; 2022 Dec; 68(1):. PubMed ID: 36533689
[No Abstract] [Full Text] [Related]
15. Learning Student Networks via Feature Embedding.
Chen H; Wang Y; Xu C; Xu C; Tao D
IEEE Trans Neural Netw Learn Syst; 2021 Jan; 32(1):25-35. PubMed ID: 32092018
[TBL] [Abstract][Full Text] [Related]
16. Resolution-Aware Knowledge Distillation for Efficient Inference.
Feng Z; Lai J; Xie X
IEEE Trans Image Process; 2021; 30():6985-6996. PubMed ID: 34347598
[TBL] [Abstract][Full Text] [Related]
17. Teacher-student complementary sample contrastive distillation.
Bao Z; Huang Z; Gou J; Du L; Liu K; Zhou J; Chen Y
Neural Netw; 2024 Feb; 170():176-189. PubMed ID: 37989039
[TBL] [Abstract][Full Text] [Related]
18. A New Deep-Learning Method for Human Activity Recognition.
Vrskova R; Kamencay P; Hudec R; Sykora P
Sensors (Basel); 2023 Mar; 23(5):. PubMed ID: 36905020
[TBL] [Abstract][Full Text] [Related]
19. Learning From Human Educational Wisdom: A Student-Centered Knowledge Distillation Method.
Yang S; Yang J; Zhou M; Huang Z; Zheng WS; Yang X; Ren J
IEEE Trans Pattern Anal Mach Intell; 2024 Jun; 46(6):4188-4205. PubMed ID: 38227419
[TBL] [Abstract][Full Text] [Related]
20. Adversarial learning-based multi-level dense-transmission knowledge distillation for AP-ROP detection.
Xie H; Liu Y; Lei H; Song T; Yue G; Du Y; Wang T; Zhang G; Lei B
Med Image Anal; 2023 Feb; 84():102725. PubMed ID: 36527770
[TBL] [Abstract][Full Text] [Related]
[Next] [New Search]