Transformer-based Automated Segmentation of the Median Nerve in Ultrasound Videos of Wrist-to-Elbow Region

Segmenting the median nerve is essential for identifying nerve entrapment syndromes, guiding surgical planning and interventions, and furthering understanding of nerve anatomy. This study aims to develop an automated tool that can assist clinicians in localizing and segmenting the median nerve from...

Full description

Saved in:
Bibliographic Details
Published inIEEE transactions on ultrasonics, ferroelectrics, and frequency control Vol. 71; no. 1; p. 1
Main Authors Gujarati, Karan R., Bathala, Lokesh, Venkatesh, Vaddadi, Mathew, Raji Susan, Yalavarthy, Phaneendra K.
Format Journal Article
LanguageEnglish
Published United States IEEE 01.01.2024
The Institute of Electrical and Electronics Engineers, Inc. (IEEE)
Subjects
Online AccessGet full text
ISSN0885-3010
1525-8955
1525-8955
DOI10.1109/TUFFC.2023.3330539

Cover

Abstract Segmenting the median nerve is essential for identifying nerve entrapment syndromes, guiding surgical planning and interventions, and furthering understanding of nerve anatomy. This study aims to develop an automated tool that can assist clinicians in localizing and segmenting the median nerve from the wrist, mid-forearm, and elbow in ultrasound videos. This is the first fully automated single deep-learning model for accurate segmentation of the median nerve from the wrist to the elbow in ultrasound videos, along with the computation of the cross-sectional area of the nerve. The visual transformer architecture, which was originally proposed to detect and classify 41 classes in YouTube videos, was modified to predict the median nerve in every frame of ultrasound videos. This is achieved by modifying the bounding box sequence matching block of the visual transformer. The median nerve segmentation is a binary class prediction, and the entire bipartite matching sequence is eliminated, enabling a direct comparison of the prediction with expert annotation in a frame-by-frame fashion. Model training, validation, and testing were performed on a dataset comprising ultrasound videos collected from 100 subjects, which were partitioned into 80, 10, and 10 subjects, respectively. The proposed model was compared with U-Net, U-Net++, Siam U-Net, Attention U-Net, LSTM U-Net, and Trans U-Net. The proposed transfomer-based model effectively leveraged the temporal and spatial information present in ultrasound video frames and efficiently segmented the median nerve with an average dice similarity coefficient of approximately 94% at the wrist and 84% in the entire forearm region.
AbstractList Segmenting the median nerve is essential for identifying nerve entrapment syndromes, guiding surgical planning and interventions, and furthering understanding of nerve anatomy. This study aims to develop an automated tool that can assist clinicians in localizing and segmenting the median nerve from the wrist, mid-forearm, and elbow in ultrasound videos. This is the first fully automated single deep learning model for accurate segmentation of the median nerve from the wrist to the elbow in ultrasound videos, along with the computation of the cross-sectional area (CSA) of the nerve. The visual transformer architecture, which was originally proposed to detect and classify 41 classes in YouTube videos, was modified to predict the median nerve in every frame of ultrasound videos. This is achieved by modifying the bounding box sequence matching block of the visual transformer. The median nerve segmentation is a binary class prediction, and the entire bipartite matching sequence is eliminated, enabling a direct comparison of the prediction with expert annotation in a frame-by-frame fashion. Model training, validation, and testing were performed on a dataset comprising ultrasound videos collected from 100 subjects, which were partitioned into 80, ten, and ten subjects, respectively. The proposed model was compared with U-Net, U-Net++, Siam U-Net, Attention U-Net, LSTM U-Net, and Trans U-Net. The proposed transformer-based model effectively leveraged the temporal and spatial information present in ultrasound video frames and efficiently segmented the median nerve with an average dice similarity coefficient (DSC) of approximately 94% at the wrist and 84% in the entire forearm region.Segmenting the median nerve is essential for identifying nerve entrapment syndromes, guiding surgical planning and interventions, and furthering understanding of nerve anatomy. This study aims to develop an automated tool that can assist clinicians in localizing and segmenting the median nerve from the wrist, mid-forearm, and elbow in ultrasound videos. This is the first fully automated single deep learning model for accurate segmentation of the median nerve from the wrist to the elbow in ultrasound videos, along with the computation of the cross-sectional area (CSA) of the nerve. The visual transformer architecture, which was originally proposed to detect and classify 41 classes in YouTube videos, was modified to predict the median nerve in every frame of ultrasound videos. This is achieved by modifying the bounding box sequence matching block of the visual transformer. The median nerve segmentation is a binary class prediction, and the entire bipartite matching sequence is eliminated, enabling a direct comparison of the prediction with expert annotation in a frame-by-frame fashion. Model training, validation, and testing were performed on a dataset comprising ultrasound videos collected from 100 subjects, which were partitioned into 80, ten, and ten subjects, respectively. The proposed model was compared with U-Net, U-Net++, Siam U-Net, Attention U-Net, LSTM U-Net, and Trans U-Net. The proposed transformer-based model effectively leveraged the temporal and spatial information present in ultrasound video frames and efficiently segmented the median nerve with an average dice similarity coefficient (DSC) of approximately 94% at the wrist and 84% in the entire forearm region.
Segmenting the median nerve is essential for identifying nerve entrapment syndromes, guiding surgical planning and interventions, and furthering understanding of nerve anatomy. This study aims to develop an automated tool that can assist clinicians in localizing and segmenting the median nerve from the wrist, mid-forearm, and elbow in ultrasound videos. This is the first fully automated single deep-learning model for accurate segmentation of the median nerve from the wrist to the elbow in ultrasound videos, along with the computation of the cross-sectional area of the nerve. The visual transformer architecture, which was originally proposed to detect and classify 41 classes in YouTube videos, was modified to predict the median nerve in every frame of ultrasound videos. This is achieved by modifying the bounding box sequence matching block of the visual transformer. The median nerve segmentation is a binary class prediction, and the entire bipartite matching sequence is eliminated, enabling a direct comparison of the prediction with expert annotation in a frame-by-frame fashion. Model training, validation, and testing were performed on a dataset comprising ultrasound videos collected from 100 subjects, which were partitioned into 80, 10, and 10 subjects, respectively. The proposed model was compared with U-Net, U-Net++, Siam U-Net, Attention U-Net, LSTM U-Net, and Trans U-Net. The proposed transfomer-based model effectively leveraged the temporal and spatial information present in ultrasound video frames and efficiently segmented the median nerve with an average dice similarity coefficient of approximately 94% at the wrist and 84% in the entire forearm region.
Segmenting the median nerve is essential for identifying nerve entrapment syndromes, guiding surgical planning and interventions, and furthering understanding of nerve anatomy. This study aims to develop an automated tool that can assist clinicians in localizing and segmenting the median nerve from the wrist, mid-forearm, and elbow in ultrasound videos. This is the first fully automated single deep learning model for accurate segmentation of the median nerve from the wrist to the elbow in ultrasound videos, along with the computation of the cross-sectional area (CSA) of the nerve. The visual transformer architecture, which was originally proposed to detect and classify 41 classes in YouTube videos, was modified to predict the median nerve in every frame of ultrasound videos. This is achieved by modifying the bounding box sequence matching block of the visual transformer. The median nerve segmentation is a binary class prediction, and the entire bipartite matching sequence is eliminated, enabling a direct comparison of the prediction with expert annotation in a frame-by-frame fashion. Model training, validation, and testing were performed on a dataset comprising ultrasound videos collected from 100 subjects, which were partitioned into 80, ten, and ten subjects, respectively. The proposed model was compared with U-Net, U-Net++, Siam U-Net, Attention U-Net, LSTM U-Net, and Trans U-Net. The proposed transformer-based model effectively leveraged the temporal and spatial information present in ultrasound video frames and efficiently segmented the median nerve with an average dice similarity coefficient (DSC) of approximately 94% at the wrist and 84% in the entire forearm region.
Author Yalavarthy, Phaneendra K.
Bathala, Lokesh
Mathew, Raji Susan
Gujarati, Karan R.
Venkatesh, Vaddadi
Author_xml – sequence: 1
  givenname: Karan R.
  surname: Gujarati
  fullname: Gujarati, Karan R.
  organization: Department of Computational and Data Sciences, Indian Institute of Science, Bangalore, India
– sequence: 2
  givenname: Lokesh
  surname: Bathala
  fullname: Bathala, Lokesh
  organization: Aster CMI Hospital, Hebbal, Bangalore, India
– sequence: 3
  givenname: Vaddadi
  surname: Venkatesh
  fullname: Venkatesh, Vaddadi
  organization: Department of Computational and Data Sciences, Indian Institute of Science, Bangalore, India
– sequence: 4
  givenname: Raji Susan
  surname: Mathew
  fullname: Mathew, Raji Susan
  organization: Department of Computational and Data Sciences, Indian Institute of Science, Bangalore, India
– sequence: 5
  givenname: Phaneendra K.
  orcidid: 0000-0003-4810-352X
  surname: Yalavarthy
  fullname: Yalavarthy, Phaneendra K.
  organization: Department of Computational and Data Sciences, Indian Institute of Science, Bangalore, India
BackLink https://www.ncbi.nlm.nih.gov/pubmed/37930930$$D View this record in MEDLINE/PubMed
BookMark eNp9kV1rFDEUhoO02G31D4jIgDfezDYfk5nJZVm6KtQWdFcvQyY5qVlmkppkFP-9WXcV6YUQOLl43vPxvufoxAcPCL0geEkIFpeb7Xq9WlJM2ZIxhjkTT9CCcMrrXnB-gha473nNMMFn6DylHcakaQR9is5YJxgub4F2m6h8siFOEOtBJTDV1ZzDpHL5fYL7CXxW2QVfBVvlr1B9AOOUr24hfofK-Wo75qhSmL2pPjsDIe3BL9GlXOdQX49D-FF9hPvS4Rk6tWpM8PxYL9B2fb1Zvatv7t6-X13d1JpxmmujwTDNWsuFbQXX1AwtbXvVMj6ApqwhnRosE6TjorG85ZpQw8BSjbXADbAL9ObQ9yGGbzOkLCeXNIyj8hDmJGnft4J1lIuCvn6E7sIcfdlOUkGoaGjfdIV6daTmYQIjH6KbVPwp_7hYAHoAdAwpRbB_EYLlPir5Oyq5j0oeoyqi_pFIu4PVxVA3_l_68iB1APDPrLKNKFf9AjhIoN0
CODEN ITUCER
CitedBy_id crossref_primary_10_1109_TETCI_2024_3377676
crossref_primary_10_1007_s00264_025_06497_1
crossref_primary_10_1007_s10278_024_01267_8
Cites_doi 10.5555/3454287.3455008
10.1016/j.compmedimag.2019.101687
10.1109/CVPR46437.2021.01212
10.1007/s11517-022-02662-5
10.1002/mp.16662
10.1109/CVPR46437.2021.00681
10.1016/j.imu.2016.06.003
10.1093/bja/aei002
10.1093/bja/aer085
10.1109/ICIVC47709.2019.8981027
10.48550/arXiv.1802.02611
10.1016/j.media.2011.02.003
10.1517/17530059.2012.692904
10.1097/ALN.0b013e3181c5dfd7
10.1007/978-3-030-00889-5_1
10.3390/s22165984
10.1109/ICCV48922.2021.00676
10.1109/ICCC54389.2021.9674267
10.1186/s42492-023-00140-9
10.1016/j.media.2023.102802
10.1109/BIBM52615.2021.9669743
10.1016/j.media.2019.101631
10.1016/S0033-8389(03)00162-3
10.1109/CVPR.2013.312
10.1002/mus.21286
10.1109/CVPR46437.2021.00863
10.1109/ICIP.2015.7351588
10.3390/diagnostics11101893
10.1007/s11910-012-0328-x
10.1109/TNSRE.2018.2833141
10.1002/mus.23912
10.1007/978-3-030-87231-1_50
10.1016/j.cnp.2017.09.001
10.1016/j.ultrasmedbio.2020.03.017
10.1007/978-3-319-24574-4_28
10.1016/j.compbiomed.2023.107268
10.1109/ICCV.2017.89
10.1016/j.compbiomed.2023.106629
10.1007/978-3-030-01261-8_1
10.1109/ICCV.2017.322
10.1109/TNSRE.2007.891393
10.1007/978-3-030-87199-4_9
10.1109/CIBCB48159.2020.9277638
10.1016/j.ultrasmedbio.2021.03.018
10.1007/978-3-030-58452-8_13
10.1109/ICCV.2017.324
ContentType Journal Article
Copyright Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2024
Copyright_xml – notice: Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2024
DBID 97E
RIA
RIE
AAYXX
CITATION
CGR
CUY
CVF
ECM
EIF
NPM
7SP
7U5
8FD
F28
FR3
L7M
7X8
DOI 10.1109/TUFFC.2023.3330539
DatabaseName IEEE All-Society Periodicals Package (ASPP) 2005–Present
IEEE All-Society Periodicals Package (ASPP) 1998–Present
IEEE Electronic Library (IEL)
CrossRef
Medline
MEDLINE
MEDLINE (Ovid)
MEDLINE
MEDLINE
PubMed
Electronics & Communications Abstracts
Solid State and Superconductivity Abstracts
Technology Research Database
ANTE: Abstracts in New Technology & Engineering
Engineering Research Database
Advanced Technologies Database with Aerospace
MEDLINE - Academic
DatabaseTitle CrossRef
MEDLINE
Medline Complete
MEDLINE with Full Text
PubMed
MEDLINE (Ovid)
Solid State and Superconductivity Abstracts
Engineering Research Database
Technology Research Database
Advanced Technologies Database with Aerospace
ANTE: Abstracts in New Technology & Engineering
Electronics & Communications Abstracts
MEDLINE - Academic
DatabaseTitleList MEDLINE - Academic

MEDLINE
Solid State and Superconductivity Abstracts
Database_xml – sequence: 1
  dbid: NPM
  name: PubMed
  url: https://proxy.k.utb.cz/login?url=http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?db=PubMed
  sourceTypes: Index Database
– sequence: 2
  dbid: EIF
  name: MEDLINE
  url: https://proxy.k.utb.cz/login?url=https://www.webofscience.com/wos/medline/basic-search
  sourceTypes: Index Database
– sequence: 3
  dbid: RIE
  name: IEEE Xplore digital library
  url: https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/
  sourceTypes: Publisher
DeliveryMethod fulltext_linktorsrc
Discipline Engineering
Physics
EISSN 1525-8955
EndPage 1
ExternalDocumentID 37930930
10_1109_TUFFC_2023_3330539
10309959
Genre orig-research
Research Support, Non-U.S. Gov't
Journal Article
GrantInformation_xml – fundername: GE Healthcare
  funderid: 10.13039/100006775
GroupedDBID ---
-~X
0R~
29I
4.4
53G
5GY
5RE
6IK
97E
AAJGR
AARMG
AASAJ
AAWTH
ABAZT
ABQJQ
ABVLG
ACGFO
ACGFS
ACIWK
AENEX
AGQYO
AHBIQ
AKJIK
AKQYR
ALMA_UNASSIGNED_HOLDINGS
ATWAV
BEFXN
BFFAM
BGNUA
BKEBE
BPEOZ
CS3
DU5
EBS
F5P
HZ~
IFIPE
IPLJI
JAVBF
LAI
M43
O9-
OCL
P2P
RIA
RIE
RNS
TN5
TWZ
.GJ
186
3EH
5VS
AAYXX
AETIX
AGSQL
AI.
AIBXA
ALLEH
CITATION
EJD
H~9
ICLAB
IFJZH
UKR
VH1
ZXP
ZY4
CGR
CUY
CVF
ECM
EIF
NPM
RIG
7SP
7U5
8FD
F28
FR3
L7M
7X8
ID FETCH-LOGICAL-c352t-dced3c36f59f695c2db6268a635bec23417abf3917594f565c12d3ef2c0c904e3
IEDL.DBID RIE
ISSN 0885-3010
1525-8955
IngestDate Sun Sep 28 10:01:19 EDT 2025
Thu Sep 11 05:41:14 EDT 2025
Mon Jul 21 06:01:29 EDT 2025
Wed Oct 01 02:33:57 EDT 2025
Thu Apr 24 23:04:09 EDT 2025
Wed Aug 27 02:35:00 EDT 2025
IsPeerReviewed true
IsScholarly true
Issue 1
Language English
License https://ieeexplore.ieee.org/Xplorehelp/downloads/license-information/IEEE.html
https://doi.org/10.15223/policy-029
https://doi.org/10.15223/policy-037
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-c352t-dced3c36f59f695c2db6268a635bec23417abf3917594f565c12d3ef2c0c904e3
Notes ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
content type line 23
ORCID 0000-0003-4810-352X
0000-0002-8353-351X
0009-0009-3554-8424
PMID 37930930
PQID 2912942847
PQPubID 85455
PageCount 1
ParticipantIDs crossref_citationtrail_10_1109_TUFFC_2023_3330539
crossref_primary_10_1109_TUFFC_2023_3330539
ieee_primary_10309959
proquest_miscellaneous_2886937259
proquest_journals_2912942847
pubmed_primary_37930930
ProviderPackageCode CITATION
AAYXX
PublicationCentury 2000
PublicationDate 2024-01-01
PublicationDateYYYYMMDD 2024-01-01
PublicationDate_xml – month: 01
  year: 2024
  text: 2024-01-01
  day: 01
PublicationDecade 2020
PublicationPlace United States
PublicationPlace_xml – name: United States
– name: New York
PublicationTitle IEEE transactions on ultrasonics, ferroelectrics, and frequency control
PublicationTitleAbbrev T-UFFC
PublicationTitleAlternate IEEE Trans Ultrason Ferroelectr Freq Control
PublicationYear 2024
Publisher IEEE
The Institute of Electrical and Electronics Engineers, Inc. (IEEE)
Publisher_xml – name: IEEE
– name: The Institute of Electrical and Electronics Engineers, Inc. (IEEE)
References ref57
ref12
Hafiane (ref13) 2017
ref56
ref15
ref14
ref58
ref53
Kingma (ref50) 2014
ref11
ref10
ref54
ref17
Naseer (ref44); 34
ref16
Zhu (ref39) 2020
ref19
ref18
Oktay (ref55) 2018
ref51
Chen (ref26) 2021
ref46
ref48
ref47
ref42
ref43
ref49
ref8
ref7
ref9
ref4
ref3
ref6
ref5
ref40
ref35
Dosovitskiy (ref38) 2020
ref34
ref37
ref36
ref31
ref30
ref33
ref32
ref2
ref1
Loshchilov (ref52) 2017
Ferreira (ref45) 2012
ref23
ref20
Bahdanau (ref25) 2014
ref22
ref21
ref28
ref27
ref29
Kumar (ref41) 2021
Kremkau (ref24) 2001
References_xml – ident: ref49
  doi: 10.5555/3454287.3455008
– year: 2017
  ident: ref52
  article-title: Decoupled weight decay regularization
  publication-title: arXiv:1711.05101
– ident: ref14
  doi: 10.1016/j.compmedimag.2019.101687
– ident: ref42
  doi: 10.1109/CVPR46437.2021.01212
– ident: ref22
  doi: 10.1007/s11517-022-02662-5
– ident: ref32
  doi: 10.1002/mp.16662
– ident: ref40
  doi: 10.1109/CVPR46437.2021.00681
– year: 2014
  ident: ref50
  article-title: Adam: A method for stochastic optimization
  publication-title: arXiv:1412.6980
– ident: ref12
  doi: 10.1016/j.imu.2016.06.003
– year: 2020
  ident: ref38
  article-title: An image is worth 16×16 words: Transformers for image recognition at scale
  publication-title: arXiv:2010.11929
– ident: ref5
  doi: 10.1093/bja/aei002
– ident: ref6
  doi: 10.1093/bja/aer085
– ident: ref56
  doi: 10.1109/ICIVC47709.2019.8981027
– year: 2020
  ident: ref39
  article-title: Deformable DETR: Deformable transformers for end-to-end object detection
  publication-title: arXiv:2010.04159
– volume-title: ImageJ User Guide: IJ 1.42 R
  year: 2012
  ident: ref45
– ident: ref18
  doi: 10.48550/arXiv.1802.02611
– ident: ref23
  doi: 10.1016/j.media.2011.02.003
– ident: ref8
  doi: 10.1517/17530059.2012.692904
– ident: ref10
  doi: 10.1097/ALN.0b013e3181c5dfd7
– ident: ref53
  doi: 10.1007/978-3-030-00889-5_1
– ident: ref33
  doi: 10.3390/s22165984
– ident: ref43
  doi: 10.1109/ICCV48922.2021.00676
– year: 2021
  ident: ref26
  article-title: TransUNet: Transformers make strong encoders for medical image segmentation
  publication-title: arXiv:2102.04306
– ident: ref28
  doi: 10.1109/ICCC54389.2021.9674267
– ident: ref37
  doi: 10.1186/s42492-023-00140-9
– year: 2017
  ident: ref13
  article-title: Deep learning with spatiotemporal consistency for nerve segmentation in ultrasound images
  publication-title: arXiv:1706.05870
– ident: ref35
  doi: 10.1016/j.media.2023.102802
– ident: ref30
  doi: 10.1109/BIBM52615.2021.9669743
– ident: ref54
  doi: 10.1016/j.media.2019.101631
– ident: ref7
  doi: 10.1016/S0033-8389(03)00162-3
– ident: ref57
  doi: 10.1109/CVPR.2013.312
– ident: ref3
  doi: 10.1002/mus.21286
– ident: ref46
  doi: 10.1109/CVPR46437.2021.00863
– ident: ref11
  doi: 10.1109/ICIP.2015.7351588
– ident: ref21
  doi: 10.3390/diagnostics11101893
– ident: ref9
  doi: 10.1007/s11910-012-0328-x
– ident: ref1
  doi: 10.1109/TNSRE.2018.2833141
– ident: ref58
  doi: 10.1002/mus.23912
– ident: ref29
  doi: 10.1007/978-3-030-87231-1_50
– ident: ref4
  doi: 10.1016/j.cnp.2017.09.001
– ident: ref17
  doi: 10.1016/j.ultrasmedbio.2020.03.017
– ident: ref15
  doi: 10.1007/978-3-319-24574-4_28
– ident: ref36
  doi: 10.1016/j.compbiomed.2023.107268
– ident: ref47
  doi: 10.1109/ICCV.2017.89
– year: 2018
  ident: ref55
  article-title: Attention U-net: Learning where to look for the pancreas
  publication-title: arXiv:1804.03999
– ident: ref34
  doi: 10.1016/j.compbiomed.2023.106629
– ident: ref48
  doi: 10.1007/978-3-030-01261-8_1
– year: 2014
  ident: ref25
  article-title: Neural machine translation by jointly learning to align and translate
  publication-title: arXiv:1409.0473
– ident: ref20
  doi: 10.1109/ICCV.2017.322
– ident: ref2
  doi: 10.1109/TNSRE.2007.891393
– ident: ref31
  doi: 10.1007/978-3-030-87199-4_9
– ident: ref51
  doi: 10.1109/CIBCB48159.2020.9277638
– ident: ref16
  doi: 10.1016/j.ultrasmedbio.2021.03.018
– volume-title: Diagnostic Ultrasound: Principles and Instrumentsa
  year: 2001
  ident: ref24
– ident: ref27
  doi: 10.1007/978-3-030-58452-8_13
– volume: 34
  start-page: 23296
  volume-title: Proc. Adv. Neural Inf. Process. Syst.
  ident: ref44
  article-title: Intriguing properties of vision transformers
– ident: ref19
  doi: 10.1109/ICCV.2017.324
– year: 2021
  ident: ref41
  article-title: Colorization transformer
  publication-title: arXiv:2102.04432
SSID ssj0014492
Score 2.4421475
Snippet Segmenting the median nerve is essential for identifying nerve entrapment syndromes, guiding surgical planning and interventions, and furthering understanding...
SourceID proquest
pubmed
crossref
ieee
SourceType Aggregation Database
Index Database
Enrichment Source
Publisher
StartPage 1
SubjectTerms Annotations
Automation
Cross-sectional Area
Elbow
Elbow (anatomy)
Electric Power Supplies
Entrapment
Forearm
Humans
Image Processing, Computer-Assisted
Image segmentation
Matching
Median (statistics)
Median nerve
Median Nerve - diagnostic imaging
Median Nerve Segmentation
Segmentation
Spatial data
Task analysis
Transformers
Ultrasonic imaging
Ultrasonography
Ultrasound Video
Video
Videos
Vision Transformer
Visualization
Wrist
Wrist - diagnostic imaging
Title Transformer-based Automated Segmentation of the Median Nerve in Ultrasound Videos of Wrist-to-Elbow Region
URI https://ieeexplore.ieee.org/document/10309959
https://www.ncbi.nlm.nih.gov/pubmed/37930930
https://www.proquest.com/docview/2912942847
https://www.proquest.com/docview/2886937259
Volume 71
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
journalDatabaseRights – providerCode: PRVIEE
  databaseName: IEEE Xplore digital library
  customDbUrl:
  eissn: 1525-8955
  dateEnd: 99991231
  omitProxy: false
  ssIdentifier: ssj0014492
  issn: 0885-3010
  databaseCode: RIE
  dateStart: 19860101
  isFulltext: true
  titleUrlDefault: https://ieeexplore.ieee.org/
  providerName: IEEE
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV3da9RAEB9soVAf1H6op1VW8E02TbObr8dSehQf7kHv2r6F7O5E2l4TuUsQ_Oud2STHIVSEPCxkstkws5nf7HwBfM6rJNUuczLKnZLahSi5eqFUxjiyiDQfRXK0xSy5Wuivt_HtkKzuc2EQ0QefYcBD78t3je34qOyUW2Jxfawd2EmzpE_W2rgMtPYdkGnXxJKkNhwzZML8dL6YTi8CbhQeKLLfY-4MvqWFfFuVpxGm1zTTlzAb19gHmDwEXWsC-_uv8o3__RGv4MWAOcV5LyQH8AzrQ3i-VYnwEPZ8JKhdH8H9fISyuJKs45w479qGgC2NvuOPxyFZqRZNJQg9Cnb1lLWYceikuKvFYtmuyjV3axLXdw6bNRPe8L9Eto28XJrml_iGHAZ9DIvp5fziSg4NGaQlnNZKZ9Epq5IqJhbnsY2cIXsoKwm0kChEpBDT0lSKLMA41xVBRXsWOYVVZEObhxrVa9itmxrfgkjZ4ZMgAQRjtKaJQlPGqSqRQWTl7ATORgYVdqhWzk0zloW3WsK88EwtmKnFwNQJfNk887Ov1fFP6mNmzhZlz5cJnIyCUAz7eV1EOeEizap8Ap82t2knsnulrLHpiCbLEgJ7EU_xphegzeSKfoMhXe-eeOl72Ke16f5s5wR221WHHwjttOajl_I_vpb48Q
linkProvider IEEE
linkToHtml http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV3db9MwED_BEAIe-BgDCgOMxBtylsV22jxO06oCow_Qwt6i2L6gQUlQmwiJv547J6kqpCGkPFjKxXF059zvfF8Ar7MyHWs_8TLJvJLaxyi5eqFU1nqyiDQfRXK0xTydLfW7C3PRJ6uHXBhEDMFnGPEw-PJ97Vo-KjvillhcH-s63DBaa9Ola22dBlqHHsi0b4wkuY2HHJk4O1osp9PTiFuFR4oseMO9wXf0UGiscjXGDLpmeg_mwyq7EJPvUdvYyP3-q4Djf3_Gfbjbo05x0onJA7iG1T7c2alFuA83Qyyo2zyEb4sBzOJaspbz4qRtaoK2NPqEX3_06UqVqEtB-FGws6eoxJyDJ8VlJZarZl1suF-T-Hzpsd4w4Rf-m8imlmcrW_8SH5EDoQ9gOT1bnM5k35JBOkJqjfQOvXIqLQ0xOTMu8ZYsoklBsIWEISGVOC5sqcgGNJkuCSy648QrLBMXuyzWqB7BXlVX-ATEmF0-KRJEsFZrmii2hRmrAhlGlt6N4HhgUO76euXcNmOVB7slzvLA1JyZmvdMHcGb7TM_u2od_6Q-YObsUHZ8GcHhIAh5v6M3eZIRMtKszEfwanub9iI7WIoK65ZoJpOU4F7CUzzuBGg7uaIfYUzX0yte-hJuzRYfzvPzt_P3z-A2rVN3Jz2HsNesW3xO2KexL4LE_wHquPw-
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Transformer-Based+Automated+Segmentation+of+the+Median+Nerve+in+Ultrasound+Videos+of+Wrist-to-Elbow+Region&rft.jtitle=IEEE+transactions+on+ultrasonics%2C+ferroelectrics%2C+and+frequency+control&rft.au=Gujarati%2C+Karan+R.&rft.au=Bathala%2C+Lokesh&rft.au=Venkatesh%2C+Vaddadi&rft.au=Mathew%2C+Raji+Susan&rft.date=2024-01-01&rft.issn=0885-3010&rft.eissn=1525-8955&rft.volume=71&rft.issue=1&rft.spage=56&rft.epage=69&rft_id=info:doi/10.1109%2FTUFFC.2023.3330539&rft.externalDBID=n%2Fa&rft.externalDocID=10_1109_TUFFC_2023_3330539
thumbnail_l http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/lc.gif&issn=0885-3010&client=summon
thumbnail_m http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/mc.gif&issn=0885-3010&client=summon
thumbnail_s http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/sc.gif&issn=0885-3010&client=summon