Training Data Augmentation for Deep Learning Radio Frequency Systems
Applications of machine learning are subject to three major components that contribute to the final performance metrics. Within the category of neural networks, and deep learning specifically, the first two are the architecture for the model being trained and the training approach used. This work fo...
Saved in:
Published in | arXiv.org |
---|---|
Main Authors | , , , |
Format | Paper Journal Article |
Language | English |
Published |
Ithaca
Cornell University Library, arXiv.org
04.01.2021
|
Subjects | |
Online Access | Get full text |
ISSN | 2331-8422 |
DOI | 10.48550/arxiv.2010.00178 |
Cover
Abstract | Applications of machine learning are subject to three major components that contribute to the final performance metrics. Within the category of neural networks, and deep learning specifically, the first two are the architecture for the model being trained and the training approach used. This work focuses on the third component, the data used during training. The primary questions that arise are ``what is in the data'' and ``what within the data matters?'' Looking into the Radio Frequency Machine Learning (RFML) field of Automatic Modulation Classification (AMC) as an example of a tool used for situational awareness, the use of synthetic, captured, and augmented data are examined and compared to provide insights about the quantity and quality of the available data necessary to achieve desired performance levels. There are three questions discussed within this work: (1) how useful a synthetically trained system is expected to be when deployed without considering the environment within the synthesis, (2) how can augmentation be leveraged within the RFML domain, and lastly, (3) what impact knowledge of degradations to the signal caused by the transmission channel contributes to the performance of a system. In general, the examined data types each have useful contributions to a final application, but captured data germane to the intended use case will always provide more significant information and enable the greatest performance. Despite the benefit of captured data, the difficulties and costs that arise from live collection often make the quantity of data needed to achieve peak performance impractical. This paper helps quantify the balance between real and synthetic data, offering concrete examples where training data is parametrically varied in size and source. |
---|---|
AbstractList | Applications of machine learning are subject to three major components that
contribute to the final performance metrics. Within the category of neural
networks, and deep learning specifically, the first two are the architecture
for the model being trained and the training approach used. This work focuses
on the third component, the data used during training. The primary questions
that arise are ``what is in the data'' and ``what within the data matters?''
Looking into the Radio Frequency Machine Learning (RFML) field of Automatic
Modulation Classification (AMC) as an example of a tool used for situational
awareness, the use of synthetic, captured, and augmented data are examined and
compared to provide insights about the quantity and quality of the available
data necessary to achieve desired performance levels. There are three questions
discussed within this work: (1) how useful a synthetically trained system is
expected to be when deployed without considering the environment within the
synthesis, (2) how can augmentation be leveraged within the RFML domain, and
lastly, (3) what impact knowledge of degradations to the signal caused by the
transmission channel contributes to the performance of a system. In general,
the examined data types each have useful contributions to a final application,
but captured data germane to the intended use case will always provide more
significant information and enable the greatest performance. Despite the
benefit of captured data, the difficulties and costs that arise from live
collection often make the quantity of data needed to achieve peak performance
impractical. This paper helps quantify the balance between real and synthetic
data, offering concrete examples where training data is parametrically varied
in size and source. Applications of machine learning are subject to three major components that contribute to the final performance metrics. Within the category of neural networks, and deep learning specifically, the first two are the architecture for the model being trained and the training approach used. This work focuses on the third component, the data used during training. The primary questions that arise are ``what is in the data'' and ``what within the data matters?'' Looking into the Radio Frequency Machine Learning (RFML) field of Automatic Modulation Classification (AMC) as an example of a tool used for situational awareness, the use of synthetic, captured, and augmented data are examined and compared to provide insights about the quantity and quality of the available data necessary to achieve desired performance levels. There are three questions discussed within this work: (1) how useful a synthetically trained system is expected to be when deployed without considering the environment within the synthesis, (2) how can augmentation be leveraged within the RFML domain, and lastly, (3) what impact knowledge of degradations to the signal caused by the transmission channel contributes to the performance of a system. In general, the examined data types each have useful contributions to a final application, but captured data germane to the intended use case will always provide more significant information and enable the greatest performance. Despite the benefit of captured data, the difficulties and costs that arise from live collection often make the quantity of data needed to achieve peak performance impractical. This paper helps quantify the balance between real and synthetic data, offering concrete examples where training data is parametrically varied in size and source. |
Author | Clark, William H Michaels, Alan J Hauser, Steven Headley, William C |
Author_xml | – sequence: 1 givenname: William surname: Clark middlename: H fullname: Clark, William H – sequence: 2 givenname: Steven surname: Hauser fullname: Hauser, Steven – sequence: 3 givenname: William surname: Headley middlename: C fullname: Headley, William C – sequence: 4 givenname: Alan surname: Michaels middlename: J fullname: Michaels, Alan J |
BackLink | https://doi.org/10.48550/arXiv.2010.00178$$DView paper in arXiv https://doi.org/10.1177/1548512921991245$$DView published paper (Access to full text may be restricted) |
BookMark | eNotj8FKw0AURQdRsNZ-gCsHXKe-eS8zkyxLY1UoCNp9mMSXkmIndZKI_fvG1NWFy-Fyz4249I1nIe4UzONEa3h04bf-mSMMBYCyyYWYIJGKkhjxWszadgcAaCxqTRORbYKrfe23MnOdk4t-u2ffua5uvKyaIDPmg1yzCyPz7j7rRq4Cf_fsy6P8OLYd79tbcVW5r5Zn_zkVm9XTZvkSrd-eX5eLdeQ0mihWBVLCBlVSlLpQhjWCoRTZlrGzJTlFVoGxZFNVIWmT6gHlEgiKstI0Fffn2VExP4R678Ix_1PNR9WBeDgTh9AMF9su3zV98MOnHOM4ATLWGDoBvS5WWA |
ContentType | Paper Journal Article |
Copyright | 2021. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License. http://arxiv.org/licenses/nonexclusive-distrib/1.0 |
Copyright_xml | – notice: 2021. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License. – notice: http://arxiv.org/licenses/nonexclusive-distrib/1.0 |
DBID | 8FE 8FG ABJCF ABUWG AFKRA AZQEC BENPR BGLVJ CCPQU DWQXO HCIFZ L6V M7S PHGZM PHGZT PIMPY PKEHL PQEST PQGLB PQQKQ PQUKI PRINS PTHSS AKY GOX |
DOI | 10.48550/arxiv.2010.00178 |
DatabaseName | ProQuest SciTech Collection ProQuest Technology Collection Materials Science & Engineering Collection ProQuest Central (Alumni Edition) ProQuest Central UK/Ireland ProQuest Central Essentials ProQuest Central Technology Collection ProQuest One Community College ProQuest Central Korea SciTech Premium Collection ProQuest Engineering Collection Engineering Database Proquest Central Premium ProQuest One Academic (New) Publicly Available Content Database ProQuest One Academic Middle East (New) ProQuest One Academic Eastern Edition (DO NOT USE) ProQuest One Applied & Life Sciences ProQuest One Academic ProQuest One Academic UKI Edition ProQuest Central China Engineering collection arXiv Computer Science arXiv.org |
DatabaseTitle | Publicly Available Content Database Engineering Database Technology Collection ProQuest One Academic Middle East (New) ProQuest Central Essentials ProQuest One Academic Eastern Edition ProQuest Central (Alumni Edition) SciTech Premium Collection ProQuest One Community College ProQuest Technology Collection ProQuest SciTech Collection ProQuest Central China ProQuest Central ProQuest One Applied & Life Sciences ProQuest Engineering Collection ProQuest One Academic UKI Edition ProQuest Central Korea Materials Science & Engineering Collection ProQuest Central (New) ProQuest One Academic ProQuest One Academic (New) Engineering Collection |
DatabaseTitleList | Publicly Available Content Database |
Database_xml | – sequence: 1 dbid: GOX name: arXiv.org url: http://arxiv.org/find sourceTypes: Open Access Repository – sequence: 2 dbid: 8FG name: ProQuest Technology Collection url: https://search.proquest.com/technologycollection1 sourceTypes: Aggregation Database |
DeliveryMethod | fulltext_linktorsrc |
Discipline | Physics |
EISSN | 2331-8422 |
ExternalDocumentID | 2010_00178 |
Genre | Working Paper/Pre-Print |
GroupedDBID | 8FE 8FG ABJCF ABUWG AFKRA ALMA_UNASSIGNED_HOLDINGS AZQEC BENPR BGLVJ CCPQU DWQXO FRJ HCIFZ L6V M7S M~E PHGZM PHGZT PIMPY PKEHL PQEST PQGLB PQQKQ PQUKI PRINS PTHSS AKY GOX |
ID | FETCH-LOGICAL-a526-41b238e6218bc5b16e5206392e7c4a7c3a13710673791f23569518bec030bcf53 |
IEDL.DBID | 8FG |
IngestDate | Tue Jul 22 21:57:09 EDT 2025 Mon Jun 30 09:17:19 EDT 2025 |
IsDoiOpenAccess | true |
IsOpenAccess | true |
IsPeerReviewed | false |
IsScholarly | false |
Language | English |
LinkModel | DirectLink |
MergedId | FETCHMERGED-LOGICAL-a526-41b238e6218bc5b16e5206392e7c4a7c3a13710673791f23569518bec030bcf53 |
Notes | SourceType-Working Papers-1 ObjectType-Working Paper/Pre-Print-1 content type line 50 |
OpenAccessLink | https://www.proquest.com/docview/2448036766?pq-origsite=%requestingapplication% |
PQID | 2448036766 |
PQPubID | 2050157 |
ParticipantIDs | arxiv_primary_2010_00178 proquest_journals_2448036766 |
PublicationCentury | 2000 |
PublicationDate | 20210104 |
PublicationDateYYYYMMDD | 2021-01-04 |
PublicationDate_xml | – month: 01 year: 2021 text: 20210104 day: 04 |
PublicationDecade | 2020 |
PublicationPlace | Ithaca |
PublicationPlace_xml | – name: Ithaca |
PublicationTitle | arXiv.org |
PublicationYear | 2021 |
Publisher | Cornell University Library, arXiv.org |
Publisher_xml | – name: Cornell University Library, arXiv.org |
SSID | ssj0002672553 |
Score | 1.7495588 |
SecondaryResourceType | preprint |
Snippet | Applications of machine learning are subject to three major components that contribute to the final performance metrics. Within the category of neural... Applications of machine learning are subject to three major components that contribute to the final performance metrics. Within the category of neural... |
SourceID | arxiv proquest |
SourceType | Open Access Repository Aggregation Database |
SubjectTerms | Computer Science - Learning Data augmentation Deep learning Machine learning Neural networks Performance measurement Training |
SummonAdditionalLinks | – databaseName: arXiv.org dbid: GOX link: http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwdV09T8MwED2VTiwIBKiFgjywRtSJP5KxIpQKCZBQkbpF_krVgbTql-Dfc3ZcMSBW6zzk2c57Tu7eAdxpmxUupyyxUuIFRRc2KbgdJjSTTBhrrQhFYS-vYvLBnmd81gFyqIVR66_FvvUH1pv7mHlFZX7kaxX9rn16m7U_J4MVV4z_jUONGYb-vFoDX4xP4SQKPTJqV-YMOq45h3IaOzKQUm0VGe3mn7H2pyGoHknp3IpEy9M5eVd2sSTjdZvt_E2iu_gFTMeP04dJEvsYJIqnImFUIy86gWSqDddUOJ56YZA6aZiSJlMIDA0NYwpapxkXqHpyxBbPnzY1zy6h2ywb1wNCa4OSjClDhwZvYkqn3EjlaF2rzBnG-tALT1-tWquKygNTBWD6MDgAUsVtuqmQ2_Oh92wTV__PvIbj1Cdy-O8ObADd7XrnbpCJt_o2LMcP3s2HBQ priority: 102 providerName: Cornell University |
Title | Training Data Augmentation for Deep Learning Radio Frequency Systems |
URI | https://www.proquest.com/docview/2448036766 https://arxiv.org/abs/2010.00178 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwfV3PT8IwFG6UxcSbPwOKpAevC7RrO3YyKgxiAhKCCbelv0Y4OOYAoxf_dttS9GDiZcm2097r3nt9_d73AXArVJToLiKhimOzQRGJChOqOiGKYsKkUoq5obDRmA1fyNOczn3Dbe1hlfuY6AK1WknbI2-bNNTtWHoxdle-hVY1yp6uegmNQxAgbFaSnRRPBz89FsxiUzFHu8NMR93V5tXH8n2P6EJWXC1wj_6EYpdf0hMQTHipq1NwoIszcORgmXJ9Dnozr-AAe3zD4f128epnhQpoqk3Y07qEniJ1AadcLVcwrXbo6E_o2cgvwCztzx6Hodc9CDnFLCRImDyqmUm-QlKBmKbYFhJYx5LwWEbcGBI5gZkE5TiizFRJXeML878KmdPoEtSKVaHrAKJcmhKOcIk60uzcuMBUxlyjPOeRloQ0QN19fVbuqC0ya5jMGaYBmnuDZH5Zr7NfJ1z9__oaHGML_rC9CtIEtU211Tcme29Ey7moBYKH_ngyNXeD57m5jr763xMmm54 |
linkProvider | ProQuest |
linkToHtml | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwtV1LTwIxEG6IxOjNZ0BRe9DjBtrtduFAjAoE5BFCMOG26bZdwkFAHio_zv_mtHT1YOKN6_aynXb6zbQz34fQbaz8ii4T5qkwhAQlriivEqiSR_yQcamU4rYprNvjzRf2PApGGfSV9sKYssr0TLQHtZpJc0deBBgqlwy9GL-fv3lGNcq8rqYSGsJJK6iqpRhzjR1tvfmAFG5ZbdVgve8obdSHT03PqQx4IqDcYyQG1NIcoC6WQUy4DqiBbapDyUQofQG_TaycS4Uk1A84xCRlmDl4RywTIxoBCJCFqMMHp8o-1nv9wc8lD-UhhOz-9jXVcocVxeJz8p6WlBGj7pa1n_5ggQW4xhHK9sVcL45RRk9P0L6tC5XLU1QbOgkJXBMrgR_W41fXrDTFEO7imtZz7Dhax3gg1GSGG4ttefYGOzr0MzTchUnO0d50NtU5hEkiIYZkQpKShNRRxDSQodAkSYSvJWN5lLOzj-Zbbo3IGCayhsmjQmqQyPnVMvrdBRf_D9-gg-aw24k6rV77Eh1SU4liLk5YAe2tFmt9BaHEKr52C4ZRtOMt8g07btfA |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Training+Data+Augmentation+for+Deep+Learning+Radio+Frequency+Systems&rft.jtitle=arXiv.org&rft.au=Clark%2C+William+H&rft.au=Hauser%2C+Steven&rft.au=Headley%2C+William+C&rft.au=Michaels%2C+Alan+J&rft.date=2021-01-04&rft.pub=Cornell+University+Library%2C+arXiv.org&rft.eissn=2331-8422&rft_id=info:doi/10.48550%2Farxiv.2010.00178 |