심층신경망으로 가는 통계 여행, 세 번째 여행: 언어모형과 트랜스포머
Over the past decade, the remarkable advancements in deep neural networks have paralleled the development and evolution of language models. Initially, language models were developed in the form of Encoder-Decoder models using early RNNs. However, with the introduction of Attention in 2015 and the em...
Saved in:
Published in | Ŭngyong tʻonggye yŏnʼgu Vol. 37; no. 5; pp. 567 - 582 |
---|---|
Main Authors | , , , |
Format | Journal Article |
Language | Korean |
Published |
한국통계학회
2024
|
Subjects | |
Online Access | Get full text |
ISSN | 1225-066X 2383-5818 |
Cover
Abstract | Over the past decade, the remarkable advancements in deep neural networks have paralleled the development and evolution of language models. Initially, language models were developed in the form of Encoder-Decoder models using early RNNs. However, with the introduction of Attention in 2015 and the emergence of the Transformer in 2017, the field saw revolutionary growth. This study briefly reviews the development process of language models and examines in detail the working mechanism and technical elements of the Transformer. Additionally, it explores statistical models and methodologies related to language models and the Transformer. 지난 10년의 기간 심층신경망의 비약적 발전은 언어모형의 개발과 그 발전을 함께 해 왔다. 언어모형은 초기 RNN을 이용한 encoder-decoder 모형의 형태로 개발되었으나, 2015년 attention이 등장하고, 2017년 transformer가 등장하여 혁명적 기술로 성장하였다. 본 연구에서는 언어모형의 발전과정을 간략하게 살펴보고, 트랜스포머의 작동원리와 기술적 요소에 대하여 구체적으로 살펴본다. 동시에 언어모형, 트랜스포머와 관련되는 통계모형과, 방법론에 대하여 함께 검토한다. |
---|---|
AbstractList | Over the past decade, the remarkable advancements in deep neural networks have paralleled the development and evolution of language models. Initially, language models were developed in the form of Encoder-Decoder models using early RNNs. However, with the introduction of Attention in 2015 and the emergence of the Transformer in 2017, the field saw revolutionary growth. This study briefly reviews the development process of language models and examines in detail the working mechanism and technical elements of the Transformer. Additionally, it explores statistical models and methodologies related to language models and the Transformer. 지난 10년의 기간 심층신경망의 비약적 발전은 언어모형의 개발과 그 발전을 함께 해 왔다. 언어모형은 초기 RNN을 이용한 encoder-decoder 모형의 형태로 개발되었으나, 2015년 attention이 등장하고, 2017년 transformer가 등장하여 혁명적 기술로 성장하였다. 본 연구에서는 언어모형의 발전과정을 간략하게 살펴보고, 트랜스포머의 작동원리와 기술적 요소에 대하여 구체적으로 살펴본다. 동시에 언어모형, 트랜스포머와 관련되는 통계모형과, 방법론에 대하여 함께 검토한다. |
Author | 황인준(In Jun Hwang) 장기석(Kisuk Jang) 김유진(Yu Jin Kim) 이윤동(Yoon Dong Lee) |
Author_xml | – sequence: 1 fullname: 김유진(Yu Jin Kim) – sequence: 2 fullname: 황인준(In Jun Hwang) – sequence: 3 fullname: 장기석(Kisuk Jang) – sequence: 4 fullname: 이윤동(Yoon Dong Lee) |
BookMark | eNpFjD9Lw0AAxQ-pYK39Dre4Gcjd5S4Xt1Lr32KXDk6Gyz-IrVEaHdwcimg7uGh1qKIgtEgHaS1kqF8od_kOFiz4lvf48d5bBbnoLPKXQB4TTjTKEc-BPMKYajpjRyugGMcn-lwMYYNbeXCsuiOVTFX3LR3_yMGL6s_kex-mX9ey8wCzm2k6aUP1NMp6dxtQtRMox7dqkCzQJlS9RPW-5ecwe35MJzOYdRL52ledj-x-JIfJGlgORDP2iwsvgPp2pV7e1aq1nb1yqao1mE61gHNh6gZ1XOQHhHIP64iagUcRRwI5hm8E1HQt4lFKBAm4yxB3ueOapqCcMEwKYP3vthHGF6EdeXHT3i8d1LCODUI4Q4yY881_L7pshae-Fwr7fB5E68o-rG1VELIoYpiSX62jebI |
ContentType | Journal Article |
DBID | DBRKI TDB JDI |
DEWEY | 519.5 |
DatabaseName | DBPIA - 디비피아 Nurimedia DBPIA Journals KoreaScience |
DatabaseTitleList | |
DeliveryMethod | fulltext_linktorsrc |
Discipline | Statistics Applied Sciences Mathematics |
DocumentTitle_FL | A statistical journey to DNN, the third trip: Language model and transformer |
EISSN | 2383-5818 |
EndPage | 582 |
ExternalDocumentID | JAKO202433861637618 NODE11951625 |
GroupedDBID | 9ZL ALMA_UNASSIGNED_HOLDINGS DBRKI JDI OK1 TDB |
ID | FETCH-LOGICAL-k605-f88a7045bc1ef358d20157fd5181a1b4e4f57c93d553a3f8c618c8bc77a583623 |
ISSN | 1225-066X |
IngestDate | Thu Dec 12 02:21:50 EST 2024 Thu Feb 06 13:29:46 EST 2025 |
IsOpenAccess | true |
IsPeerReviewed | false |
IsScholarly | true |
Issue | 5 |
Keywords | 인코더-디코더 multi-head attention 다중어텐션 언어모형 transformer positional encoding 트랜스포머 위치인코딩 language model encoder-decoder - |
Language | Korean |
LinkModel | OpenURL |
MergedId | FETCHMERGED-LOGICAL-k605-f88a7045bc1ef358d20157fd5181a1b4e4f57c93d553a3f8c618c8bc77a583623 |
Notes | KISTI1.1003/JNL.JAKO202433861637618 |
OpenAccessLink | http://click.ndsl.kr/servlet/LinkingDetailView?cn=JAKO202433861637618&dbt=JAKO&org_code=O481&site_code=SS1481&service_code=01 |
PageCount | 16 |
ParticipantIDs | kisti_ndsl_JAKO202433861637618 nurimedia_primary_NODE11951625 |
PublicationCentury | 2000 |
PublicationDate | 2024 |
PublicationDateYYYYMMDD | 2024-01-01 |
PublicationDate_xml | – year: 2024 text: 2024 |
PublicationDecade | 2020 |
PublicationTitle | Ŭngyong tʻonggye yŏnʼgu |
PublicationTitleAlternate | The Korean journal of applied statistics |
PublicationYear | 2024 |
Publisher | 한국통계학회 |
Publisher_xml | – name: 한국통계학회 |
SSID | ssj0000612489 ssib053377530 ssib001150021 ssib044750966 ssib022238561 |
Score | 2.2580338 |
Snippet | Over the past decade, the remarkable advancements in deep neural networks have paralleled the development and evolution of language models. Initially, language... |
SourceID | kisti nurimedia |
SourceType | Open Access Repository Publisher |
StartPage | 567 |
Title | 심층신경망으로 가는 통계 여행, 세 번째 여행: 언어모형과 트랜스포머 |
URI | https://www.dbpia.co.kr/journal/articleDetail?nodeId=NODE11951625 http://click.ndsl.kr/servlet/LinkingDetailView?cn=JAKO202433861637618&dbt=JAKO&org_code=O481&site_code=SS1481&service_code=01 |
Volume | 37 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
journalDatabaseRights | – providerCode: PRVHPJ databaseName: ROAD: Directory of Open Access Scholarly Resources customDbUrl: eissn: 2383-5818 dateEnd: 99991231 omitProxy: true ssIdentifier: ssib044750966 issn: 1225-066X databaseCode: M~E dateStart: 20130101 isFulltext: true titleUrlDefault: https://road.issn.org providerName: ISSN International Centre |
link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwnR1Na9RANNRerAfxEz9LDs4pRrKbTDLjLR9b6pa2lwr14pLPUiqp2F2kHsRDEW0PXrR6qKIgtJQepLWwh_qHdrP_wfcm2d1gC1Yvy-yb9-Z9JZn3hpl5knQniPRqrOmGGnI_VI2oEqqcRVTVaRJGQVLhsY8HnKdnzMmHRn2ezo-c2S3tWmo1g3vhixPPlfyPVwEGfsVTsv_g2cGgAIA2-Bd-wcPweyofk5pLmENsFxsOIw7tQzRSs4lTJY5HavDXItzDLg5oroBUoK0IJA03OwCI2YQbAPIIM8VI0KcThiAgtAQXj3CTMI5-EZwM4KogMbBiDEHASoCOkeh2DjVR0KJhCFFsYjOBxwjvs3XcXBJbYDuE11B2ZAnYhuiaEMM7SOywcoxNXApd6cIqllFqEhfGcKC5sBorq9jHJlIBdBda_cdNcOXCEMJItlboAhpW2aOWUl9MFVF4mg9JQGJObC5IvEIrkA03U7IHqVJvpcrkc7EcXyJyURU7V5Oh8XMzwgBVNrW40lpS6ieQeMJWuWjCaKA9MAfRluHt9VBRsaOKlxdxqsPlWyEsFSa00be2d9zPfRwuGjZhbmnCgu9xvxokzOcCBlGYrlJWzGvFLJdfrVO8zbQ0ZdG8HEoR_dC8FNQfd5DPzHo1vBuwAtkyhDsQ1GKVlOmXtWHkDYmFNowEMepkpSPTeK2kVsqsIcmwIFHWBsuhGGkbomDlQCNIFzGHWjwnnU1bWOsCPpil2G_ugnS-SNpkO38DL0ojS8uXpDHM0_Jrzi9Lj7ONvax9mG187ez_6m5_zraOut-25M6PV93193Lv9WHnYE3OPu71Nt_elbO1ttzdf5NttwvQfTnbbGebP7u7O71PHzoHR3Jvvd39spWtf--92-vutK9IcxO1OXdSLUqXqEumRtWEMd-CZCkIK3GiUxZBmE2tJKIQT_uVwIiNhFoh1yNKdV9PWGhWWMiC0LJ8yiCk1K9Ko-lyGl-TZL8aGVrkUyMy8WpGgyV4fZYWJ2juODCuS-PCSo00WnnSqNtTs_h86TozIdOyYFxAGJiv8TS_wqZRduiNvyHclMZwzHzh8ZY02nzWim9DKN4MxsVD8BsQtrsD |
linkProvider | ISSN International Centre |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=%EC%8B%AC%EC%B8%B5%EC%8B%A0%EA%B2%BD%EB%A7%9D%EC%9C%BC%EB%A1%9C+%EA%B0%80%EB%8A%94+%ED%86%B5%EA%B3%84+%EC%97%AC%ED%96%89%2C+%EC%84%B8+%EB%B2%88%EC%A7%B8+%EC%97%AC%ED%96%89%3A+%EC%96%B8%EC%96%B4%EB%AA%A8%ED%98%95%EA%B3%BC+%ED%8A%B8%EB%9E%9C%EC%8A%A4%ED%8F%AC%EB%A8%B8&rft.jtitle=%C5%ACngyong+t%CA%BBonggye+y%C5%8Fn%CA%BCgu&rft.au=%EA%B9%80%EC%9C%A0%EC%A7%84%28Yu+Jin+Kim%29&rft.au=%ED%99%A9%EC%9D%B8%EC%A4%80%28In+Jun+Hwang%29&rft.au=%EC%9E%A5%EA%B8%B0%EC%84%9D%28Kisuk+Jang%29&rft.au=%EC%9D%B4%EC%9C%A4%EB%8F%99%28Yoon+Dong+Lee%29&rft.date=2024&rft.pub=%ED%95%9C%EA%B5%AD%ED%86%B5%EA%B3%84%ED%95%99%ED%9A%8C&rft.issn=1225-066X&rft.eissn=2383-5818&rft.volume=37&rft.issue=5&rft.spage=567&rft.epage=582&rft.externalDocID=NODE11951625 |
thumbnail_l | http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/lc.gif&issn=1225-066X&client=summon |
thumbnail_m | http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/mc.gif&issn=1225-066X&client=summon |
thumbnail_s | http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/sc.gif&issn=1225-066X&client=summon |