심층신경망으로 가는 통계 여행, 세 번째 여행: 언어모형과 트랜스포머

Over the past decade, the remarkable advancements in deep neural networks have paralleled the development and evolution of language models. Initially, language models were developed in the form of Encoder-Decoder models using early RNNs. However, with the introduction of Attention in 2015 and the em...

Full description

Saved in:
Bibliographic Details
Published inŬngyong tʻonggye yŏnʼgu Vol. 37; no. 5; pp. 567 - 582
Main Authors 김유진(Yu Jin Kim), 황인준(In Jun Hwang), 장기석(Kisuk Jang), 이윤동(Yoon Dong Lee)
Format Journal Article
LanguageKorean
Published 한국통계학회 2024
Subjects
Online AccessGet full text
ISSN1225-066X
2383-5818

Cover

Abstract Over the past decade, the remarkable advancements in deep neural networks have paralleled the development and evolution of language models. Initially, language models were developed in the form of Encoder-Decoder models using early RNNs. However, with the introduction of Attention in 2015 and the emergence of the Transformer in 2017, the field saw revolutionary growth. This study briefly reviews the development process of language models and examines in detail the working mechanism and technical elements of the Transformer. Additionally, it explores statistical models and methodologies related to language models and the Transformer. 지난 10년의 기간 심층신경망의 비약적 발전은 언어모형의 개발과 그 발전을 함께 해 왔다. 언어모형은 초기 RNN을 이용한 encoder-decoder 모형의 형태로 개발되었으나, 2015년 attention이 등장하고, 2017년 transformer가 등장하여 혁명적 기술로 성장하였다. 본 연구에서는 언어모형의 발전과정을 간략하게 살펴보고, 트랜스포머의 작동원리와 기술적 요소에 대하여 구체적으로 살펴본다. 동시에 언어모형, 트랜스포머와 관련되는 통계모형과, 방법론에 대하여 함께 검토한다.
AbstractList Over the past decade, the remarkable advancements in deep neural networks have paralleled the development and evolution of language models. Initially, language models were developed in the form of Encoder-Decoder models using early RNNs. However, with the introduction of Attention in 2015 and the emergence of the Transformer in 2017, the field saw revolutionary growth. This study briefly reviews the development process of language models and examines in detail the working mechanism and technical elements of the Transformer. Additionally, it explores statistical models and methodologies related to language models and the Transformer. 지난 10년의 기간 심층신경망의 비약적 발전은 언어모형의 개발과 그 발전을 함께 해 왔다. 언어모형은 초기 RNN을 이용한 encoder-decoder 모형의 형태로 개발되었으나, 2015년 attention이 등장하고, 2017년 transformer가 등장하여 혁명적 기술로 성장하였다. 본 연구에서는 언어모형의 발전과정을 간략하게 살펴보고, 트랜스포머의 작동원리와 기술적 요소에 대하여 구체적으로 살펴본다. 동시에 언어모형, 트랜스포머와 관련되는 통계모형과, 방법론에 대하여 함께 검토한다.
Author 황인준(In Jun Hwang)
장기석(Kisuk Jang)
김유진(Yu Jin Kim)
이윤동(Yoon Dong Lee)
Author_xml – sequence: 1
  fullname: 김유진(Yu Jin Kim)
– sequence: 2
  fullname: 황인준(In Jun Hwang)
– sequence: 3
  fullname: 장기석(Kisuk Jang)
– sequence: 4
  fullname: 이윤동(Yoon Dong Lee)
BookMark eNpFjD9Lw0AAxQ-pYK39Dre4Gcjd5S4Xt1Lr32KXDk6Gyz-IrVEaHdwcimg7uGh1qKIgtEgHaS1kqF8od_kOFiz4lvf48d5bBbnoLPKXQB4TTjTKEc-BPMKYajpjRyugGMcn-lwMYYNbeXCsuiOVTFX3LR3_yMGL6s_kex-mX9ey8wCzm2k6aUP1NMp6dxtQtRMox7dqkCzQJlS9RPW-5ecwe35MJzOYdRL52ledj-x-JIfJGlgORDP2iwsvgPp2pV7e1aq1nb1yqao1mE61gHNh6gZ1XOQHhHIP64iagUcRRwI5hm8E1HQt4lFKBAm4yxB3ueOapqCcMEwKYP3vthHGF6EdeXHT3i8d1LCODUI4Q4yY881_L7pshae-Fwr7fB5E68o-rG1VELIoYpiSX62jebI
ContentType Journal Article
DBID DBRKI
TDB
JDI
DEWEY 519.5
DatabaseName DBPIA - 디비피아
Nurimedia DBPIA Journals
KoreaScience
DatabaseTitleList
DeliveryMethod fulltext_linktorsrc
Discipline Statistics
Applied Sciences
Mathematics
DocumentTitle_FL A statistical journey to DNN, the third trip: Language model and transformer
EISSN 2383-5818
EndPage 582
ExternalDocumentID JAKO202433861637618
NODE11951625
GroupedDBID 9ZL
ALMA_UNASSIGNED_HOLDINGS
DBRKI
JDI
OK1
TDB
ID FETCH-LOGICAL-k605-f88a7045bc1ef358d20157fd5181a1b4e4f57c93d553a3f8c618c8bc77a583623
ISSN 1225-066X
IngestDate Thu Dec 12 02:21:50 EST 2024
Thu Feb 06 13:29:46 EST 2025
IsOpenAccess true
IsPeerReviewed false
IsScholarly true
Issue 5
Keywords 인코더-디코더
multi-head attention
다중어텐션
언어모형
transformer
positional encoding
트랜스포머
위치인코딩
language model
encoder-decoder
-
Language Korean
LinkModel OpenURL
MergedId FETCHMERGED-LOGICAL-k605-f88a7045bc1ef358d20157fd5181a1b4e4f57c93d553a3f8c618c8bc77a583623
Notes KISTI1.1003/JNL.JAKO202433861637618
OpenAccessLink http://click.ndsl.kr/servlet/LinkingDetailView?cn=JAKO202433861637618&dbt=JAKO&org_code=O481&site_code=SS1481&service_code=01
PageCount 16
ParticipantIDs kisti_ndsl_JAKO202433861637618
nurimedia_primary_NODE11951625
PublicationCentury 2000
PublicationDate 2024
PublicationDateYYYYMMDD 2024-01-01
PublicationDate_xml – year: 2024
  text: 2024
PublicationDecade 2020
PublicationTitle Ŭngyong tʻonggye yŏnʼgu
PublicationTitleAlternate The Korean journal of applied statistics
PublicationYear 2024
Publisher 한국통계학회
Publisher_xml – name: 한국통계학회
SSID ssj0000612489
ssib053377530
ssib001150021
ssib044750966
ssib022238561
Score 2.2580338
Snippet Over the past decade, the remarkable advancements in deep neural networks have paralleled the development and evolution of language models. Initially, language...
SourceID kisti
nurimedia
SourceType Open Access Repository
Publisher
StartPage 567
Title 심층신경망으로 가는 통계 여행, 세 번째 여행: 언어모형과 트랜스포머
URI https://www.dbpia.co.kr/journal/articleDetail?nodeId=NODE11951625
http://click.ndsl.kr/servlet/LinkingDetailView?cn=JAKO202433861637618&dbt=JAKO&org_code=O481&site_code=SS1481&service_code=01
Volume 37
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
journalDatabaseRights – providerCode: PRVHPJ
  databaseName: ROAD: Directory of Open Access Scholarly Resources
  customDbUrl:
  eissn: 2383-5818
  dateEnd: 99991231
  omitProxy: true
  ssIdentifier: ssib044750966
  issn: 1225-066X
  databaseCode: M~E
  dateStart: 20130101
  isFulltext: true
  titleUrlDefault: https://road.issn.org
  providerName: ISSN International Centre
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwnR1Na9RANNRerAfxEz9LDs4pRrKbTDLjLR9b6pa2lwr14pLPUiqp2F2kHsRDEW0PXrR6qKIgtJQepLWwh_qHdrP_wfcm2d1gC1Yvy-yb9-Z9JZn3hpl5knQniPRqrOmGGnI_VI2oEqqcRVTVaRJGQVLhsY8HnKdnzMmHRn2ezo-c2S3tWmo1g3vhixPPlfyPVwEGfsVTsv_g2cGgAIA2-Bd-wcPweyofk5pLmENsFxsOIw7tQzRSs4lTJY5HavDXItzDLg5oroBUoK0IJA03OwCI2YQbAPIIM8VI0KcThiAgtAQXj3CTMI5-EZwM4KogMbBiDEHASoCOkeh2DjVR0KJhCFFsYjOBxwjvs3XcXBJbYDuE11B2ZAnYhuiaEMM7SOywcoxNXApd6cIqllFqEhfGcKC5sBorq9jHJlIBdBda_cdNcOXCEMJItlboAhpW2aOWUl9MFVF4mg9JQGJObC5IvEIrkA03U7IHqVJvpcrkc7EcXyJyURU7V5Oh8XMzwgBVNrW40lpS6ieQeMJWuWjCaKA9MAfRluHt9VBRsaOKlxdxqsPlWyEsFSa00be2d9zPfRwuGjZhbmnCgu9xvxokzOcCBlGYrlJWzGvFLJdfrVO8zbQ0ZdG8HEoR_dC8FNQfd5DPzHo1vBuwAtkyhDsQ1GKVlOmXtWHkDYmFNowEMepkpSPTeK2kVsqsIcmwIFHWBsuhGGkbomDlQCNIFzGHWjwnnU1bWOsCPpil2G_ugnS-SNpkO38DL0ojS8uXpDHM0_Jrzi9Lj7ONvax9mG187ez_6m5_zraOut-25M6PV93193Lv9WHnYE3OPu71Nt_elbO1ttzdf5NttwvQfTnbbGebP7u7O71PHzoHR3Jvvd39spWtf--92-vutK9IcxO1OXdSLUqXqEumRtWEMd-CZCkIK3GiUxZBmE2tJKIQT_uVwIiNhFoh1yNKdV9PWGhWWMiC0LJ8yiCk1K9Ko-lyGl-TZL8aGVrkUyMy8WpGgyV4fZYWJ2juODCuS-PCSo00WnnSqNtTs_h86TozIdOyYFxAGJiv8TS_wqZRduiNvyHclMZwzHzh8ZY02nzWim9DKN4MxsVD8BsQtrsD
linkProvider ISSN International Centre
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=%EC%8B%AC%EC%B8%B5%EC%8B%A0%EA%B2%BD%EB%A7%9D%EC%9C%BC%EB%A1%9C+%EA%B0%80%EB%8A%94+%ED%86%B5%EA%B3%84+%EC%97%AC%ED%96%89%2C+%EC%84%B8+%EB%B2%88%EC%A7%B8+%EC%97%AC%ED%96%89%3A+%EC%96%B8%EC%96%B4%EB%AA%A8%ED%98%95%EA%B3%BC+%ED%8A%B8%EB%9E%9C%EC%8A%A4%ED%8F%AC%EB%A8%B8&rft.jtitle=%C5%ACngyong+t%CA%BBonggye+y%C5%8Fn%CA%BCgu&rft.au=%EA%B9%80%EC%9C%A0%EC%A7%84%28Yu+Jin+Kim%29&rft.au=%ED%99%A9%EC%9D%B8%EC%A4%80%28In+Jun+Hwang%29&rft.au=%EC%9E%A5%EA%B8%B0%EC%84%9D%28Kisuk+Jang%29&rft.au=%EC%9D%B4%EC%9C%A4%EB%8F%99%28Yoon+Dong+Lee%29&rft.date=2024&rft.pub=%ED%95%9C%EA%B5%AD%ED%86%B5%EA%B3%84%ED%95%99%ED%9A%8C&rft.issn=1225-066X&rft.eissn=2383-5818&rft.volume=37&rft.issue=5&rft.spage=567&rft.epage=582&rft.externalDocID=NODE11951625
thumbnail_l http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/lc.gif&issn=1225-066X&client=summon
thumbnail_m http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/mc.gif&issn=1225-066X&client=summon
thumbnail_s http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/sc.gif&issn=1225-066X&client=summon