서브 헤더

공지사항

공지사항

행사세미나 전문가 초청 세미나 개최( Prof. Keisuke Sakaguchi @ Tohoku University)

페이지 정보

profile_image
작성자 관리자 댓글 0건 조회 937회 작성일 23.03.27

본문

전문가 초청 세미나 개최 안내드리오니 많은 관심과 참여 부탁드립니다. 감사합니다. 


Title: Large Language Models: What's Happening Now?


Speaker: Prof. Keisuke Sakaguchi @ Tohoku University


Time : 2023 April 5th 14:00 ~ 15:00


Location: Online

Online: https://us02web.zoom.us/j/87084697053?pwd=U1ZqUXpJMHczcW1pKzJXck01NHZhUT09 (Passcode: 0405)


Language: English speech & English slides


Abstract

Large Language Models (LLMs) such as BERT, GPT-3, and the recently introduced GPT-4 have ushered in a new era in natural language processing, enabling applications like machine translation, summarization, proofreading, and chatbot dialogues across diverse industries and start-ups. Powered by the Encoder-Decoder (Transformer) architecture and self-attention mechanisms, LLMs harness vast amounts of web data to achieve remarkable results. GPT-4 has introduced advancements in instruction fine-tuning and reinforcement learning with human feedback (RLHF), further enhancing its capabilities. In this presentation, we will explore the current landscape of LLMs, investigating their inner workings, recent advancements, and the implementation of these methods. Additionally, we will discuss their limitations and identify emerging research directions, incorporating insights from our recent projects in the field. (By the way, a portion of this abstract was crafted by GPT-4. Can you identify which part?)


Bio:

Keisuke Sakaguchi is an associate professor at Tohoku University. His research interests lie at the intersection of Natural Language Processing, Machine Learning, and Psycholinguistics. The long-term research goals are 1) to understand human intelligence, especially natural langauge processing, and 2) to build embodied AI that is as robust and efficient as humans. More specifically, his research interests include: robust NLP models for noisy texts (e.g., text normalization, parsing, automated grammatical error correction), commonsense knowledge acquisition and reasoning, NLP for educational purposes (first and second language acquisition), text generation (decoding algorithms), efficient data collection, and reliable evaluation metric design (meta-evaluation).