About
Yeon-Soo Lim
WORK EXPERIENCE
- AIR Lab@42dot(Hyundai Global Software Center) (2022.09 ~ )
- AIR Lab@Hyundai Motor Company (2022.03 ~ 2022.09)
Education
- M.S., in Computer Engineering, Kyung Hee University (2020.03 ~ 2022.02)
- B.S., in Computer Engineering, Kyung Hee University (2014.03 ~ 2020.02)
Technical Skills
- Hand-on experience in applying PyTorch, NLP tools (NLTK, etc.) and open libraries (huggingface, fairseq, etc.) for Korean NLP tasks
- Experience in several NLP tasks with Non-autoregressive decoding, text summarization and Natural Language Understanding
- Build ML pipeline system use Airflow, MLflow, Docker and kubernetes
- Python: Skillful, C++: Middle level
Research Experience
Non-Autoregressive Machine Translation (2020.10 ~ 2022.01)
- Research Non-Autoregressive Decoding technology for machine translation
- Proposes a novel machine translation model which changes the length of a target sentence iteratively to an optimal length
- Propsed model expand or delete token to find optimal target sentence length
Korean single and multi-document summary (2020.03 ~ 2021.12)
- Develop a topic-centric text summarization model with statistical attribute model. Apply PPLM to the text summarization for generate a topic-centric summary.
- Develop a text summarization model using BART
- Propose post-processing method to merge semantically similar summaries for multi-document summarization.
Korean dialogue model based on deep learning (2020.03 ~ 2020.12)
- Research on dialogue generation technology in the form of a single model for multiple domains. Using Continual Learning, a single conversational model can respond to multiple domains.
- Develop a conversation model using seq2seq structure and GPT2 that reflects the keyword score of tokens measured by the keyword extractor.
Publications
International Journal Articles
- Yeon-Soo Lim, Eun-Ju Park, Hyun-Je Song and Seong-Bae Park. A Non-Autoregressive Neural Machine Translation Model with Iterative Length Update of Target Sentence, IEEE Access, vol. 10, pp.43341-43350, 2022.
- So-Eon Kim, Yeon-Soo Lim and Seong-Bae Parkk. Strong Influence of Responses in Training Dialogue Response Generator, Applied Sciences, 11(16), 7415, 2021.
Domestic Conference Papers
- Yeon-Soo Lim, BongMin Kim, Choong Seon Hong, and Seong-Bae Park. 2021. Hate Speech Detection Model using Noise Self-training. In Proceedings of the Korea Computer Congress 2021, pp. 376-378, 2021.
- Sunggoo Kwon, Yeon-Soo Lim, and Seong-Bae Park. The Study on Korean Single Document Summarization Using Pre-trained Language Models. In Proceedings of the Korea Computer Congress 2021, pp. 379-381, 2021.
- Yeon-Soo Lim, So-Eon Kim, Bong-Min Kim, Heejae Jung, and Seong-Bae Park. A Query-aware Dialog Model for Open-domain Dialog. In Proceedings of the 32th Annual Conference on Human and Cognitive Language Technology, pp. 274-279, 2020.
Patents
Domestic Patent (in Korean)
- Cheoneum Park, Yeonsoo Lim. Method, Apparatus, and Computer-readable Medium for Error Correction on Document Summarization. 2023/03, (application umber) 10-2023-0033520.