• Title/Summary/Keyword: LLM

Search Result 79, Processing Time 0.025 seconds

Automation of M.E.P Design Using Large Language Models (대형 언어 모델을 활용한 설비설계의 자동화)

  • Park, Kyung Kyu;Lee, Seung-Been;Seo, Min Jo;Kim, Si Uk;Choi, Won Jun;Kim, Chee Kyung
    • Proceedings of the Korean Institute of Building Construction Conference
    • /
    • 2023.11a
    • /
    • pp.237-238
    • /
    • 2023
  • Urbanization and the increase in building scale have amplified the complexity of M.E.P design. Traditional design methods face limitations when considering intricate pathways and variables, leading to an emergent need for research in automated design. Initial algorithmic approaches encountered challenges in addressing complex architectural structures and the diversity of M.E.P types. However, with the launch of OpenAI's ChatGPT-3.5 beta version in 2022, new opportunities in the automated design sector were unlocked. ChatGPT, based on the Large Language Model (LLM), has the capability to deeply comprehend the logical structures and meanings within training data. This study analyzed the potential application and latent value of LLMs in M.E.P design. Ultimately, the implementation of LLM in M.E.P design will make genuine automated design feasible, which is anticipated to drive advancements across designs in the construction sector.

  • PDF

A Proposal of Evaluation of Large Language Models Built Based on Research Data (연구데이터 관점에서 본 거대언어모델 품질 평가 기준 제언)

  • Na-eun Han;Sujeong Seo;Jung-ho Um
    • Journal of the Korean Society for information Management
    • /
    • v.40 no.3
    • /
    • pp.77-98
    • /
    • 2023
  • Large Language Models (LLMs) are becoming the major trend in the natural language processing field. These models were built based on research data, but information such as types, limitations, and risks of using research data are unknown. This research would present how to analyze and evaluate the LLMs that were built with research data: LLaMA or LLaMA base models such as Alpaca of Stanford, Vicuna of the large model systems organization, and ChatGPT from OpenAI from the perspective of research data. This quality evaluation focuses on the validity, functionality, and reliability of Data Quality Management (DQM). Furthermore, we adopted the Holistic Evaluation of Language Models (HELM) to understand its evaluation criteria and then discussed its limitations. This study presents quality evaluation criteria for LLMs using research data and future development directions.

Development of a Regulatory Q&A System for KAERI Utilizing Document Search Algorithms and Large Language Model (거대언어모델과 문서검색 알고리즘을 활용한 한국원자력연구원 규정 질의응답 시스템 개발)

  • Hongbi Kim;Yonggyun Yu
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.28 no.5
    • /
    • pp.31-39
    • /
    • 2023
  • The evolution of Natural Language Processing (NLP) and the rise of large language models (LLM) like ChatGPT have paved the way for specialized question-answering (QA) systems tailored to specific domains. This study outlines a system harnessing the power of LLM in conjunction with document search algorithms to interpret and address user inquiries using documents from the Korea Atomic Energy Research Institute (KAERI). Initially, the system refines multiple documents for optimized search and analysis, breaking the content into managable paragraphs suitable for the language model's processing. Each paragraph's content is converted into a vector via an embedding model and archived in a database. Upon receiving a user query, the system matches the extracted vectors from the question with the stored vectors, pinpointing the most pertinent content. The chosen paragraphs, combined with the user's query, are then processed by the language generation model to formulate a response. Tests encompassing a spectrum of questions verified the system's proficiency in discerning question intent, understanding diverse documents, and delivering rapid and precise answers.

Enhancing Empathic Reasoning of Large Language Models Based on Psychotherapy Models for AI-assisted Social Support (인공지능 기반 사회적 지지를 위한 대형언어모형의 공감적 추론 향상: 심리치료 모형을 중심으로)

  • Yoon Kyung Lee;Inju Lee;Minjung Shin;Seoyeon Bae;Sowon Hahn
    • Korean Journal of Cognitive Science
    • /
    • v.35 no.1
    • /
    • pp.23-48
    • /
    • 2024
  • Building human-aligned artificial intelligence (AI) for social support remains challenging despite the advancement of Large Language Models. We present a novel method, the Chain of Empathy (CoE) prompting, that utilizes insights from psychotherapy to induce LLMs to reason about human emotional states. This method is inspired by various psychotherapy approaches-Cognitive-Behavioral Therapy (CBT), Dialectical Behavior Therapy (DBT), Person-Centered Therapy (PCT), and Reality Therapy (RT)-each leading to different patterns of interpreting clients' mental states. LLMs without CoE reasoning generated predominantly exploratory responses. However, when LLMs used CoE reasoning, we found a more comprehensive range of empathic responses aligned with each psychotherapy model's different reasoning patterns. For empathic expression classification, the CBT-based CoE resulted in the most balanced classification of empathic expression labels and the text generation of empathic responses. However, regarding emotion reasoning, other approaches like DBT and PCT showed higher performance in emotion reaction classification. We further conducted qualitative analysis and alignment scoring of each prompt-generated output. The findings underscore the importance of understanding the emotional context and how it affects human-AI communication. Our research contributes to understanding how psychotherapy models can be incorporated into LLMs, facilitating the development of context-aware, safe, and empathically responsive AI.

Exploring the Relationship Between Machine and Human Performance in Natural Language Processing Tasks (자연어 처리 태스크에 대한 기계와 인간의 성능 상관관계 연구)

  • Seoyoon Park;Heejae Kim;Seong-Woo Lee;Yejee Kang;Yeonji Jang;Hansaem Kim
    • Annual Conference on Human and Language Technology
    • /
    • 2023.10a
    • /
    • pp.485-490
    • /
    • 2023
  • 언어 모델 발전에 따라 사람과 유사하게 글을 생성하고 태스크를 수행하는 LLM들이 등장하고 있다. 하지만 아직까지도 기계와 사람의 수행 과정에 초점을 맞추어 차이점을 드러내는 연구는 활성화되지 않았다. 본 연구는 자연어 이해 및 생성 태스크 수행 시 기계와 인간의 수행 과정 차이를 밝히고자 하였다. 이에 이해 태스크로는 문법성 판단, 생성 태스크로는 요약 태스크를 대상 태스크로 선정하였고, 기존 주류 사전학습 모델이었던 transformer 계열 모델과 LLM인 ChatGPT 3.5를 사용하여 실험을 진행하였다. 실험 결과 문법성 판단 시 기계들이 인간의 언어적 직관을 반영하지 못하는 양상을 발견하였고, 요약 태스크에서는 인간과 기계의 성능 판단 기준이 다름을 확인하였다.

  • PDF

Hallucination Detection for Generative Large Language Models Exploiting Consistency and Fact Checking Technique (생성형 거대 언어 모델에서 일관성 확인 및 사실 검증을 활 용한 Hallucination 검출 기법)

  • Myeong Jin;Gun-Woo Kim
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2023.11a
    • /
    • pp.461-464
    • /
    • 2023
  • 최근 GPT-3 와 LLaMa 같은 생성형 거대 언어모델을 활용한 서비스가 공개되었고, 실제로 많은 사람들이 사용하고 있다. 해당 모델들은 사용자들의 다양한 질문에 대해 유창한 답변을 한다는 이유로 주목받고 있다. 하지만 LLMs 의 답변에는 종종 Inconsistent content 와 non-factual statement 가 존재하며, 이는 사용자들로 하여금 잘못된 정보의 전파 등의 문제를 야기할 수 있다. 이에 논문에서는 동일한 질문에 대한 LLM 의 답변 샘플과 외부 지식을 활용한 Hallucination Detection 방법을 제안한다. 제안한 방법은 동일한 질문에 대한 LLM 의 답변들을 이용해 일관성 점수(Consistency score)를 계산한다. 거기에 외부 지식을 이용한 사실검증을 통해 사실성 점수(Factuality score)를 계산한다. 계산된 일관성 점수와 사실성 점수를 활용하여 문장 수준의 Hallucination Detection 을 가능하게 했다. 실험에는 GPT-3 를 이용하여 WikiBio dataset 에 있는 인물에 대한 passage 를 생성한 데이터셋을 사용하였으며, 우리는 해당 방법을 통해 문장 수준에서의 Hallucination Detection 성능이 baseline 보다 AUC-PR scores 에서 향상됨을 보였다.

Analysis of Discriminatory Patterns in Performing Arts Recognized by Large Language Models (LLMs): Focused on ChatGPT (거대언어모델(LLM)이 인식하는 공연예술의 차별 양상 분석: ChatGPT를 중심으로)

  • Jiae Choi
    • Journal of Intelligence and Information Systems
    • /
    • v.29 no.3
    • /
    • pp.401-418
    • /
    • 2023
  • Recently, the socio-economic interest in Large Language Models (LLMs) has been growing due to the emergence of ChatGPT. As a type of generative AI, LLMs have reached the level of script creation. In this regard, it is important to address the issue of discrimination (sexism, racism, religious discrimination, ageism, etc.) in the performing arts in general or in specific performing arts works or organizations in a large language model that will be widely used by the general public and professionals. However, there has not yet been a full-scale investigation and discussion on the issue of discrimination in the performing arts in large-scale language models. Therefore, the purpose of this study is to textually analyze the perceptions of discrimination issues in the performing arts from LMMs and to derive implications for the performing arts field and the development of LMMs. First, BBQ (Bias Benchmark for QA) questions and measures for nine discrimination issues were used to measure the sensitivity to discrimination of the giant language models, and the answers derived from the representative giant language models were verified by performing arts experts to see if there were any parts of the giant language models' misperceptions, and then the giant language models' perceptions of the ethics of discriminatory views in the performing arts field were analyzed through the content analysis method. As a result of the analysis, implications for the performing arts field and points to be noted in the development of large-scale linguistic models were derived and discussed.

A Study on the Development of AI Presentation Automation Service Base on AWS (AWS 기반 AI 프레젠테이션 자동화 서비스 개발에 관한 연구)

  • Taein Kang;Juyeon Kim;Gayeon Park
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2023.11a
    • /
    • pp.943-944
    • /
    • 2023
  • 본 프로젝트는 AWS 에서 제공되는 AI/ML 플랫폼과 LLM 모델을 기반으로 문서 및 텍스트와 ppt 를 자동 변환하는 서비스로, 실생활 및 업무에서 활용이 가능하며 별도 사용자 조작 없이 사용가능한 발표자료 및 대본 제작 도구를 통해 일의 효율성을 향상시키고자 한다.

A Kth Shortest Path Algorithm with the Link-Based Label Setting Approach and Its Application for An Alternative Routes Selection (링크표지확정 다수경로탐색 알고리즘과 대안경로선정을 위한 활용)

  • Lee, Mee-Young;Baik, Nam-Cheol;Kang, Weon-Eui;Shin, Seong-Il
    • Journal of Korean Society of Transportation
    • /
    • v.22 no.4 s.75
    • /
    • pp.85-96
    • /
    • 2004
  • Given a Path represented by a sequence of link numbers in a transportation network, the reasonable path is defined as a path that any link is appeared multiple times in it. Application of the link labelmethod(LLM) to the shortest path algorithms(SPA) enables to model the reasonable path choice behavior in urban networks. This study aims at expanding the LLM to a Kth shortest path algorithms(KPSA), which adopts the node label setting method. The small-scaled network test demonstrated that the proposed algorithm works correctly and the revised Sioux fall network test showed that the path choice behaviors are reasonably reflected. In the large-scaled network based on the South Korea peninsula, drivers' route diversion perceptions are included as cost terms in total cost. The algorithm may be applied as an alternative route information tools for the deployment of ATIS.

Performance Analysis and Comparison on Mobility Management Schemes for NGN(Next Generation Network) (NGN에서의 이동성 관리 방안 성능 분석 및 비교)

  • Yu, Myoung-Ju;Lee, Jong-Min;Choi, Seong-Gon
    • The Journal of the Korea Contents Association
    • /
    • v.7 no.4
    • /
    • pp.56-65
    • /
    • 2007
  • For supporting seamless service in NGN, various schemes for mobility are studying in ITU-T, 3GPP, IETF, and so on. Currently, mobility support schemes such as MIPv6, IP2, Q.MMF, IMS was introduced. However the performance analysis and comparison of all existing schemes is not examined as yet. That is, the existing mobility methods need to be compared according to same performance parameter. So, in this paper, we investigate and compare the performances of MIPv6, IP2, Q.MMF on latency during location registration. The latency in location registration is consider as cost and time. Then, the performance analysis of Q.MMF Case1 has shown the best performance as it is measured the minimum cost and time. Q.MMF Case1 is the location registration process in case that MT moves within a local network. As LLM in a access network manages the location of MT in this case, Q.MMF Case1 has lower location cost and time than other cases.