• Title/Summary/Keyword: vision AI

Search Result 141, Processing Time 0.026 seconds

Beyond Platforms to Ecosystems: Research on the Metaverse Industry Ecosystem Utilizing Information Ecology Theory (플랫폼을 넘어 생태계로: Information Ecology Theory를 활용한 메타버스 산업 생태계연구 )

  • Seokyoung Shin;Jaiyeol Son
    • Information Systems Review
    • /
    • v.25 no.4
    • /
    • pp.131-159
    • /
    • 2023
  • Recently, amidst the backdrop of the COVID-19 pandemic shifting towards an endemic phase, there has been a rise in discussions and debates about the future of the metaverse. Simultaneously, major metaverse platforms like Roblox have been launching services integrated with generative AI, and Apple's mixed reality hardware, Vision Pro, has been announced, creating new expectations for the metaverse. In this situation where the outlook for the metaverse is divided, it is crucial to diagnose the metaverse from an ecosystem perspective, examine its key ecological features, driving forces for development, and future possibilities for advancement. This study utilized Wang's (2021) Information Ecology Theory (IET) framework, which is representative of ecosystem research in the field of Information Systems (IS), to derive the Metaverse Industrial Ecosystem (MIE). The analysis revealed that the MIE consists of four main domains: Tech Landscape, Category Ecosystem, Metaverse Platform, and Product/Service Ecosystem. It was found that the MIE exhibits characteristics such as digital connectivity, the integration of real and virtual worlds, value creation capabilities, and value sharing (Web 3.0). Furthermore, the interactions among the domains within the MIE and the four characteristics of the ecosystem were identified as driving forces for the development of the MIE at an ecosystem level. Additionally, the development of the MIE at an ecosystem level was categorized into three distinct stages: Narrow Ecosystem, Expanded Ecosystem, and Everywhere Ecosystem. It is anticipated that future advancements in related technologies and industries, such as robotics, AI, and 6G, will promote the transition from the current Expanded Ecosystem level of the MIE to an Everywhere Ecosystem level, where the connection between the real and virtual worlds is pervasive. This study provides several implications. Firstly, it offers a foundational theory and analytical framework for ecosystem research, addressing a gap in previous metaverse studies. It also presents various research topics within the metaverse domain. Additionally, it establishes an academic foundation that integrates concept definition research and impact studies, which are key areas in metaverse research. Lastly, referring to the developmental stages and conditions proposed in this study, businesses and governments can explore future metaverse markets and related technologies. They can also consider diverse metaverse business strategies. These implications are expected to guide the exploration of the emerging metaverse market and facilitate the evaluation of various metaverse business strategies.

Vision-based Low-cost Walking Spatial Recognition Algorithm for the Safety of Blind People (시각장애인 안전을 위한 영상 기반 저비용 보행 공간 인지 알고리즘)

  • Sunghyun Kang;Sehun Lee;Junho Ahn
    • Journal of Internet Computing and Services
    • /
    • v.24 no.6
    • /
    • pp.81-89
    • /
    • 2023
  • In modern society, blind people face difficulties in navigating common environments such as sidewalks, elevators, and crosswalks. Research has been conducted to alleviate these inconveniences for the visually impaired through the use of visual and audio aids. However, such research often encounters limitations when it comes to practical implementation due to the high cost of wearable devices, high-performance CCTV systems, and voice sensors. In this paper, we propose an artificial intelligence fusion algorithm that utilizes low-cost video sensors integrated into smartphones to help blind people safely navigate their surroundings during walking. The proposed algorithm combines motion capture and object detection algorithms to detect moving people and various obstacles encountered during walking. We employed the MediaPipe library for motion capture to model and detect surrounding pedestrians during motion. Additionally, we used object detection algorithms to model and detect various obstacles that can occur during walking on sidewalks. Through experimentation, we validated the performance of the artificial intelligence fusion algorithm, achieving accuracy of 0.92, precision of 0.91, recall of 0.99, and an F1 score of 0.95. This research can assist blind people in navigating through obstacles such as bollards, shared scooters, and vehicles encountered during walking, thereby enhancing their mobility and safety.

Literature Review of AI Hallucination Research Since the Advent of ChatGPT: Focusing on Papers from arXiv (챗GPT 등장 이후 인공지능 환각 연구의 문헌 검토: 아카이브(arXiv)의 논문을 중심으로)

  • Park, Dae-Min;Lee, Han-Jong
    • Informatization Policy
    • /
    • v.31 no.2
    • /
    • pp.3-38
    • /
    • 2024
  • Hallucination is a significant barrier to the utilization of large-scale language models or multimodal models. In this study, we collected 654 computer science papers with "hallucination" in the abstract from arXiv from December 2022 to January 2024 following the advent of Chat GPT and conducted frequency analysis, knowledge network analysis, and literature review to explore the latest trends in hallucination research. The results showed that research in the fields of "Computation and Language," "Artificial Intelligence," "Computer Vision and Pattern Recognition," and "Machine Learning" were active. We then analyzed the research trends in the four major fields by focusing on the main authors and dividing them into data, hallucination detection, and hallucination mitigation. The main research trends included hallucination mitigation through supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF), inference enhancement via "chain of thought" (CoT), and growing interest in hallucination mitigation within the domain of multimodal AI. This study provides insights into the latest developments in hallucination research through a technology-oriented literature review. This study is expected to help subsequent research in both engineering and humanities and social sciences fields by understanding the latest trends in hallucination research.

Computer Vision-based Continuous Large-scale Site Monitoring System through Edge Computing and Small-Object Detection

  • Kim, Yeonjoo;Kim, Siyeon;Hwang, Sungjoo;Hong, Seok Hwan
    • International conference on construction engineering and project management
    • /
    • 2022.06a
    • /
    • pp.1243-1244
    • /
    • 2022
  • In recent years, the growing interest in off-site construction has led to factories scaling up their manufacturing and production processes in the construction sector. Consequently, continuous large-scale site monitoring in low-variability environments, such as prefabricated components production plants (precast concrete production), has gained increasing importance. Although many studies on computer vision-based site monitoring have been conducted, challenges for deploying this technology for large-scale field applications still remain. One of the issues is collecting and transmitting vast amounts of video data. Continuous site monitoring systems are based on real-time video data collection and analysis, which requires excessive computational resources and network traffic. In addition, it is difficult to integrate various object information with different sizes and scales into a single scene. Various sizes and types of objects (e.g., workers, heavy equipment, and materials) exist in a plant production environment, and these objects should be detected simultaneously for effective site monitoring. However, with the existing object detection algorithms, it is difficult to simultaneously detect objects with significant differences in size because collecting and training massive amounts of object image data with various scales is necessary. This study thus developed a large-scale site monitoring system using edge computing and a small-object detection system to solve these problems. Edge computing is a distributed information technology architecture wherein the image or video data is processed near the originating source, not on a centralized server or cloud. By inferring information from the AI computing module equipped with CCTVs and communicating only the processed information with the server, it is possible to reduce excessive network traffic. Small-object detection is an innovative method to detect different-sized objects by cropping the raw image and setting the appropriate number of rows and columns for image splitting based on the target object size. This enables the detection of small objects from cropped and magnified images. The detected small objects can then be expressed in the original image. In the inference process, this study used the YOLO-v5 algorithm, known for its fast processing speed and widely used for real-time object detection. This method could effectively detect large and even small objects that were difficult to detect with the existing object detection algorithms. When the large-scale site monitoring system was tested, it performed well in detecting small objects, such as workers in a large-scale view of construction sites, which were inaccurately detected by the existing algorithms. Our next goal is to incorporate various safety monitoring and risk analysis algorithms into this system, such as collision risk estimation, based on the time-to-collision concept, enabling the optimization of safety routes by accumulating workers' paths and inferring the risky areas based on workers' trajectory patterns. Through such developments, this continuous large-scale site monitoring system can guide a construction plant's safety management system more effectively.

  • PDF

Enhancing the performance of the facial keypoint detection model by improving the quality of low-resolution facial images (저화질 안면 이미지의 화질 개선를 통한 안면 특징점 검출 모델의 성능 향상)

  • KyoungOok Lee;Yejin Lee;Jonghyuk Park
    • Journal of Intelligence and Information Systems
    • /
    • v.29 no.2
    • /
    • pp.171-187
    • /
    • 2023
  • When a person's face is recognized through a recording device such as a low-pixel surveillance camera, it is difficult to capture the face due to low image quality. In situations where it is difficult to recognize a person's face, problems such as not being able to identify a criminal suspect or a missing person may occur. Existing studies on face recognition used refined datasets, so the performance could not be measured in various environments. Therefore, to solve the problem of poor face recognition performance in low-quality images, this paper proposes a method to generate high-quality images by performing image quality improvement on low-quality facial images considering various environments, and then improve the performance of facial feature point detection. To confirm the practical applicability of the proposed architecture, an experiment was conducted by selecting a data set in which people appear relatively small in the entire image. In addition, by choosing a facial image dataset considering the mask-wearing situation, the possibility of expanding to real problems was explored. As a result of measuring the performance of the feature point detection model by improving the image quality of the face image, it was confirmed that the face detection after improvement was enhanced by an average of 3.47 times in the case of images without a mask and 9.92 times in the case of wearing a mask. It was confirmed that the RMSE for facial feature points decreased by an average of 8.49 times when wearing a mask and by an average of 2.02 times when not wearing a mask. Therefore, it was possible to verify the applicability of the proposed method by increasing the recognition rate for facial images captured in low quality through image quality improvement.

The Low Carbon & Green Growth Policy and Green Life-Style, The Practical Implication and Vision on Family (저탄소녹색성장정책과 녹색생활양식, 가족에 대한 실천적 함의와 전망)

  • Choi, Youn-Shil;Sung, Mi-Ai
    • Journal of the Korean Home Economics Association
    • /
    • v.49 no.1
    • /
    • pp.79-91
    • /
    • 2011
  • The purposes of this study were firstly to explore the practical implications that of 'low carbon and green growth' policy, which is at the top of the Government's agenda provides to family, and secondly to propose some visions for a future based on those implications. The results of this study were as follows: Firstly, in terms of a global perspective, there is now a worldwide trend towards the adoption of 'low carbon and green growth' policies. Secondly, the Government-driven 'green growth policy' demands a total transformation, that is, revolution, not only in terms of our industries, but also in terms of our mentality and ordinary life. Thirdly, the driving force for this life revolution lies in having green life style, and the family is the primary agent for making the green life style a practical reality.

A Study on the 3-D Information Abstraction of object using Triangulation System (물체의 3-D 형상 복원을 위한 삼각측량 시스템)

  • Kim, Kuk-Se;Lee, Jeong-Ki;Cho, Ai-Ri;Ba, Il-Ho;Lee, Joon
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2003.05a
    • /
    • pp.409-412
    • /
    • 2003
  • The 3-D shape use to effect of movie, animation, industrial design, medical treatment service, education, engineering etc... But it is not easy to make 3-D shape from the information of 2-D image. There are two methods in restoring 3-D video image through 2-D image; First the method of using a laser; Second, the method of acquiring 3-D image through stereo vision. Instead of doing two methods with many difficulties, I study the method of simple 3-D image in this research paper. We present here a simple and efficient method, called direct calibration, which does not require any equations at all. The direct calibration procedure builds a lookup table(LUT) linking image and 3-D coordinates by a real 3-D triangulation system. The LUT is built by measuring the image coordinates of a grid of known 3-D points, and recording both image and world coordinates for each point; the depth values of all other visible points are obtained by interpolation.

  • PDF

Analysis of Mentors' Roles using IPA in the Workplace Mentoring : From the Perspective of Mentors and Mentees (IPA를 이용한 직장멘토링에서 멘토의 역할 분석 : 멘토와 멘티의 관점에서)

  • Kim, Jae Kyeong;Choi, Bhang Gil;Choi, Il Young;Son, Yu Kyung
    • Journal of Information Technology Services
    • /
    • v.20 no.4
    • /
    • pp.69-80
    • /
    • 2021
  • Many studies have discussed the effectiveness of mentoring from a mentor or mentee perspective. However, it is is necessary to deeply understand the formal mentoring relationship from the perspective of both the mentor and the mentee because the mentoring relationship is the interaction between the mentor and the mentee. Therefore, in this study, the mentors' role through IPA was compared and analyzed from the perspective of mentors and mentees. A survey was conducted on 376 employees of the financial bank, and the managers in charge of the company's official workplace mentoring and employees who participated in the mentoring program were interviewed. As a result of the analysis, mentors are more satisfied with the rewarding experience, while mentees are satisfied with commitment, and organizational ascendency and impact. In addition, mentees judge that "Coach", "Provides support", "Provides vision & widens horizons", "Broaden experience", "Cooperation", "Motivates", "Networking ability", "Provide cross-functional information", "Role model", "Share credit", "Teacher", and "Transfer skills, leadership, & technology" are important as mentor's roles are important. Therefore, in order to foster mentors for effective workplace mentoring, it is necessary to educate the mentor in advance about the mentors' role that the mentee considers to be important.

Light-weight Gender Classification and Age Estimation based on Ensemble Multi-tasking Deep Learning (앙상블 멀티태스킹 딥러닝 기반 경량 성별 분류 및 나이별 추정)

  • Huy Tran, Quoc Bao;Park, JongHyeon;Chung, SunTae
    • Journal of Korea Multimedia Society
    • /
    • v.25 no.1
    • /
    • pp.39-51
    • /
    • 2022
  • Image-based gender classification and age estimation of human are classic problems in computer vision. Most of researches in this field focus just only one task of either gender classification or age estimation and most of the reported methods for each task focus on accuracy performance and are not computationally light. Thus, running both tasks together simultaneously on low cost mobile or embedded systems with limited cpu processing speed and memory capacity are practically prohibited. In this paper, we propose a novel light-weight gender classification and age estimation method based on ensemble multitasking deep learning with light-weight processing neural network architecture, which processes both gender classification and age estimation simultaneously and in real-time even for embedded systems. Through experiments over various well-known datasets, it is shown that the proposed method performs comparably to the state-of-the-art gender classification and/or age estimation methods with respect to accuracy and runs fast enough (average 14fps) on a Jestson Nano embedded board.

Trends and Implications of Digital Transformation in Vehicle Experience and Audio User Interface (차내 경험의 디지털 트랜스포메이션과 오디오 기반 인터페이스의 동향 및 시사점)

  • Kim, Kihyun;Kwon, Seong-Geun
    • Journal of Korea Multimedia Society
    • /
    • v.25 no.2
    • /
    • pp.166-175
    • /
    • 2022
  • Digital transformation is driving so many changes in daily life and industry. The automobile industry is in a similar situation. In some cases, element techniques in areas called metabuses are also being adopted, such as 3D animated digital cockpit, around view, and voice AI, etc. Through the growth of the mobile market, the norm of human-computer interaction (HCI) has been evolving from keyboard-mouse interaction to touch screen. The core area was the graphical user interface (GUI), and recently, the audio user interface (AUI) has partially replaced the GUI. Since it is easy to access and intuitive to the user, it is quickly becoming a common area of the in-vehicle experience (IVE), especially. The benefits of a AUI are freeing the driver's eyes and hands, using fewer screens, lower interaction costs, more emotional and personal, effective for people with low vision. Nevertheless, when and where to apply a GUI or AUI are actually different approaches because some information is easier to process as we see it. In other cases, there is potential that AUI is more suitable. This is a study on a proposal to actively apply a AUI in the near future based on the context of various scenes occurring to improve IVE.