• Title/Summary/Keyword: research data

Search Result 69,841, Processing Time 0.085 seconds

Data Science and Deep Learning in Natural Sciences

  • Cha, Meeyoung
    • The Bulletin of The Korean Astronomical Society
    • /
    • v.44 no.2
    • /
    • pp.56.1-56.1
    • /
    • 2019
  • We are producing and consuming more data than ever before. Massive data allow us to better understand the world around us, yet they bring a new set of challenges due to their inherent noise and sheer enormity in size. Without smart algorithms and infrastructures, big data problems will remain intractable, and the same is true in natural science research. The mission of data science as a research field is to develop and apply computational methods in support of and in the replacement of costly practices in handling data. In this talk, I will introduce how data science and deep learning has been used for solving various problems in natural sciences. In particular, I will present a case study of analyzing high-resolution satellite images to infer socioeconomic scales of developing countries.

  • PDF

Feasibility Study of Case-Finding for Breast Cancer by Community Health Workers in Rural Bangladesh

  • Chowdhury, Touhidul Imran;Love, Richard Reed;Chowdhury, Mohammad Touhidul Imran;Artif, Abu Saeem;Ahsan, Hasib;Mamun, Anwarul;Khanam, Tahmina;Woods, James;Salim, Reza
    • Asian Pacific Journal of Cancer Prevention
    • /
    • v.16 no.17
    • /
    • pp.7853-7857
    • /
    • 2015
  • Background: Mortality from breast cancer is high in low- and middle-income countries, in part because most patients have advanced stage disease when first diagnosed. Case-finding may be one approach to changing this situation. Materials and Methods: We conducted a pilot study to explore the feasibility of population-based case finding for breast cancer by community health workers (CHWs) using different data collection methods and approaches to management of women found to have breast abnormalities. After training 8 CHWs in breast problem recognition, manual paper data collection and operation of a cell-phone software platform for reporting demographic, history and physical finding information, these CHWs visited 3150 women >age 18 and over they could find-- from 2356 households in 8 villages in rural Bangladesh. By 4 random assignments of villages, data were collected manually (Group 1), or with the cell-phone program alone (Group 2) or with management algorithms (Groups 3 and 4), and women adjudged to have a serious breast problem were shown a motivational video (Group 3), or navigated/accompanied to a breast problem center for evaluation (Group 4). Results: Only three visited women refused evaluation. The manual data acquisition group (1) had missing data in 80% of cases, and took an average of 5 minutes longer to acquire, versus no missing data in the cell phone-reporting groups (2,3 and 4). One woman was identified with stage III breast cancer, and was appropriately treated. Conclusions: Among very poor rural Bangladeshi women, there was very limited reluctance to undergo breast evaluation. The estimated rarity of clinical breast cancer is supported by these population-based findings. The feasibility and efficient use of mobile technology in this setting is supported. Successor studies may most appropriately be trials focusing on improving the suggested benefits of motivation and navigation, on increasing the numbers of cases found, and on stage of disease at diagnosis as the primary endpoint.

Business Intelligence and Marketing Insights in an Era of Big Data: The Q-sorting Approach

  • Kim, Ki Youn
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.8 no.2
    • /
    • pp.567-582
    • /
    • 2014
  • The purpose of this study is to qualitatively identify the typologies and characteristics of the big data marketing strategy in major companies that are taking advantage of the big data business in Korea. Big data means piles accumulated from converging platforms such as computing infrastructures, smart devices, social networking and new media, and big data is also an analytic technique itself. Numerous enterprises have grown conscious that big data can be a most significant resource or capability since the issue of big data recently surfaced abruptly in Korea. Companies will be obliged to design their own implementing plans for big data marketing and to customize their own analytic skills in the new era of big data, which will fundamentally transform how businesses operate and how they engage with customers, suppliers, partners and employees. This research employed a Q-study, which is a methodology, model, and theory used in 'subjectivity' research to interpret professional panels' perceptions or opinions through in-depth interviews. This method includes a series of q-sorting analysis processes, proposing 40 stimuli statements (q-sample) compressed out of about 60 (q-population) and explaining the big data marketing model derived from in-depth interviews with 20 marketing managers who belong to major companies(q-sorters). As a result, this study makes fundamental contributions to proposing new findings and insights for small and medium-size enterprises (SMEs) and policy makers that need guidelines or direction for future big data business.

Pseudo-standard and Its Implementation for the Maintenance Data of Ship and Offshore Structures (선박 및 해양 구조물에 있어서 유지보수용 데이터 교환을 위한 준표준 분석과 사례 구현)

  • Son, Gum-Jun;Lee, Jang-Hyun;Lee, Jeongyoul;Han, Eun-Jung
    • Korean Journal of Computational Design and Engineering
    • /
    • v.18 no.4
    • /
    • pp.267-274
    • /
    • 2013
  • This study focuses on the data schema and data content, which includes maintenance data, data structures and illustration data relevant with the maintenance process of ship and offshore structures. Product lifecycle management (PLM) is expected to encompass all the product data generated for the operation and maintenance information as well as the design and production. This paper introduces a data exchange schema in PLM of ship and offshore, serving as the basis for the role of standards required by the middle-of-life PLM. Also this paper identifies a typology of standards relevant to PLM that addresses the schema of evolving standards and identifies a XML schema supporting the exchange of data related with maintenance operations. Technical document based on standards in accordance with S1000D and Shipdex is explained. A case study illustrating the use of standard data exchange and technical document is presented.

A Comparison of Data Extraction Techniques and an Implementation of Data Extraction Technique using Index DB -S Bank Case- (원천 시스템 환경을 고려한 데이터 추출 방식의 비교 및 Index DB를 이용한 추출 방식의 구현 -ㅅ 은행 사례를 중심으로-)

  • 김기운
    • Korean Management Science Review
    • /
    • v.20 no.2
    • /
    • pp.1-16
    • /
    • 2003
  • Previous research on data extraction and integration for data warehousing has concentrated mainly on the relational DBMS or partly on the object-oriented DBMS. Mostly, it describes issues related with the change data (deltas) capture and the incremental update by using the triggering technique of active database systems. But, little attention has been paid to data extraction approaches from other types of source systems like hierarchical DBMS, etc. and from source systems without triggering capability. This paper argues, from the practical point of view, that we need to consider not only the types of information sources and capabilities of ETT tools but also other factors of source systems such as operational characteristics (i.e., whether they support DBMS log, user log or no log, timestamp), and DBMS characteristics (i.e., whether they have the triggering capability or not, etc), in order to find out appropriate data extraction techniques that could be applied to different source systems. Having applied several different data extraction techniques (e.g., DBMS log, user log, triggering, timestamp-based extraction, file comparison) to S bank's source systems (e.g., IMS, DB2, ORACLE, and SAM file), we discovered that data extraction techniques available in a commercial ETT tool do not completely support data extraction from the DBMS log of IMS system. For such IMS systems, a new date extraction technique is proposed which first creates Index database and then updates the data warehouse using the Index database. We illustrates this technique using an example application.

Development of Data Acquisition System to Obtain Vessel and Weather Information in Around Mokpo Harbor Bridge (목포대교 주변의 선박 항행 정보 및 기상 정보 획득 시스템 개발)

  • Yim, Jeong-Bin
    • Journal of Navigation and Port Research
    • /
    • v.35 no.1
    • /
    • pp.1-7
    • /
    • 2011
  • This paper describes the development of Data Acquisition System (DAS) to obtain the vessel and weather information needing to evaluate collision risks levels between Mokpo harbor bridge and passing vessels. DAS consists of Signal Receiving and Processing Unit to obtain the data sets of passing vessels and weather status, Networking Unit to transmit and distribute the acquisition data sets and Data Management Unit. Through the field tests on the deck of shuttle car ferry between Mokpo Port Passenger Terminal and An-Jua island, Sinan-Gun, we found that the DAS can provide useful data sets for adequate the collision risk evaluation. In addition, the noise-like data sets appeared in the weather data can be suppressed fully using 5-th order Butterworth digital filter.

A Study on Data Classification of Raman OIM Hyperspectral Bone Data

  • Jung, Sung-Hwan
    • Journal of Korea Multimedia Society
    • /
    • v.14 no.8
    • /
    • pp.1010-1019
    • /
    • 2011
  • This was a preliminary research for the goal of understanding between internal structure of Osteogenesis Imperfecta Murine (OIM) bone and its fragility. 54 hyperspectral bone data sets were captured by using JASCO 2000 Raman spectrometer at UMKC-CRISP (University of Missouri-Kansas City Center for Research on Interfacial Structure and Properties). Each data set consists of 1,091 data points from 9 OIM bones. The original captured hyperspectral data sets were noisy and base-lined ones. We removed the noise and corrected the base-lined data for the final efficient classification. High dimensional Raman hyperspectral data on OIM bones was reduced by Principal Components Analysis (PCA) and Linear Discriminant Analysis (LDA) and efficiently classified for the first time. We confirmed OIM bones could be classified such as strong, middle and weak one by using the coefficients of their PCA or LDA. Through experiment, we investigated the efficiency of classification on the reduced OIM bone data by the Bayesian classifier and K -Nearest Neighbor (K-NN) classifier. As the experimental result, the case of LDA reduction showed higher classification performance than that of PCA reduction in the two classifiers. K-NN classifier represented better classification rate, compared with Bayesian classifier. The classification performance of K-NN was about 92.6% in case of LDA.

Estimation of Material Requirement of Piping Materials in an Offshore Structure using Big Data Analysis (빅데이터 분석을 이용한 해양 구조물 배관 자재의 소요량 예측)

  • Oh, Min-Jae;Roh, Myung-Il;Park, Sung-Woo;Kim, Seong-Hoon
    • Journal of the Society of Naval Architects of Korea
    • /
    • v.55 no.3
    • /
    • pp.243-251
    • /
    • 2018
  • In the shipyard, a lot of data is generated, stored, and managed during design, construction, and operation phases to build ships and offshore structures. However, it is difficult to handle such big data efficiently using existing data-handling technologies. As the big data technology is developed, the ship and offshore industries start to focus on the existing big data to find valuable information from it. In this paper, the material requirement estimation method of offshore structure piping materials using big data analysis is proposed. A big data platform for the data analysis in the shipyard is introduced and it is applied to the analysis of material requirement estimation to solve the problems in piping design by a designer. The regression model is developed from the big data of piping materials and verified using the existing data. This analysis can help a piping designer to estimate the exact amount of material requirement and schedule the purchase time.

Construction of Street Trees Information Management Program Using GIS and Database (GIS와 데이터베이스를 이용한 가로수정보 관리프로그램 구축)

  • Kim, Hee-Nyeon;Jung, Sung-Gwan;Park, Kyung-Hun;You, Ju-Han
    • Current Research on Agriculture and Life Sciences
    • /
    • v.26
    • /
    • pp.45-54
    • /
    • 2008
  • The purpose of this research is to develope street trees management program for more an effective street trees management. The principal point of this program is to relate spatial data and attribute data that is the main concept in GIS(Geographic Information System). To do this function, MapObjects which is ESRI's mapping and GIS components was used to process spatial data and Access which had been developed by MS was used to manipulate attribute data in this program. Visual Basic also was used to design and develop user interfaces and procedures, relate two sort of data, and lastly complete Application. Relational data model was adopted to design tables and their relation, Antenucci's GIS development model was selected to design and complete this program. The configuration of this application is composed of management data and reference data. The management data includes the location of street tree, a growth condition, a surrounding environment, the characters of tree, an equipments, a management records and etc. The reference data include general information about tree, blight and insects.

  • PDF

Classification of Domestic Freight Data and Application for Network Models in the Era of 'Government 3.0' ('정부 3.0' 시대를 맞이한 국내 화물 자료의 집계 수준에 따른 분류체계 구축 및 네트워크 모형 적용방안)

  • YOO, Han Sol;KIM, Nam Seok
    • Journal of Korean Society of Transportation
    • /
    • v.33 no.4
    • /
    • pp.379-392
    • /
    • 2015
  • Freight flow data in Korea has been collected for a variety of purposes by various organizations. However, since the representation and format of the data varies, it has not been substantially used for freight analyses and furthermore for freight policies. In order to increase the applicability of those data sets, it is required to bring them in a table and compare for finding the differences. Then, it is shown that the raw data can be aggregated by a particular criterion such as mode, origin and destination, and type commodity. This study aims to examine the freight data issue in terms of three different points of view. First, we investigated various freight volume data sets which are released by several organizations. Second, we tried to develop formulations for freight volume data. Third, we discussed how to apply the formulations to network models in which particular OR (Operations Research) techniques are used. The results emphasized that some data might be useless for modeling once they are aggregated. As a result of examining the freight volume data, this study found that 14 organizations share their data sets at various aggregation levels. This study is not an ordinary research article, which normally includes data analysis, because it seems to be impossible to conduct extensive case studies. The reason is that the data dealt in this study are diverse. Nevertheless, this study might guide the research direction in the freight transport research society in terms of data issue. Especially, it can be concluded that this study is a timely research because the governmemt has emphasized the importance of sharing data to public throughout 'government 3.0' for research purpose.