• Title/Summary/Keyword: parallel performance

Search Result 2,860, Processing Time 0.035 seconds

Characteristics of Electrode Potential and AC Impendance of Perchlorate Ion-Selective Electrodes Based on Quaternary Phosphonium Salts in PVC Membranes (제4급 인산염을 이용한 과염소산 이온선택성 PVC막 전극의 전극전위와 AC 임피던스 특성)

  • 안형환
    • Membrane Journal
    • /
    • v.9 no.4
    • /
    • pp.230-239
    • /
    • 1999
  • Perchlorate ion-selective electrodes in PVC membranes that respond linearly to concentration 106 M were developed by incorporating the quaternary phosphonium salts as a canier. The effects of the chemical structure, the contents of canier, the kind of plasticizer and the membrane thickness on electrode characteristics such as the electrode slope, the linear respone range and the detection limit were studied. With this results, the detectable pH range, selectivity coefficients and AC impedance characteristics were compared and investigated. The perchlorate ion substituents of the quaternary phosphonium salts like tetraoctylphosphonium perchlorate (TOPP) , tetraphenylphosphonium perchlorate(TPPP), and tetrabutylphosphonium perchlorate(TBPP) as a canier were used. The electrode characteristics were better in the ascending order of TBPP < TPPP < TOPP, with the increase of carbon chain length of the alkyl group. Dioctylsebacate(OOS) was best as a plasticizer, the canier contents were better with 11.76 wt% and the optimum membrane thickness was 0.19 mm. Under the above condition, the electrode slope was 56.58 mV/$^P{ClO}_4$,the linear response range was $10^{-1}$\times$10^{-6}$ M, the detection limit was 9.66 x $10^{-7}$ M. The performance of electrode was better than Orion electrode. The electrode potential was stable within the pH range from 3 to 11. The order of the selectivity coefficients for the perchlorate ion was sol < F < Br < 1. With the result of impedance spectrum, it was found that the equivalent circuit for the electrode could be expressed by a series combination of solution resistance, parallel circuit consisting of the double layer capacitance and bulk resistance and Warburg impedance. And solution resistance was almost not appeared and Warburg impedance was highly appeared by diffusion. Then Warburg coefficient was 1.32$\times$$10^74 $\Omega$ $\cdot$ ${cm}^2/s^{1/2}$.

  • PDF

Study of system using load cell for real time weight sensing of artificial incubator (인공부화기의 실시간 중량감지를 위한 로드셀을 이용한 시스템 연구)

  • jeong, Jin-hyoung;Kim, Ae-kyung;Lee, Sang-Sik
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.11 no.2
    • /
    • pp.144-149
    • /
    • 2018
  • The eggs are incubated for 18 days through the generator and incubated in the developing incubator. During the developmental period, the weight loss of the fetus is correlated with the ventricular formation, and the proper ventricular formation is also associated with the healthy embryonic hatching and the egg hatching rate. However, in the incubator period of the domestic hatchery, it is a reality to acquire the resultant side by the Iranian standard weight measurement with the experience of the hatchery and the person concerned and the development period without the apparatus for measuring the present weight. As a result, prevalence of early mortality, hunger and illness during hatching are frequent. Monitoring the reduction of weaning weight is crucial to obtaining chick quality and hatching performance with weight changes within the development machine. Water loss is different depending on the size of eggs, egg shell, and elder group. We can expect to increase the hatching rate by measuring the weight change in real time and optimizing the ventilation change accordingly. There is a need to develop a real-time measurement system that can control 10 to 13% reduction of the total weight during hatching. The system through this study is a way to check the one - time directly when moving the existing egg, and it is impossible to control the measurement of the fetal water evaporation within the development period. Unlike systems that do not affect the hatching rate, four load cells are connected in parallel on the Arduino sketch board and the AT-command command is used to connect the mobile phone and computer in real time. The communication speed of Bluetooth was set to 15200 to match the communication speed of Arduino and Hyper-terminal program. The real - time monitoring system was designed to visually check the change of the weight of the fetus in the artificial incubator. In this way, we aimed to improve the hatching rate and health condition of the hatching eggs.

A Study on Training Dataset Configuration for Deep Learning Based Image Matching of Multi-sensor VHR Satellite Images (다중센서 고해상도 위성영상의 딥러닝 기반 영상매칭을 위한 학습자료 구성에 관한 연구)

  • Kang, Wonbin;Jung, Minyoung;Kim, Yongil
    • Korean Journal of Remote Sensing
    • /
    • v.38 no.6_1
    • /
    • pp.1505-1514
    • /
    • 2022
  • Image matching is a crucial preprocessing step for effective utilization of multi-temporal and multi-sensor very high resolution (VHR) satellite images. Deep learning (DL) method which is attracting widespread interest has proven to be an efficient approach to measure the similarity between image pairs in quick and accurate manner by extracting complex and detailed features from satellite images. However, Image matching of VHR satellite images remains challenging due to limitations of DL models in which the results are depending on the quantity and quality of training dataset, as well as the difficulty of creating training dataset with VHR satellite images. Therefore, this study examines the feasibility of DL-based method in matching pair extraction which is the most time-consuming process during image registration. This paper also aims to analyze factors that affect the accuracy based on the configuration of training dataset, when developing training dataset from existing multi-sensor VHR image database with bias for DL-based image matching. For this purpose, the generated training dataset were composed of correct matching pairs and incorrect matching pairs by assigning true and false labels to image pairs extracted using a grid-based Scale Invariant Feature Transform (SIFT) algorithm for a total of 12 multi-temporal and multi-sensor VHR images. The Siamese convolutional neural network (SCNN), proposed for matching pair extraction on constructed training dataset, proceeds with model learning and measures similarities by passing two images in parallel to the two identical convolutional neural network structures. The results from this study confirm that data acquired from VHR satellite image database can be used as DL training dataset and indicate the potential to improve efficiency of the matching process by appropriate configuration of multi-sensor images. DL-based image matching techniques using multi-sensor VHR satellite images are expected to replace existing manual-based feature extraction methods based on its stable performance, thus further develop into an integrated DL-based image registration framework.

Development of heat exchanger for underground water heat. II - Design and manufacture for heat exchanger of underground water - (지하수 이용을 위한 열교환기 개발. II - 지하수이용 냉·난방기 설계제작 -)

  • Lee, W.Y.;Ahn, D.H.;Kim, S.C.;Park, W.P.;Kang, Y.G.;Kim, S.B.
    • Journal of Practical Agriculture & Fisheries Research
    • /
    • v.4 no.1
    • /
    • pp.128-137
    • /
    • 2002
  • This study was conducted to develop the heat exchanger by utilizing the heat energy of underground water(15℃), which might be used for cooling and heating system of the agricultural facilities. We developed the heat exchanger by using the parallel type plat fin tube made of Aluminum(Al 6063), which was named Aloo-Heat(No. 0247164, offered by Korean Intellectual property Office). The trial manufactures were made from Aloo-heat which was 600mm, 700mm length respectively, and It were welded to the end "U" type in order to direct flow of the underground water. The performance test was carried out under the condition of open space and room temperature with the change of flow rate of the underground water and air. The results are as follows. 1. The trial manufactures had convection heat value from 33 to 156 W/m2℃, and It was coincided with design assumption. 2. The amount of energy transfer was increased with the increment of the area of heat transfer, the air flow, the gap of temperature inlet & outlet the underground water and the air. 3. The heat value was 6,825W when the air flow was 6,000m3/h and the gap of temperature between inlet and outlet of the underground water was 6℃, and It dropped from 25.8℃ to 23.2℃(-2.6℃ difference). The convection heat value was 88.5W/m2℃. 4. The heat value was 2.625W when the air flow was 4,000m3/h and the gap of temperature between inlet and outlet the underground water was 2℃, and It dropped from 27℃ to 22.5℃(-4.5℃ difference). The convection heat value was 33.6W/m2℃. 5. Correlation values(R2) of the testing heat values of the trial manufacture type I, II, and III were 0.9141, 0.8935, and 0.9323 respectively, and correlation values(R2) of the amount of the air flow 6,000m3/h, 5,000m3/h, 4,000m3/h were 0.9513, 0.9414, and 0.9003 respectively.

The Presence of Related Personnel Effects on the IPO of Special Listed Firms on KOSDAQ Market: Based on the Signal Effect of Third-party Social Recognition (관계인사 영입이 코스닥 기술특례기업 IPO성과에 미치는 영향: 제3자 사회적 인정의 신호 효과를 바탕으로)

  • Kiyong, Kim;Young-Hee, Ko
    • Asia-Pacific Journal of Business Venturing and Entrepreneurship
    • /
    • v.17 no.6
    • /
    • pp.13-24
    • /
    • 2022
  • The purpose of this study is to examine whether the existence of related personnel in KOSDAQ technology special listed firms has a signal effect on the market and affects performance when listed. The KOSDAQ technology special listing system is a system introduced to enable future growth by securing financing through corporate public offering based on the technology and marketability of technology-based startups and venture companies. As a result of analyzing 135 special technology companies listed from 2005 to 21 (excluding SPAC mergers and foreign companies) whether or not related personnel affect corporate value and listing period when they are listed, it was analyzed that the presence of related personnel did not significantly affect corporate value or listing period. The same was found in the results of the verification by reducing the scope to related personnel such as public officials and related agencies. However, under certain conditions, significant results were derived from the presence of related personnel on the listing of companies listed in special technology cases. It was found that the presence of related personnel and VC investment had a significant effect on corporate value, and in the case of bio-industry, there was a slight significant effect on the duration of listing. This study is significant in that it systematically analyzed the signal effect of the existence of related personnel for the first time for all 135 companies. In addition, as a result of the analysis, the results suggest that internalized efforts to secure technology and marketability are more important, such as parallel to VC investment, rather than simply recruiting related personnel.

COVID-19 Rapid Antigen Test Results in Preschool and School (March 2 to May 1, 2022) (유치원·학교 구성원의 코로나19 신속항원검사 결과(2022년 3월 2일부터 5월 1일까지))

  • Gowoon Yun;Young-Joon Park;Eun Jung Jang;Sangeun Lee;Ryu Kyung Kim;Heegwon Jeong;Jin Gwack
    • Pediatric Infection and Vaccine
    • /
    • v.31 no.1
    • /
    • pp.113-121
    • /
    • 2024
  • Purpose: In response to the surge in coronavirus disease 2019 (COVID-19) omicron variant cases, we have implemented preemptive testing for preschool and school. The purpose is to quickly detect COVID-19 cases using a rapid antigen test (RAT) kit so that normal school activities can continue. Methods: The results entered in The Healthcare Self-Test App were merged with the information on the status of confirmed cases in the COVID-19 Information Management System by Korea Disease Control and Prevention Agency (KDCA) for preschool and school of students and staffs March 2 to May 1, 2022 to analyze the RAT positive rate and positive predictive value of RAT. Results: In preschool and school 19,458,575 people were tested, weekly RAT positive rate ranged from 1.10% to 5.90%, positive predictive value of RAT ranged from 86.42% to 93.18%. By status, RAT positive rate ranged from 1.13% to 6.16% for students, 0.99% to 3.93% for staffs, positive predictive value of RAT ranged from 87.19% to 94.03% for students, 77.55% to 83.10% for staffs. RAT positive rate by symptoms ranged from 76.32% to 88.02% for those with symptoms and 0.34% to 1.11% for those without symptoms. As a result of preschool and school RAT, 943,342 confirmed cases were preemptively detected, before infection spread in preschool and school. Conclusions: RAT was well utilized to detect confirmed cases at an early stage, reducing the risk of transmission to minimize the educational gap in preschool and school. To compensate for the limitations of RAT, further research should continue to reevaluate the performance of RAT as new strains of viruses continue to emerge. We will have to come up with various ways to utilize it, such as performing periodic and repeated RAT and parallel polymerase chain reaction.

Methods for Integration of Documents using Hierarchical Structure based on the Formal Concept Analysis (FCA 기반 계층적 구조를 이용한 문서 통합 기법)

  • Kim, Tae-Hwan;Jeon, Ho-Cheol;Choi, Joong-Min
    • Journal of Intelligence and Information Systems
    • /
    • v.17 no.3
    • /
    • pp.63-77
    • /
    • 2011
  • The World Wide Web is a very large distributed digital information space. From its origins in 1991, the web has grown to encompass diverse information resources as personal home pasges, online digital libraries and virtual museums. Some estimates suggest that the web currently includes over 500 billion pages in the deep web. The ability to search and retrieve information from the web efficiently and effectively is an enabling technology for realizing its full potential. With powerful workstations and parallel processing technology, efficiency is not a bottleneck. In fact, some existing search tools sift through gigabyte.syze precompiled web indexes in a fraction of a second. But retrieval effectiveness is a different matter. Current search tools retrieve too many documents, of which only a small fraction are relevant to the user query. Furthermore, the most relevant documents do not nessarily appear at the top of the query output order. Also, current search tools can not retrieve the documents related with retrieved document from gigantic amount of documents. The most important problem for lots of current searching systems is to increase the quality of search. It means to provide related documents or decrease the number of unrelated documents as low as possible in the results of search. For this problem, CiteSeer proposed the ACI (Autonomous Citation Indexing) of the articles on the World Wide Web. A "citation index" indexes the links between articles that researchers make when they cite other articles. Citation indexes are very useful for a number of purposes, including literature search and analysis of the academic literature. For details of this work, references contained in academic articles are used to give credit to previous work in the literature and provide a link between the "citing" and "cited" articles. A citation index indexes the citations that an article makes, linking the articleswith the cited works. Citation indexes were originally designed mainly for information retrieval. The citation links allow navigating the literature in unique ways. Papers can be located independent of language, and words in thetitle, keywords or document. A citation index allows navigation backward in time (the list of cited articles) and forwardin time (which subsequent articles cite the current article?) But CiteSeer can not indexes the links between articles that researchers doesn't make. Because it indexes the links between articles that only researchers make when they cite other articles. Also, CiteSeer is not easy to scalability. Because CiteSeer can not indexes the links between articles that researchers doesn't make. All these problems make us orient for designing more effective search system. This paper shows a method that extracts subject and predicate per each sentence in documents. A document will be changed into the tabular form that extracted predicate checked value of possible subject and object. We make a hierarchical graph of a document using the table and then integrate graphs of documents. The graph of entire documents calculates the area of document as compared with integrated documents. We mark relation among the documents as compared with the area of documents. Also it proposes a method for structural integration of documents that retrieves documents from the graph. It makes that the user can find information easier. We compared the performance of the proposed approaches with lucene search engine using the formulas for ranking. As a result, the F.measure is about 60% and it is better as about 15%.

Transfer Learning using Multiple ConvNet Layers Activation Features with Principal Component Analysis for Image Classification (전이학습 기반 다중 컨볼류션 신경망 레이어의 활성화 특징과 주성분 분석을 이용한 이미지 분류 방법)

  • Byambajav, Batkhuu;Alikhanov, Jumabek;Fang, Yang;Ko, Seunghyun;Jo, Geun Sik
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.1
    • /
    • pp.205-225
    • /
    • 2018
  • Convolutional Neural Network (ConvNet) is one class of the powerful Deep Neural Network that can analyze and learn hierarchies of visual features. Originally, first neural network (Neocognitron) was introduced in the 80s. At that time, the neural network was not broadly used in both industry and academic field by cause of large-scale dataset shortage and low computational power. However, after a few decades later in 2012, Krizhevsky made a breakthrough on ILSVRC-12 visual recognition competition using Convolutional Neural Network. That breakthrough revived people interest in the neural network. The success of Convolutional Neural Network is achieved with two main factors. First of them is the emergence of advanced hardware (GPUs) for sufficient parallel computation. Second is the availability of large-scale datasets such as ImageNet (ILSVRC) dataset for training. Unfortunately, many new domains are bottlenecked by these factors. For most domains, it is difficult and requires lots of effort to gather large-scale dataset to train a ConvNet. Moreover, even if we have a large-scale dataset, training ConvNet from scratch is required expensive resource and time-consuming. These two obstacles can be solved by using transfer learning. Transfer learning is a method for transferring the knowledge from a source domain to new domain. There are two major Transfer learning cases. First one is ConvNet as fixed feature extractor, and the second one is Fine-tune the ConvNet on a new dataset. In the first case, using pre-trained ConvNet (such as on ImageNet) to compute feed-forward activations of the image into the ConvNet and extract activation features from specific layers. In the second case, replacing and retraining the ConvNet classifier on the new dataset, then fine-tune the weights of the pre-trained network with the backpropagation. In this paper, we focus on using multiple ConvNet layers as a fixed feature extractor only. However, applying features with high dimensional complexity that is directly extracted from multiple ConvNet layers is still a challenging problem. We observe that features extracted from multiple ConvNet layers address the different characteristics of the image which means better representation could be obtained by finding the optimal combination of multiple ConvNet layers. Based on that observation, we propose to employ multiple ConvNet layer representations for transfer learning instead of a single ConvNet layer representation. Overall, our primary pipeline has three steps. Firstly, images from target task are given as input to ConvNet, then that image will be feed-forwarded into pre-trained AlexNet, and the activation features from three fully connected convolutional layers are extracted. Secondly, activation features of three ConvNet layers are concatenated to obtain multiple ConvNet layers representation because it will gain more information about an image. When three fully connected layer features concatenated, the occurring image representation would have 9192 (4096+4096+1000) dimension features. However, features extracted from multiple ConvNet layers are redundant and noisy since they are extracted from the same ConvNet. Thus, a third step, we will use Principal Component Analysis (PCA) to select salient features before the training phase. When salient features are obtained, the classifier can classify image more accurately, and the performance of transfer learning can be improved. To evaluate proposed method, experiments are conducted in three standard datasets (Caltech-256, VOC07, and SUN397) to compare multiple ConvNet layer representations against single ConvNet layer representation by using PCA for feature selection and dimension reduction. Our experiments demonstrated the importance of feature selection for multiple ConvNet layer representation. Moreover, our proposed approach achieved 75.6% accuracy compared to 73.9% accuracy achieved by FC7 layer on the Caltech-256 dataset, 73.1% accuracy compared to 69.2% accuracy achieved by FC8 layer on the VOC07 dataset, 52.2% accuracy compared to 48.7% accuracy achieved by FC7 layer on the SUN397 dataset. We also showed that our proposed approach achieved superior performance, 2.8%, 2.1% and 3.1% accuracy improvement on Caltech-256, VOC07, and SUN397 dataset respectively compare to existing work.

The Characteristics and Performances of Manufacturing SMEs that Utilize Public Information Support Infrastructure (공공 정보지원 인프라 활용한 제조 중소기업의 특징과 성과에 관한 연구)

  • Kim, Keun-Hwan;Kwon, Taehoon;Jun, Seung-pyo
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.4
    • /
    • pp.1-33
    • /
    • 2019
  • The small and medium sized enterprises (hereinafter SMEs) are already at a competitive disadvantaged when compared to large companies with more abundant resources. Manufacturing SMEs not only need a lot of information needed for new product development for sustainable growth and survival, but also seek networking to overcome the limitations of resources, but they are faced with limitations due to their size limitations. In a new era in which connectivity increases the complexity and uncertainty of the business environment, SMEs are increasingly urged to find information and solve networking problems. In order to solve these problems, the government funded research institutes plays an important role and duty to solve the information asymmetry problem of SMEs. The purpose of this study is to identify the differentiating characteristics of SMEs that utilize the public information support infrastructure provided by SMEs to enhance the innovation capacity of SMEs, and how they contribute to corporate performance. We argue that we need an infrastructure for providing information support to SMEs as part of this effort to strengthen of the role of government funded institutions; in this study, we specifically identify the target of such a policy and furthermore empirically demonstrate the effects of such policy-based efforts. Our goal is to help establish the strategies for building the information supporting infrastructure. To achieve this purpose, we first classified the characteristics of SMEs that have been found to utilize the information supporting infrastructure provided by government funded institutions. This allows us to verify whether selection bias appears in the analyzed group, which helps us clarify the interpretative limits of our study results. Next, we performed mediator and moderator effect analysis for multiple variables to analyze the process through which the use of information supporting infrastructure led to an improvement in external networking capabilities and resulted in enhancing product competitiveness. This analysis helps identify the key factors we should focus on when offering indirect support to SMEs through the information supporting infrastructure, which in turn helps us more efficiently manage research related to SME supporting policies implemented by government funded institutions. The results of this study showed the following. First, SMEs that used the information supporting infrastructure were found to have a significant difference in size in comparison to domestic R&D SMEs, but on the other hand, there was no significant difference in the cluster analysis that considered various variables. Based on these findings, we confirmed that SMEs that use the information supporting infrastructure are superior in size, and had a relatively higher distribution of companies that transact to a greater degree with large companies, when compared to the SMEs composing the general group of SMEs. Also, we found that companies that already receive support from the information infrastructure have a high concentration of companies that need collaboration with government funded institution. Secondly, among the SMEs that use the information supporting infrastructure, we found that increasing external networking capabilities contributed to enhancing product competitiveness, and while this was no the effect of direct assistance, we also found that indirect contributions were made by increasing the open marketing capabilities: in other words, this was the result of an indirect-only mediator effect. Also, the number of times the company received additional support in this process through mentoring related to information utilization was found to have a mediated moderator effect on improving external networking capabilities and in turn strengthening product competitiveness. The results of this study provide several insights that will help establish policies. KISTI's information support infrastructure may lead to the conclusion that marketing is already well underway, but it intentionally supports groups that enable to achieve good performance. As a result, the government should provide clear priorities whether to support the companies in the underdevelopment or to aid better performance. Through our research, we have identified how public information infrastructure contributes to product competitiveness. Here, we can draw some policy implications. First, the public information support infrastructure should have the capability to enhance the ability to interact with or to find the expert that provides required information. Second, if the utilization of public information support (online) infrastructure is effective, it is not necessary to continuously provide informational mentoring, which is a parallel offline support. Rather, offline support such as mentoring should be used as an appropriate device for abnormal symptom monitoring. Third, it is required that SMEs should improve their ability to utilize, because the effect of enhancing networking capacity through public information support infrastructure and enhancing product competitiveness through such infrastructure appears in most types of companies rather than in specific SMEs.

Information Privacy Concern in Context-Aware Personalized Services: Results of a Delphi Study

  • Lee, Yon-Nim;Kwon, Oh-Byung
    • Asia pacific journal of information systems
    • /
    • v.20 no.2
    • /
    • pp.63-86
    • /
    • 2010
  • Personalized services directly and indirectly acquire personal data, in part, to provide customers with higher-value services that are specifically context-relevant (such as place and time). Information technologies continue to mature and develop, providing greatly improved performance. Sensory networks and intelligent software can now obtain context data, and that is the cornerstone for providing personalized, context-specific services. Yet, the danger of overflowing personal information is increasing because the data retrieved by the sensors usually contains privacy information. Various technical characteristics of context-aware applications have more troubling implications for information privacy. In parallel with increasing use of context for service personalization, information privacy concerns have also increased such as an unrestricted availability of context information. Those privacy concerns are consistently regarded as a critical issue facing context-aware personalized service success. The entire field of information privacy is growing as an important area of research, with many new definitions and terminologies, because of a need for a better understanding of information privacy concepts. Especially, it requires that the factors of information privacy should be revised according to the characteristics of new technologies. However, previous information privacy factors of context-aware applications have at least two shortcomings. First, there has been little overview of the technology characteristics of context-aware computing. Existing studies have only focused on a small subset of the technical characteristics of context-aware computing. Therefore, there has not been a mutually exclusive set of factors that uniquely and completely describe information privacy on context-aware applications. Second, user survey has been widely used to identify factors of information privacy in most studies despite the limitation of users' knowledge and experiences about context-aware computing technology. To date, since context-aware services have not been widely deployed on a commercial scale yet, only very few people have prior experiences with context-aware personalized services. It is difficult to build users' knowledge about context-aware technology even by increasing their understanding in various ways: scenarios, pictures, flash animation, etc. Nevertheless, conducting a survey, assuming that the participants have sufficient experience or understanding about the technologies shown in the survey, may not be absolutely valid. Moreover, some surveys are based solely on simplifying and hence unrealistic assumptions (e.g., they only consider location information as a context data). A better understanding of information privacy concern in context-aware personalized services is highly needed. Hence, the purpose of this paper is to identify a generic set of factors for elemental information privacy concern in context-aware personalized services and to develop a rank-order list of information privacy concern factors. We consider overall technology characteristics to establish a mutually exclusive set of factors. A Delphi survey, a rigorous data collection method, was deployed to obtain a reliable opinion from the experts and to produce a rank-order list. It, therefore, lends itself well to obtaining a set of universal factors of information privacy concern and its priority. An international panel of researchers and practitioners who have the expertise in privacy and context-aware system fields were involved in our research. Delphi rounds formatting will faithfully follow the procedure for the Delphi study proposed by Okoli and Pawlowski. This will involve three general rounds: (1) brainstorming for important factors; (2) narrowing down the original list to the most important ones; and (3) ranking the list of important factors. For this round only, experts were treated as individuals, not panels. Adapted from Okoli and Pawlowski, we outlined the process of administrating the study. We performed three rounds. In the first and second rounds of the Delphi questionnaire, we gathered a set of exclusive factors for information privacy concern in context-aware personalized services. The respondents were asked to provide at least five main factors for the most appropriate understanding of the information privacy concern in the first round. To do so, some of the main factors found in the literature were presented to the participants. The second round of the questionnaire discussed the main factor provided in the first round, fleshed out with relevant sub-factors. Respondents were then requested to evaluate each sub factor's suitability against the corresponding main factors to determine the final sub-factors from the candidate factors. The sub-factors were found from the literature survey. Final factors selected by over 50% of experts. In the third round, a list of factors with corresponding questions was provided, and the respondents were requested to assess the importance of each main factor and its corresponding sub factors. Finally, we calculated the mean rank of each item to make a final result. While analyzing the data, we focused on group consensus rather than individual insistence. To do so, a concordance analysis, which measures the consistency of the experts' responses over successive rounds of the Delphi, was adopted during the survey process. As a result, experts reported that context data collection and high identifiable level of identical data are the most important factor in the main factors and sub factors, respectively. Additional important sub-factors included diverse types of context data collected, tracking and recording functionalities, and embedded and disappeared sensor devices. The average score of each factor is very useful for future context-aware personalized service development in the view of the information privacy. The final factors have the following differences comparing to those proposed in other studies. First, the concern factors differ from existing studies, which are based on privacy issues that may occur during the lifecycle of acquired user information. However, our study helped to clarify these sometimes vague issues by determining which privacy concern issues are viable based on specific technical characteristics in context-aware personalized services. Since a context-aware service differs in its technical characteristics compared to other services, we selected specific characteristics that had a higher potential to increase user's privacy concerns. Secondly, this study considered privacy issues in terms of service delivery and display that were almost overlooked in existing studies by introducing IPOS as the factor division. Lastly, in each factor, it correlated the level of importance with professionals' opinions as to what extent users have privacy concerns. The reason that it did not select the traditional method questionnaire at that time is that context-aware personalized service considered the absolute lack in understanding and experience of users with new technology. For understanding users' privacy concerns, professionals in the Delphi questionnaire process selected context data collection, tracking and recording, and sensory network as the most important factors among technological characteristics of context-aware personalized services. In the creation of a context-aware personalized services, this study demonstrates the importance and relevance of determining an optimal methodology, and which technologies and in what sequence are needed, to acquire what types of users' context information. Most studies focus on which services and systems should be provided and developed by utilizing context information on the supposition, along with the development of context-aware technology. However, the results in this study show that, in terms of users' privacy, it is necessary to pay greater attention to the activities that acquire context information. To inspect the results in the evaluation of sub factor, additional studies would be necessary for approaches on reducing users' privacy concerns toward technological characteristics such as highly identifiable level of identical data, diverse types of context data collected, tracking and recording functionality, embedded and disappearing sensor devices. The factor ranked the next highest level of importance after input is a context-aware service delivery that is related to output. The results show that delivery and display showing services to users in a context-aware personalized services toward the anywhere-anytime-any device concept have been regarded as even more important than in previous computing environment. Considering the concern factors to develop context aware personalized services will help to increase service success rate and hopefully user acceptance for those services. Our future work will be to adopt these factors for qualifying context aware service development projects such as u-city development projects in terms of service quality and hence user acceptance.