• Title/Summary/Keyword: 인덱스

Search Result 1,089, Processing Time 0.027 seconds

Odysseus/Parallel-OOSQL: A Parallel Search Engine using the Odysseus DBMS Tightly-Coupled with IR Capability (오디세우스/Parallel-OOSQL: 오디세우스 정보검색용 밀결합 DBMS를 사용한 병렬 정보 검색 엔진)

  • Ryu, Jae-Joon;Whang, Kyu-Young;Lee, Jae-Gil;Kwon, Hyuk-Yoon;Kim, Yi-Reun;Heo, Jun-Suk;Lee, Ki-Hoon
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.14 no.4
    • /
    • pp.412-429
    • /
    • 2008
  • As the amount of electronic documents increases rapidly with the growth of the Internet, a parallel search engine capable of handling a large number of documents are becoming ever important. To implement a parallel search engine, we need to partition the inverted index and search through the partitioned index in parallel. There are two methods of partitioning the inverted index: 1) document-identifier based partitioning and 2) keyword-identifier based partitioning. However, each method alone has the following drawbacks. The former is convenient in inserting documents and has high throughput, but has poor performance for top h query processing. The latter has good performance for top-k query processing, but is inconvenient in inserting documents and has low throughput. In this paper, we propose a hybrid partitioning method to compensate for the drawback of each method. We design and implement a parallel search engine that supports the hybrid partitioning method using the Odysseus DBMS tightly coupled with information retrieval capability. We first introduce the architecture of the parallel search engine-Odysseus/parallel-OOSQL. We then show the effectiveness of the proposed system through systematic experiments. The experimental results show that the query processing time of the document-identifier based partitioning method is approximately inversely proportional to the number of blocks in the partition of the inverted index. The results also show that the keyword-identifier based partitioning method has good performance in top-k query processing. The proposed parallel search engine can be optimized for performance by customizing the methods of partitioning the inverted index according to the application environment. The Odysseus/parallel OOSQL parallel search engine is capable of indexing, storing, and querying 100 million web documents per node or tens of billions of web documents for the entire system.

A Study on the Exploratory Spatial Data Analysis of the Distribution of Longevity Population and the Scale Effect of the Modifiable Areal Unit Problem(MAUP) (장수 인구의 분포 패턴에 관한 탐색적 공간 데이터 분석과 수정 가능한 공간단위 문제(MAUP)의 Scale Effect에 관한 연구)

  • Choi, Don-Jeong;Suh, Yong-Cheol
    • Journal of the Korean Association of Geographic Information Studies
    • /
    • v.16 no.3
    • /
    • pp.40-53
    • /
    • 2013
  • Most of the existing domestic studies to identify the distribution of longevity population and influencing factors oriented confirmatory approach. Furthermore, most of the studies in this research topic simply have used their own definition of spatial unit of analysis or employed arbitrary spatial units of analysis according to data availability. These research approaches can not sufficiently reflect the spatial characteristic of longevity phenomenon and exposed to the Modifiable Aerial Unit Problem(MAUP). This research performed the Exploratory Spatial Data Analysis(ESDA) to identify the spatial autocorrelation of the distribution of longevity population and investigated whether the modifiable areal unit problem in the aspect of scale effect using spatial population data in Korea. We used Si_Gun_Gu and Eup_Myeon_Dong as two different spatial units of regional longevity indicators measured. Then, we applied Getis-Ord Gi* to investigate the existence of spatial hot spots and cold spots. The results from our analysis show that there exist statistically significant spatial autocorrelation and spatial hot spots and cold spots of regional longevity at both Si_Gun_Gu and Eup_Myeon_Dong levels. This result implies that the modifiable areal unit problem does exist in the studies of spatial patterns of longevity population distribution. The demand for longevity researches would be increased inevitably. In addition, there were apparent differences for the global spatial autocorrelation and local spatial cluster which calculated different spatial units such as Si_Gun_Gu and Eup_Myeon_Dong and this can be seen as scale effect of MAUP. The findings from our analysis show that any study in this topic can mislead results when the modifiable areal unit problem and spatial autocorrelation are not explicitly considered.

Cerebral-perfusion Reserve after Carotid-artery Stenting: Relationship with Power Spectrum of Electroencephalography (경동맥스텐트삽입술 후의 뇌관류예비능: 뇌파파워스펙트럼과의 연관성)

  • Jeong, Da-hye;Jung, Seokwon;Kwak, Byeonggeun;Kim, Young-Soo;Kim, Soo-kyoung;Kwon, Oh-Young
    • Korean Journal of Clinical Laboratory Science
    • /
    • v.48 no.2
    • /
    • pp.144-152
    • /
    • 2016
  • Carotid-artery stenosis may reduce cerebral perfusion, and affect cerebral neuronal activities. We examined the question of whether the recovery of cerebral-perfusion reserve after carotid-artery stenting (CAS) can affect the EEG power-spectrum. Nineteen candidates for CAS were initially recruited. Subtraction imaging of single photon emissary computerized tomography (SPECT) and an electroencephalogram (EEG) were taken twice, before and 1 month after CAS. At each time point, the EEGs were recorded before and after injection of acetazolamide (pre-ACZ EEG and post-ACZ EEG). Finally, 7 patients were enrolled after exclusion of incomplete studies. We obtained the spectral ratio (SR) of each hemisphere. SR was defined as the divided value of the power-spectrum sum of fast activities by that of slow activities. The power-spectrum values between hemispheres were compared using the inter-hemispheric index of spectral ratio (IHISR), and we examined the correlation between the power-spectrum and the cerebral-perfusion reserve. Cerebral-perfusion reserve improved after CAS on the stent side in 6 of 7 patients. In 3 patients with unilateral carotid-artery stenosis, CAS increased SR on the pre-ACZ EEGs, and IHISR on the post-ACZ EEGs. The increases of SR and IHISR were concordant with the increment of cerebral-perfusion reserve. In contrast, the results in the other patients with bilateral stenosis showed complex patterns. The SR of pre-ACZ EEGs and IHISR of post-ACZ EEGs may be useful electrophysiological markers for the blood-flow reserve after CAS in patients with unilateral carotid-artery stenosis, but not in those with bilateral stenosis.

Evaluation on the Accuracy of Targeting Error Correction Through the Application of Target Locating System in Robotic CyberKnife (로봇 사이버나이프에서 위치인식시스템을 이용한 Targeting Error값 보정의 정확성 평가)

  • Jeong, Young-Joon;Jung, Jae-Hong;Lim, Kwang-Chae;Cho, Eun-Ju
    • The Journal of Korean Society for Radiation Therapy
    • /
    • v.21 no.1
    • /
    • pp.1-7
    • /
    • 2009
  • Purpose: The purpose is to evaluate the accuracy of correcting the targeting error through the Target Location System (TLS) for the location change error of the reference point which arises from the movement or motion of patient during the treatment using the CyberKnife. Materials and Methods: In this test, Gafchromic MD-55 film was inserted into the head and neck phantom to analyze the accuracy of the targeting, and then the 6 MV X-ray of CyberKnife (CyberKnife Robotic Radiosurgery System G4, Accuray, US) was irradiated. End to End (E2E) program was used to analyze the accuracy of targeting, which is provided by Accuray Corporation. To compute the error of the targeting, the test was carried out with the films that were irradiated 12 times by maintaining the distance within the rage of $0{\pm}0.2\;mm$ toward x, y, z from the reference point and maintaining the angle within the rage of $0{\pm}0.2^{\circ}$ toward roll, pitch, yaw, and then with the films which were irradiated 6 times by applying intentional movement. And the correlation in the average value of the reference film and the test film were analyzed through independent samples t-test. In addition, the consistency of dose distribution through gamma-index method (dose difference: 3%) was quantified, compared, and analyzed by varying the distance to agreement (DTA) to 1 mm, 1.5 mm, 2 mm, respectively. Results: E2E test result indicated that the average error of the reference film was 0.405 mm and the standard deviation was 0.069 mm. The average error of the test film was 0.413 mm with the standard deviation of 0.121 mm. The result of independent sampling t-test for both averages showed that the significant probability was P=0.836 (confidence level: 95%). Besides, by comparing the consistency of dose distribution of DTA through 1 mm, 1.5 mm, 2 mm, it was found that the average dose distribution of axial film was 95.04%, 97.56%, 98.13%, respectively in 3,314 locations of the reference film, consistent with the average dose distribution of sagittal film that was 95.47%, 97.68%, 98.47%, respectively. By comparing with the test film, it was found that the average dose distribution of axial film was 96.38%, 97.57%, 98.04%, respectively, at 3,323 locations, consistent with the average dose distribution of sagittal film which was 95.50%, 97.87%, 98.36%, respectively. Conclusion: Robotic CyberKnife traces and complements in real time the error in the location change of the reference point caused by the motion or movement of patient during the treatment and provides the accuracy with the consistency of over 95% dose distribution and the targeting error below 1 mm.

  • PDF

Estimation of the Characteristics for the Dose Distribution in the Polymer Gel by Means of Monte Carlo Simulation (몬테카를로 시뮬레이션을 이용한 양성자 조사에 따른 Polymer Gel 내부의 선량 분포 특성 평가)

  • Park, Min-Seok;Kim, Gi-Sub;Jung, Hai-Jo;Park, Se-Young;Choi, In-Seok;Kim, Hyun-Ji;Yoon, Yong-Su;Kim, Jung-Min
    • Journal of radiological science and technology
    • /
    • v.36 no.2
    • /
    • pp.165-173
    • /
    • 2013
  • This study was the estimation of the dose distribution for proton, prompt gamma rays and proton induced neutron particles, in case of exposing the proton beam to polymer gel dosimeter and water phantom. The polymer gel dosimeter was compositeness material of Gelatin, Methacrylic acid, Hydroquinone, Tetrakis and Distilled water. The density of gel dosimeter was $1.04g/cm^3$ which was similar to water. The 72, 116 and 140 MeV proton beams were used in the simulation. Proton beam interacted with the nuclei of the phantom and the nuclei in excited states emitted prompt gamma rays and proton induced neutron particles during the process of de-excitation. The proton particles, prompt gamma rays, proton induced neutron particles were detected by polymer gel dosimeter and water phantom, respectively. The gap of the axis for gel was 2 mm. The Bragg-peak for proton particles in gel dosimeter was similar to water phantom. The dose distribution for proton and prompt gamma rays in gel dosimeter and water phantom was approximately identical in case of 72, 116 and 140 MeV for proton beam. However, in case of proton induced neutron particles for 72, 116 and 140 MeV proton beam, particles were not detected in gel dosimeter, while the Water phantom absorbed neutron particles. Considering the resulting data, gel dosimeter which was developed in the normoxic state attentively detected the dose distribution for proton beam exposure except proton induced neutron particles.

Examination of Dose Change at the Junction at the Time of Treatment Using Multi-Isocenter Volumetric Modulated Arc Therapy (용적조절호형방사선치료(VMAT)의 다중치료중심(Multi- Isocenter)을 이용한 치료 시, 접합부(Junction)의 선량 변화에 대한 고찰)

  • Jung, Dong Min;Park, Kwang Soon;Ahn, Hyuk Jin;Choi, Yoon Won;Park, Byul Nim;Kwon, Yong Jae;Moon, Sung Gong;Lee, Jong Oon;Jeong, Tae Sik;Park, Ryeong Hwang;Kim, Se young;Kim, Mi Jung;Baek, Jong Geol;Cho, Jeong Hee
    • The Journal of Korean Society for Radiation Therapy
    • /
    • v.33
    • /
    • pp.9-14
    • /
    • 2021
  • This study examined dose change depending on the reposition error of the junction at the time of treatment with multi-isocenter volumetric modulated arc therapy. This study selected a random treatment region in the Arccheck Phantom and established the treatment plan for multi-isocenter volumetric modulated arc therapy. Then, after setting the error of the junction at 0 ~ 4 mm in the X (left), Y (upper), and Z (inner and outer) directions, the area was irradiated using a linear accelerator; the point doses and gamma indexes obtained through the Phantom were subsequently analyzed. It was found that when errors of 2 and 4 mm took place in the X and Y directions, the gamma pass rates (point doses) were 99.3% (2.085) and 98% (2.079 Gy) in the former direction and 98.5% (2.088) and 95.5% (2.093 Gy) in the latter direction, respectively. In addition, when errors of 1, 2, and 4 mm occurred in the inner and outer parts of the Z direction, the gamma pass rates (point doses) were found to be 94.8% (2.131), 82.6% (2.164), and 72.8% (2.22 Gy) in the former part and 93.4% (2.069), 90.6% (2.047), and 79.7% (1.962 Gy) in the latter part, respectively. In the X and Y directions, errors up to 4 mm were tolerable; however, in the Z direction, error values exceeding 1 mm were beyond the tolerance level. This suggests that for high and low dose areas, errors in the direction same as the progress direction in the treatment region have a more sensitive dose distribution. If the guidelines for set-up errors are established at the institutional level through continuous research in the future, it will be possible to provide good quality treatment using junctions.

An Iterative, Interactive and Unified Seismic Velocity Analysis (반복적 대화식 통합 탄성파 속도분석)

  • Suh Sayng-Yong;Chung Bu-Heung;Jang Seong-Hyung
    • Geophysics and Geophysical Exploration
    • /
    • v.2 no.1
    • /
    • pp.26-32
    • /
    • 1999
  • Among the various seismic data processing sequences, the velocity analysis is the most time consuming and man-hour intensive processing steps. For the production seismic data processing, a good velocity analysis tool as well as the high performance computer is required. The tool must give fast and accurate velocity analysis. There are two different approches in the velocity analysis, batch and interactive. In the batch processing, a velocity plot is made at every analysis point. Generally, the plot consisted of a semblance contour, super gather, and a stack pannel. The interpreter chooses the velocity function by analyzing the velocity plot. The technique is highly dependent on the interpreters skill and requires human efforts. As the high speed graphic workstations are becoming more popular, various interactive velocity analysis programs are developed. Although, the programs enabled faster picking of the velocity nodes using mouse, the main improvement of these programs is simply the replacement of the paper plot by the graphic screen. The velocity spectrum is highly sensitive to the presence of the noise, especially the coherent noise often found in the shallow region of the marine seismic data. For the accurate velocity analysis, these noise must be removed before the spectrum is computed. Also, the velocity analysis must be carried out by carefully choosing the location of the analysis point and accuarate computation of the spectrum. The analyzed velocity function must be verified by the mute and stack, and the sequence must be repeated most time. Therefore an iterative, interactive, and unified velocity analysis tool is highly required. An interactive velocity analysis program, xva(X-Window based Velocity Analysis) was invented. The program handles all processes required in the velocity analysis such as composing the super gather, computing the velocity spectrum, NMO correction, mute, and stack. Most of the parameter changes give the final stack via a few mouse clicks thereby enabling the iterative and interactive processing. A simple trace indexing scheme is introduced and a program to nike the index of the Geobit seismic disk file was invented. The index is used to reference the original input, i.e., CDP sort, directly A transformation techinique of the mute function between the T-X domain and NMOC domain is introduced and adopted to the program. The result of the transform is simliar to the remove-NMO technique in suppressing the shallow noise such as direct wave and refracted wave. However, it has two improvements, i.e., no interpolation error and very high speed computing time. By the introduction of the technique, the mute times can be easily designed from the NMOC domain and applied to the super gather in the T-X domain, thereby producing more accurate velocity spectrum interactively. The xva program consists of 28 files, 12,029 lines, 34,990 words and 304,073 characters. The program references Geobit utility libraries and can be installed under Geobit preinstalled environment. The program runs on X-Window/Motif environment. The program menu is designed according to the Motif style guide. A brief usage of the program has been discussed. The program allows fast and accurate seismic velocity analysis, which is necessary computing the AVO (Amplitude Versus Offset) based DHI (Direct Hydrocarn Indicator), and making the high quality seismic sections.

  • PDF

Dosimetric Effect on Selectable Optimization Parameters of Volumatric Modulated Arc Therapy (선택적 최적화 변수(Selectable Optimization Parameters)에 따른 부피적조절회전방사선치료(VMAT)의 선량학적 영향)

  • Jung, Jae-Yong;Shin, Yong-Joo;Sohn, Seung-Chang;Kim, Yeon-Rae;Min, Jung-Wan;Suh, Tae-Suk
    • Progress in Medical Physics
    • /
    • v.23 no.1
    • /
    • pp.15-25
    • /
    • 2012
  • The aim of this study is to evaluate plan quality and dose accuracy for Volumetric Modulated Arc Therapy (VMAT) on the TG-119 and is to investigate the effects on variation of the selectable optimization parameters of VMAT. VMAT treatment planning was implemented on a Varian iX linear accelerator with ARIA record and verify system (Varian Mecical System Palo Alto, CA) and Oncentra MasterPlan treatment planning system (Nucletron BV, Veenendaal, Netherlands). Plan quality and dosimetric accuracy were evaluated by effect of varying a number of arc, gantry spacing and delivery time for the test geometries provided in TG-119. Plan quality for the target and OAR was evaluated by the mean value and the standard deviation of the Dose Volume Histograms (DVHs). The ionization chamber and $Delta^{4PT}$ bi-planar diode array were used for the dose evaluation. For treatment planning evaluation, all structure sets closed to the goals in the case of single arc, except for the C-shape (hard), and all structure sets achieved the goals in the case of dual arc, except for C-shape (hard). For the variation of a number of arc, the simple structure such as a prostate did not have the difference between single arc and dual arc, whereas the complex structure such as a head and neck showed a superior result in the case of dual arc. The dose distribution with gantry spacing of $4^{\circ}$ was shown better plan quality than the gantry spacing of $6^{\circ}$, but was similar results compared with gantry spacing of $2^{\circ}$. For the verification of dose accuracy with single arc and dual arc, the mean value of a relative error between measured and calculated value were within 3% and 4% for point dose and confidence limit values, respectively. For the verification on dose accuracy with the gantry intervals of $2^{\circ}$, $4^{\circ}$ and $6^{\circ}$, the mean values of relative error were within 3% and 5% for point dose and confidence limit values, respectively. In the verification of dose distribution with $Delta^{4PT}$ bi-planar diode array, gamma passing rate was $98.72{\pm}1.52%$ and $98.3{\pm}1.5%$ for single arc and dual arc, respectively. The confidence limit values were within 4%. The smaller the gantry spacing, the more accuracy results were shown. In this study, we performed the VMAT QA based on TG-119 procedure, and demonstrated that all structure sets were satisfied with acceptance criteria. And also, the results for the selective optimization variables informed the importance of selection for the suitable variables according to the clinical cases.

Development of Yóukè Mining System with Yóukè's Travel Demand and Insight Based on Web Search Traffic Information (웹검색 트래픽 정보를 활용한 유커 인바운드 여행 수요 예측 모형 및 유커마이닝 시스템 개발)

  • Choi, Youji;Park, Do-Hyung
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.3
    • /
    • pp.155-175
    • /
    • 2017
  • As social data become into the spotlight, mainstream web search engines provide data indicate how many people searched specific keyword: Web Search Traffic data. Web search traffic information is collection of each crowd that search for specific keyword. In a various area, web search traffic can be used as one of useful variables that represent the attention of common users on specific interests. A lot of studies uses web search traffic data to nowcast or forecast social phenomenon such as epidemic prediction, consumer pattern analysis, product life cycle, financial invest modeling and so on. Also web search traffic data have begun to be applied to predict tourist inbound. Proper demand prediction is needed because tourism is high value-added industry as increasing employment and foreign exchange. Among those tourists, especially Chinese tourists: Youke is continuously growing nowadays, Youke has been largest tourist inbound of Korea tourism for many years and tourism profits per one Youke as well. It is important that research into proper demand prediction approaches of Youke in both public and private sector. Accurate tourism demands prediction is important to efficient decision making in a limited resource. This study suggests improved model that reflects latest issue of society by presented the attention from group of individual. Trip abroad is generally high-involvement activity so that potential tourists likely deep into searching for information about their own trip. Web search traffic data presents tourists' attention in the process of preparation their journey instantaneous and dynamic way. So that this study attempted select key words that potential Chinese tourists likely searched out internet. Baidu-Chinese biggest web search engine that share over 80%- provides users with accessing to web search traffic data. Qualitative interview with potential tourists helps us to understand the information search behavior before a trip and identify the keywords for this study. Selected key words of web search traffic are categorized by how much directly related to "Korean Tourism" in a three levels. Classifying categories helps to find out which keyword can explain Youke inbound demands from close one to far one as distance of category. Web search traffic data of each key words gathered by web crawler developed to crawling web search data onto Baidu Index. Using automatically gathered variable data, linear model is designed by multiple regression analysis for suitable for operational application of decision and policy making because of easiness to explanation about variables' effective relationship. After regression linear models have composed, comparing with model composed traditional variables and model additional input web search traffic data variables to traditional model has conducted by significance and R squared. after comparing performance of models, final model is composed. Final regression model has improved explanation and advantage of real-time immediacy and convenience than traditional model. Furthermore, this study demonstrates system intuitively visualized to general use -Youke Mining solution has several functions of tourist decision making including embed final regression model. Youke Mining solution has algorithm based on data science and well-designed simple interface. In the end this research suggests three significant meanings on theoretical, practical and political aspects. Theoretically, Youke Mining system and the model in this research are the first step on the Youke inbound prediction using interactive and instant variable: web search traffic information represents tourists' attention while prepare their trip. Baidu web search traffic data has more than 80% of web search engine market. Practically, Baidu data could represent attention of the potential tourists who prepare their own tour as real-time. Finally, in political way, designed Chinese tourist demands prediction model based on web search traffic can be used to tourism decision making for efficient managing of resource and optimizing opportunity for successful policy.