• Title/Summary/Keyword: Using Distance

Search Result 12,889, Processing Time 0.05 seconds

High Resolution HC$_3$N Observations toward the Central Region of Sagittarius B2

  • H.S-Ching;Oh, M.ishi;M.Morimoto
    • Bulletin of the Korean Space Science Society
    • /
    • 1993.10a
    • /
    • pp.17-17
    • /
    • 1993
  • The wall shear stress in the vicinity of end-to end anastomoses under steady flow conditions was measured using a flush-mounted hot-film anemometer(FMHFA) probe. The experimental measurements were in good agreement with numerical results except in flow with low Reynolds numbers. The wall shear stress increased proximal to the anastomosis in flow from the Penrose tubing (simulating an artery) to the PTFE: graft. In flow from the PTFE graft to the Penrose tubing, low wall shear stress was observed distal to the anastomosis. Abnormal distributions of wall shear stress in the vicinity of the anastomosis, resulting from the compliance mismatch between the graft and the host artery, might be an important factor of ANFH formation and the graft failure. The present study suggests a correlation between regions of the low wall shear stress and the development of anastomotic neointimal fibrous hyperplasia(ANPH) in end-to-end anastomoses. 30523 T00401030523 ^x Air pressure decay(APD) rate and ultrafiltration rate(UFR) tests were performed on new and saline rinsed dialyzers as well as those roused in patients several times. C-DAK 4000 (Cordis Dow) and CF IS-11 (Baxter Travenol) reused dialyzers obtained from the dialysis clinic were used in the present study. The new dialyzers exhibited a relatively flat APD, whereas saline rinsed and reused dialyzers showed considerable amount of decay. C-DAH dialyzers had a larger APD(11.70${\pm}$1.32mmHg/min)compared to CF dialyzers(4.32${\pm}$0.55mmHg/min)(p<0.05). However, there was no observable difference in the UFR between the two dialyzers. Neither APD nor UFR showed any significant increase with an increasing number of reuses for up to more than 20reuses. A substantial number of failures observed in APD(larger than 20mmHe/min)on the reused dialyzers(2 out of 40 CP and S out 26 C-DAK) were attributed to the Possible damage on the fibers. The CF 15-11 HFDs which failed APD test did not show changes in the UFR compared to normal dialyzers indicating that APD is a more sensitive test than UFR test to evaluate the integrity of the fibers. 30527 T00401030527 ^x For quantitative measurement of reflected light from a clinical diagnostic strip, a prototype old reflectance photometer was designed. The strip loader and cassette were made to obtain more accurate reflectance parameters. The strip was illuminated at 45˚c through optical fiber and the intensity of reflected light was determined at rectanguLat angle using a photodiode. The kubelka-munk coefficient and reflection optical density were determined ar four different wavelengths(500,550,570 and 610nm) for blood glucose strip. For higher concentration than 300mg/41 about glucose, a saturation state of abforbance was observed at 500,550 and 570nm. The correlation between glucose concentration and parameters was the best at 610nm. 30535 T00401030535 ^x Radiation-induced fibrosarcoma tumors were grown on the flanks of C3H mice. The mice were divided into two groups. One group was injected with Photofrin II, intravenously (2.5mg/kg body weight). The other group received no Photofrin II. Mice from both groups were irradialed for approximately 15 minutes at 100,300, or 500 mW/cm2 with the argon (488nm/514.5 nm), dye(628nm) and gold vapor (pulsed 628 nm) laser light. A photosensitizer behaved as an added absorber. Under our experimental conditions, the presence of Photolfrin II increased surface temperature by at least 40% and the temperature rise due to 300 mW/cm2 irradiation exceeded values for hyperthermia. Light and temperature distributions with depth were estimated by a computer model. The model demonstrated the influence of wavelength on the thermal process and proved to be a valuable tool to investigate internal temperature rise. 30536 T00401030536 ^x We investigated the structural geometry of thirty-eight Korean femurs. The purpose of this study is to identify major geometrical differences between Korean femurs 3nd others that we believe belong to Caucasians so that we would be able to get insights into the femoral component design that fits Asians including Koreans. We utilized computerized tomography (CT) images of femurs extracted from cadavers. The CT images were transformed into bitmap data by using a film scanner, and then analyzed by using a commercially available software called Image v.1.0 and a Macintosh IIci computer.The resulting data were compared with already published data. The major results show that the geometry of the Korean femurs is significantly different from that of Caucasians: (1) the anteversion angle and the canal flare index are greater by the amount of approximately 8˚ and 0.5, respectively, (2) the shape of the isthmus cross section is more round, and (3) the distance between the teaser trochanter and the proximal border of the isthmus is shelter by about 15 mm. The results suggested that the femoral component suitable for Asians should be different from the currently-used components designed and manufactured mostly by European or American companies. 30537 T00401030537 ^x It is well known that nonlinear propagation characteristics of the wave in the tissue may give very useful information for the medical diagnoisis. In this paper, a new method to detect nonlinear propagation characteristics of the internal vibration in the tissue for the low frequency mechanical vibration by using bispectral analysis is proposed. In the method, low frequency vibration of f0( = 100Hz) is applied on the surface of the object, and the waveform of the internal vibration x (t) is measured from Doppler frequency modulation of silmultaneously transmitted probing ultrasonic waves. Then, the bispectra of the signal x (t) at the frequencies (f0, f0) and (f0, 2f0) are calculated to estimate the nonlinear propagation characteristics as their magnitude ratio, w here since bispectrum is free from the gaussian additive noise we can get the value with high S/N. Basic experimental system is constructed by using 3.0 MHz probing ultrasonic waves and the several experiments are carried out for some phantoms. Results show the superiority of the proposed method to the conventional method using power spectrum and also its usefulness for the tissue characterization. 30541 T00401030541 ^x This paper describes the implementation of a computerized radial pulse diagnosis by aids of a clinical expert. On this base, we composed of the radial pulse diagnosis system in korean traditional medicine. The system composed of a radial pulse wave detection system and a radial pulse diagnosis system. With a detection system, we detected Inyoung and Cheongu radial pulse wave and processed it. Then, we have got the characteristic parameters of radial pulse wave and also quantified that according to the method of Inyoung-Cheongu Comparison Radial Pulse Diagnosis. We defined the jugement standard of radial pulse diagnosis system and then we confirmed the possibility for realization of automatic radial pulse diagnosis in korean traditional medicine. 30545 T00401030545 ^x Microspheres are expected to be applied to biomedical areas such as solid-phase immunoassays, drug delivery systems, immunomagnetic cell separation. To synthesize microspheres for biomedical application, "two stage shot growth method" was developed. The uniformity ratio of synthesized microspheres was always smaller than 1.05. And the surface charge density (or the number of ionizable functional groups) of the microspheres synthesized by "two stage shot growth method" was 6~13 times higher than that of the microspheres synthesized by conventional seeded batch copolymerization. As a previous step for biomedical application, adsorption experiments of bovine albumin on microspheres were carried out under various conditions. The maximum adsorbed amount was obtained in the neighborhood of pH 4.5. Isoelectric point of bovine albumin is pH 5.0, so experimental result shows that it shifted to acid area. The adsorption isotherm was obtained, the plateau region was always reached at 2.Og/L (bulk concentration of bovine albumin).The effect of the kind and the amount of surface functional group was also examined. 30575 T00401030575 ^x A medical image workstation was developed using multimedia technique. The system based on PC-486DX was designed to acquire medical images produced by medical imaging instruments and related audio information, that is, doctors' reporting results. Input information was processed and analyzed, then the results were presented in the form of graph and animation. All the informations of the system were hierarchically related with the image as the apex. Processing and analysis algorithms were implemented so that the diagnostic accuracy could be improved. The diagnosed information can be transferred for patient diagnosis through LAN(local area network). 30592 T00401030592 ^x In the conventional infrared imaging system, complex infrared lens systems are usually used for directing collimated narrow infrared beams into the high speed 2-dimensional optic scanner. In this paper, a simple reflective infrared optic system with a 2-dimensional optic scanner is proposed for the realization of medical infrared thermography system. It has been experimentally proven that the intfrared thermography system composed of the proposed optic system has the temperature resolution of 0.1˚c under the spatial resolution of lmrad, the image matrix size of 256 X 240, and tile imaging time of 4 seconds. 30593 T00401030593 ^x In this paper, MIIS (Medical Image Information System) has been designed and implemented using INGRES RDBMS, which is based on a client/server architecture. The implemented system allows users to register and retrieve patient information, medical images and diagnostic reports. It also provides the function to display these information on workstation windows simultaneously by using the designed menu-driven graphic user interface. The medical image compression/decompression techniques are implemented and integrated into the medical image database system for the efficient data storage and the fast access through the network. 30594 T00401030594 ^x In this paper, computerized BEAM was implemented for the space domain analysis of EEG. Trans-formation from temporal summation to two-dimensional mappings is formed by 4 nearest point inter-polaton method. Methods of representation of BEAM are two. One is dot density method which classify brain electrical potential 9 levels by dot density of gray levels and the other is colour method which classify brain electrical 12 levels by red-green colours. In this BEAM, instantaneous change and average energy distribution over any arbitrary time interval of brain electrical activity could be observed and analyzed easily. In the frequency domain, the distribution of energy spectrum of a special band can easily be distinguished normality and abnormality. 30608 T00401030608 ^x Laboratory information system (LIS) is a key tool to manage laboratory data in clinical pathology. Our department has developed an information system for routine hematology using down-sized computer system. We have used an IBM 486 compatible PC with 16MB main memory, 210 MB hard disk drive, 9 RS-232C port and 24 pin dot printer. The operating system and database management system were SCO UNIX and SCO foxbase, respectively. For program development, we used Xbase language provided by SCO foxbase. The C language was used for interface purpose. To make the system use friendly, pull-down menu was used. The system connected to our hospital information system via application program interface (API), so the information related to patient and request details is automatically transmitted to our computer. Our system interfaced with fwd complete blood count analyzers(Sysmex NE-8000 and Coulter STKS) for unidirectional data tansmission from analyzer to computer. The authors suggests that this system based on down-sized computer could provide a progressive approach to total LIS based on local area network, and the implemented system could serve as a model for other hospital's LIS for routine hematology. 30609 T00401030609 ^x To develop an artificial bone substitute that is gradually degraded and replaced by the regenerated natural bone, the authors designed a composite that is consisted of calcium phosphate and collagen. To use as the structural matrix of the composite, collagen was purified from human umbilical cord. The obtained collagen was treated by pepsin to remove telopeptides, and finally, the immune-free atelocollagen was produced: The cross linked atelocollagen was highly resistant to the collagenase induced collagenolysis. The cross linked collagen demonstrated an improved tensile strength. 30618 T00401030618 ^x This paper is a study on the design of adptive filter for QRS complex detection. We propose a simple adaptive algorithm to increase capability of noise cancelation in QRS complex detection with two stage adaptive filter. At the first stage, background noise is removed and at the next stage, only spectrum of QRS complex components is passed. Two adaptive filters can afford to keep track of the changes of both noise and QRS complex. Each adaptive filter consists of prediction error filter and FIR filter The impulse response of FIR filter uses coefficients of prediction error filter. The detection rates for 105 and 108 of MIT/BIH data base were 99.3% and 97.4% respectively. 30619 T00401030619 ^x To develop an artificial bone substitute that is gradually degraded and replaced by the regenerated natural bone, the authors designed and produced a composite that is consisted of calcium phosphate and collagen. Human umbilical cord origin pepsin treated type I atelocollagen was used as the structural matrix, by which sintered or non-sintered carbonate apatite was encapsulated to form an inorganic-organic composite. With cross linking atelocollagen by UV ray irradiation, the resistance to both compressive and tensile strength was increased. Collagen degradation by the collagenase induced collagenolysis was also decreased. 30620 T00401030620 ^x We have developed a monoleaflet polymer valve as an inexpensive and viable alternative, especially for short-term use in the ventricular assist device or total artificial heart. The frame and leaflet of the polymer valve were made from polyurethane, To evaluate the hemodynamic performance of the polymer valve a comparative study of flow dynamics past a polymer valve and a St. Jude Medical prosthetic valve under physiological pulsatile flow conditions in vitro was made. Comparisons between the valves were made on the transvalvular pressure drop, regurgitation volume and maximum valve opening area. The polymer valve showed smaller regurgitation volume and transvalvular pressure drop compared to the mechanical valve at higher heart rate. The results showed that the functional characteristics of the polymer valve compared favorably with those of the mechanical valve at higher heart rate. 30621 T00401030621 ^x Explosive evaporative removal process of biological tissue by absorption of a CW laser has been simulated by using gelatin and a multimode Nd:YAG laser. Because the point of maximun temperature of laser-irradiated gelatin exists below the surface due to surface cooling, evaporation at the boiling temperature is made explosively from below the surface. The important parameters of this process are the conduction loss to laser power absorption (defined as the conduction-to-laser power parameter, Nk), the convection heat transfer at the surface to conduction loss (defined as Bi), dimensionless extinction coefficient (defined as Br.), and dimensionless irradiation time (defined as Fo). Dependence of Fo on Nk and Bi has been observed by experiment, and the results have been compared with the numerical results obtained by solving a 2-dimensional conduction equation. Fo and explosion depth (from the surface to the point of maximun temperature) are increased when Nk and Bi are increased.To find out the minimum laser power for explosive evaporative removal process, steady state analysis has been also made. The limit of Nk to induce evaporative removal, which is proportional to the inverse of the laser power, has been obtained. 30622 T00401030622 ^x N1 and N2 gross neural action potentials were measured from the round window of the guinea pig cochlea at the onset of the acoustic stimuli. N1-N2 audiograms were made by means of regulating stimulant intensities in order to produce constant N1-N2 potentials as criteria for different input tone pip frequencies. The lowest threshold was measured with an input tone pip I5 dB SPL in intensity and 12 KHz in frequency when the animal was in normal physiological condition. The procedure of experimental measurements is explained in detail. This experimental approach is very useful for the investigation of the Cochlear function. Both noN1inear and active functions of the Cochlea can be monitored by N1-N2 audiograms. 30623 T00401030623 ^x In electrical impedance tomography(EIT), we use boundary current and voltage measurements toprovide the information about the cross-sectional distribution of electrical impedance or resistivity. One of the major problems in EIT has been the inaccessibility of internal voltage or current data in finding the internal impedance values. We propose a new image reconstruction method using internal current density data measured by NMR. We obtained a two-dimensional current density distribution within a phantom by processing the real and imaginary MR images from a 4.77 NMR machine. We implemented a resistivity mage reconstruction algorithm using the finite element method and sensitivity matrix. We presented computer simulation results of the mage reconstruction algorithm and furture direction of the research. 30624 T00401030624 ^x A new method of digital image analysis technique for discrimination of cancer cell was presented in this paper. The object image was the Thyroid eland cells image that was diagnosed as normal and abnormal (two types of abnormal: follicular neoplastic cell, and papillary neoplastic cell), respectively. By using the proposed region segmentation algorithm, the cells were segmented into nucleus. The 16 feature parameters were used to calculate the features of each nucleus. A9 a consequence of using dominant feature parameters method proposed in this paper, discrimination rate of 91.11% was obtained for Thyroid Gland cells. 30625 T00401030625 ^x An electrical stimulator was designed to induce locomotion for paraplegic patients caused by central nervous system injury. Optimal stimulus parameters, which can minimize muscle fatigue and can achieve effective muscle contraction were determined in slow and fast muscles in Sprague-Dawley rats. Stimulus patterns of our stimulator were designed to simulate electromyographic activity monitored during locomotion of normal subjects. Muscle types of the lower extremity were classified according to their mechanical property of contraction, which are slow muscle (msoleus m.) and fast muscle (medial gastrocneminus m., rectus femoris m., vastus lateralis m.). Optimal parameters of electrical stimulation for slow muscles were 20 Hz, 0.2 ms square pulse. For fast muscle, 40 Hz, 0.3 ms square pulse was optimal to produce repeated contraction. Higher stimulus intensity was required when synergistic muscles were stimulated simultaneously than when they were stimulated individually. Electrical stimulation for each muscle was designed to generate bipedal locomotion, so that individual muscles alternate contraction and relaxation to simulate stance and swing phases. Portable electrical stimulator with 16 channels built in microprocessor was constructed and applied to paraplegic patients due to lumbar cord injury. The electrical stimulator restored partially gait function in paraplegic patients. 30626 T00401030626 ^x Two-Dimensional modelling of the Cochlear biomechanics is presented in this paper. The Laplace partial differential equation which represents the fluid mechanics of the Cochlea has been transformed into two-dimensional electrical transmission line. The procedure of this transformation is explained in detail. The comparison between one and two dimensional models is also presented. This electrical modelling of the basilar membrane (BM) is clearly useful for the next approach to the further. Development of active elements which are essential in the producing of the sharp tuning of the BM. This paper shows that two-dimension model is qualitatively better than one-dimensional model both in amplitude and phase responses of the BM displacement. The present model is only for frequency response. However because the model is electrical, the two-dimensional transmission line model can be extended to time response without any difficult. 30627 T00401030627 ^x A method has been proposed for the fully automatic detection of left ventricular endocardial boundary in 2D short axis echocardiogram using geometric model. The procedure has the following three distinct stages. First, the initial center is estimated by the initial center estimation algorithm which is applied to decimated image. Second, the center estimation algorithm is applied to original image and then best-fit elliptic model estimation is processed. Third, best-fit boundary is detected by the cost function which is based on the best-fit elliptic model. The proposed method shows effective result without manual intervention by a human operator. 30628 T00401030628 ^x The intelligent trajectory control method that controls moving direction and average velocity for a prosthetic arm is proposed by pattern recognition and force estimations using EMG signals. Also, we propose the real time trajectory planning method which generates continuous accelleration paths using 3 stage linear filters to minimize the impact to human body induced by arm motions and to reduce the muscle fatigue. We use combination of MLP and fuzzy filter for pattern recognition to estimate the direction of a muscle and Hogan's method for the force estimation. EMG signals are acquired by using a amputation simulator and 2 dimensional joystick motion. The simulation results of proposed prosthetic arm control system using the EMf signals show that the arm is effectively followed the desired trajectory depended on estimated force and direction of muscle movements. 30638 T00401030638 ^x A new neural network architecture for the recognition of patterns from images is proposed, which is partially based on the results of physiological studies. The proposed network is composed of multi-layers and the nerve cells in each layer are connected by spatial filters which approximate receptive fields in optic nerve fields. In the proposed method, patterns recognition for complicated images is carried out using global features as well as local features such as lines and end-points. A new generating method of matched filers representing global features is proposed in this network. 30659 T00401030659 ^x An implementation scheme of the magnetic nerve stimulator using a switching mode power supply is proposed. By using a switching mode power supply rather than a conventional linear power supply for charging high voltage capacitors, the weight and size of the magnetic nerve stimulator can be considerably reduced. Maximum output voltage of the developed magnetic nerve stimulator using the switching mode power supply is 3,000 volts and switching time is about 100 msec. Experimental results or human nerve stimulations using the developed stimulator are presented. 30768 T00401030768 ^x In this paper, we describe the design methodology and specifications of the developed module-based bedside monitors for patient monitoring. The bedside monitor consists of a main unit and module cases with various parameter modules. The main unit includes a 12.1" TFT color LCD, a main CPU board, and peripherals such as a module controller, Ethernet LAN card, video card, rotate/push button controller, etc. The main unit can connect at maximum three module cases each of which can accommodate up to 7 parameter modules. They include the modules for electrocardiograph, respiration, invasive blood pressure, noninvasive blood pressure, temperature, and SpO2 with Plethysmograph.

  • PDF

The Study of Dose Change by Field Effect on Atomic Number of Shielding Materals in 6 MeV Electron Beam (6 MeV 전자선의 차폐물질 원자번호와 조사야 크기에 따른 선량변화 연구)

  • Lee, Seung Hoon;Kwak, Keun Tak;Park, Ju Kyeong;Gim, Yang Soo;Cha, Seok Yong
    • The Journal of Korean Society for Radiation Therapy
    • /
    • v.25 no.2
    • /
    • pp.145-151
    • /
    • 2013
  • Purpose: In this study, we analyzed how the dose change by field size effects on atomic number of shielding materials while using 6 MeV election beam. Materials and Methods: The parallel plate chamber is mounted in $25{\times}25cm^2$ the phantom such that the entrance window of the detector is flush with the phantom surface. phantom was covered laterally with aluminum, copper and lead which thickness have 5% of allowable transmission and then the doses were measured in field size $6{\times}6$, $10{\times}10$ and $20{\times}20cm^2$ respectively. 100 cGy was irradiated using 6 MeV electron beam and SSD (Source Surface Distance) was 100 cm with $10{\times}10cm^2$ field size. To calculate the photon flux, electron flux and Energy deposition produced after pass materals respectively, MCNPX code was used. Results: The results according to the various shielding materials which have 5% of allowable transmission are as in the following. Thickness change rate with field size of $6{\times}6cm^2$ and $20{\times}20cm^2$ that compared to the field size of $10{\times}10cm^2$ found to be +0.06% and -0.06% with aluminum, +0.13% and -0.1% with copper, -1.53% and +1.92% with lead respectively. Compare to the field size $10{\times}10cm^2$, energy deposition for $6{\times}6cm^2$ and $20{\times}20cm^2$ had -4.3% and +4.85% respectively without shielding material. With aluminum it had -0.87% and +6.93% respectively and with lead it had -4.16% and +5.57% respectively. When it comes to photon flux with $6{\times}6cm^2$ and $20{\times}20cm^2$ of field sizes the chance -8.95% and +15.92% without shielding material respectively, with aluminum the number -15.56% and +16.06% respectively and with copper the chance -12.27% and +15.53% respectively, with lead the number +12.36% and -19.81% respectively. In case of electron flux in the same condition, the number -3.92% and +4.55% respectively without shielding material respectively, with aluminum the number +0.59% and +6.87% respectively, with copper the number -1.59% and +3.86% respectively, with lead the chance -5.15% and +4.00% respectively. Conclusion: In this study, we found that the required thickness of the shielding materials got thinner with low atomic number substance as the irradiation field is increasing. On the other hand, with high atomic number substance the required thickness had increased. In addition, bremsstrahlung radiation have an influence on low atomic number materials and high atomic number materials are effected by scattered electrons.

  • PDF

Methodology for Identifying Issues of User Reviews from the Perspective of Evaluation Criteria: Focus on a Hotel Information Site (사용자 리뷰의 평가기준 별 이슈 식별 방법론: 호텔 리뷰 사이트를 중심으로)

  • Byun, Sungho;Lee, Donghoon;Kim, Namgyu
    • Journal of Intelligence and Information Systems
    • /
    • v.22 no.3
    • /
    • pp.23-43
    • /
    • 2016
  • As a result of the growth of Internet data and the rapid development of Internet technology, "big data" analysis has gained prominence as a major approach for evaluating and mining enormous data for various purposes. Especially, in recent years, people tend to share their experiences related to their leisure activities while also reviewing others' inputs concerning their activities. Therefore, by referring to others' leisure activity-related experiences, they are able to gather information that might guarantee them better leisure activities in the future. This phenomenon has appeared throughout many aspects of leisure activities such as movies, traveling, accommodation, and dining. Apart from blogs and social networking sites, many other websites provide a wealth of information related to leisure activities. Most of these websites provide information of each product in various formats depending on different purposes and perspectives. Generally, most of the websites provide the average ratings and detailed reviews of users who actually used products/services, and these ratings and reviews can actually support the decision of potential customers in purchasing the same products/services. However, the existing websites offering information on leisure activities only provide the rating and review based on one stage of a set of evaluation criteria. Therefore, to identify the main issue for each evaluation criterion as well as the characteristics of specific elements comprising each criterion, users have to read a large number of reviews. In particular, as most of the users search for the characteristics of the detailed elements for one or more specific evaluation criteria based on their priorities, they must spend a great deal of time and effort to obtain the desired information by reading more reviews and understanding the contents of such reviews. Although some websites break down the evaluation criteria and direct the user to input their reviews according to different levels of criteria, there exist excessive amounts of input sections that make the whole process inconvenient for the users. Further, problems may arise if a user does not follow the instructions for the input sections or fill in the wrong input sections. Finally, treating the evaluation criteria breakdown as a realistic alternative is difficult, because identifying all the detailed criteria for each evaluation criterion is a challenging task. For example, if a review about a certain hotel has been written, people tend to only write one-stage reviews for various components such as accessibility, rooms, services, or food. These might be the reviews for most frequently asked questions, such as distance between the nearest subway station or condition of the bathroom, but they still lack detailed information for these questions. In addition, in case a breakdown of the evaluation criteria was provided along with various input sections, the user might only fill in the evaluation criterion for accessibility or fill in the wrong information such as information regarding rooms in the evaluation criteria for accessibility. Thus, the reliability of the segmented review will be greatly reduced. In this study, we propose an approach to overcome the limitations of the existing leisure activity information websites, namely, (1) the reliability of reviews for each evaluation criteria and (2) the difficulty of identifying the detailed contents that make up the evaluation criteria. In our proposed methodology, we first identify the review content and construct the lexicon for each evaluation criterion by using the terms that are frequently used for each criterion. Next, the sentences in the review documents containing the terms in the constructed lexicon are decomposed into review units, which are then reconstructed by using the evaluation criteria. Finally, the issues of the constructed review units by evaluation criteria are derived and the summary results are provided. Apart from the derived issues, the review units are also provided. Therefore, this approach aims to help users save on time and effort, because they will only be reading the relevant information they need for each evaluation criterion rather than go through the entire text of review. Our proposed methodology is based on the topic modeling, which is being actively used in text analysis. The review is decomposed into sentence units rather than considering the whole review as a document unit. After being decomposed into individual review units, the review units are reorganized according to each evaluation criterion and then used in the subsequent analysis. This work largely differs from the existing topic modeling-based studies. In this paper, we collected 423 reviews from hotel information websites and decomposed these reviews into 4,860 review units. We then reorganized the review units according to six different evaluation criteria. By applying these review units in our methodology, the analysis results can be introduced, and the utility of proposed methodology can be demonstrated.

Spatio-Temporal Monitoring of Soil CO2 Fluxes and Concentrations after Artificial CO2 Release (인위적 CO2 누출에 따른 토양 CO2 플럭스와 농도의 시공간적 모니터링)

  • Kim, Hyun-Jun;Han, Seung Hyun;Kim, Seongjun;Yun, Hyeon Min;Jun, Seong-Chun;Son, Yowhan
    • Journal of Environmental Impact Assessment
    • /
    • v.26 no.2
    • /
    • pp.93-104
    • /
    • 2017
  • CCS (Carbon Capture and Storage) is a technical process to capture $CO_2$ from industrial and energy-based sources, to transfer and sequestrate impressed $CO_2$ in geological formations, oceans, or mineral carbonates. However, potential $CO_2$ leakage exists and causes environmental problems. Thus, this study was conducted to analyze the spatial and temporal variations of $CO_2$ fluxes and concentrations after artificial $CO_2$ release. The Environmental Impact Evaluation Test Facility (EIT) was built in Eumseong, Korea in 2015. Approximately 34kg $CO_2$ /day/zone were injected at Zones 2, 3, and 4 among the total of 5 zones from October 26 to 30, 2015. $CO_2$ fluxes were measured every 30 minutes at the surface at 0m, 1.5m, 2.5m, and 10m from the $CO_2$ releasing well using LI-8100A until November 13, 2015, and $CO_2$ concentrations were measured once a day at 15cm, 30cm, and 60cm depths at every 0m, 1.5m, 2.5m, 5m, and 10m from the well using GA5000 until November 28, 2015. $CO_2$ flux at 0m from the well started increasing on the fifth day after $CO_2$ release started, and continued to increase until November 13 even though the artificial $CO_2$ release stopped. $CO_2$ fluxes measured at 2.5m, 5.0m, and 10m from the well were not significantly different with each other. On the other hand, soil $CO_2$ concentration was shown as 38.4% at 60cm depth at 0m from the well in Zone 3 on the next day after $CO_2$ release started. Soil $CO_2$ was horizontally spreaded overtime, and detected up to 5m away from the well in all zones until $CO_2$ release stopped. Also, soil $CO_2$ concentrations at 30cm and 60cm depths at 0m from the well were measured similarly as $50.6{\pm}25.4%$ and $55.3{\pm}25.6%$, respectively, followed by 30cm depth ($31.3{\pm}17.2%$) which was significantly lower than those measured at the other depths on the final day of $CO_2$ release period. Soil $CO_2$ concentrations at all depths in all zones were gradually decreased for about 1 month after $CO_2$ release stopped, but still higher than those of the first day after $CO_2$ release stared. In conclusion, the closer the distance from the well and the deeper the depth, the higher $CO_2$ fluxes and concentrations occurred. Also, long-term monitoring should be required because the leaked $CO_2$ gas can remains in the soil for a long time even if the leakage stopped.

Study on the Difference in Intake Rate by Kidney in Accordance with whether the Bladder is Shielded and Injection method in 99mTc-DMSA Renal Scan for Infants (소아 99mTc-DMSA renal scan에서 방광차폐유무와 방사성동위원소 주입방법에 따른 콩팥섭취율 차이에 관한 연구)

  • Park, Jeong Kyun;Cha, Jae Hoon;Kim, Kwang Hyun;An, Jong Ki;Hong, Da Young;Seong, Hyo Jin
    • The Korean Journal of Nuclear Medicine Technology
    • /
    • v.20 no.2
    • /
    • pp.27-31
    • /
    • 2016
  • Purpose $^{99m}Tc-DMSA$ renal scan is a test for the comparison of the function by imaging the parenchyma of the kidneys by the cortex of a kidney and by computing the intake ratio of radiation by the left and right kidney. Since the distance between the kidneys and the bladder is not far given the bodily structure of an infant, the bladder is included in the examination domain. Research was carried out with the presumption that counts of bladder would impart an influence on the kidneys at the time of this renal scan. In consideration of the special feature that only a trace amount of a RI is injected in a pediatric examination, research on the method of injection was also carried out concurrently. Materials and Methods With 34 infants aged between 1 month to 12 months for whom a $^{99m}Tc-DMSA$ renal scan was implemented on the subjects, a Post IMAGE was acquired in accordance with the test time after having injected the same quantity of DMSA of 0.5mCi. Then, after having acquired an additional image by shielding the bladder by using a circular lead plate for comparison purposes, a comparison was made by illustrating the percentile of (Lt. Kidney counts + Rt. Kidney counts)/ Total counts, by drawing the same sized ROI (length of 55.2mm X width of 70.0mm). In addition, in the format of a 3-way stopcock, a Heparin cap and direct injection into the patient were performed in accordance with RI injection methods. The differences in the count changes in accordance with each of the methods were compared by injecting an additional 2cc of saline into the 3-way stopcock and Heparin cap. Results The image prior to shielding of the bladder displayed a kidney intake rate with a deviation of $70.9{\pm}3.18%$ while the image after the shielding of the bladder displayed a kidney intake rate with a deviation of $79.4{\pm}5.19%$, thereby showing approximately 6.5~8.5% of difference. In terms of the injection method, the method that used the 3-way form, a deviation of $68.9{\pm}2.80%$ prior to the shielding and a deviation of $78.1{\pm}5.14%$ after the shielding were displayed. In the method of using a Heparin cap, a deviation of $71.3{\pm}5.14%$ prior to the shielding and a deviation of $79.8{\pm}3.26%$ after the shielding were displayed. Lastly, in the method of direct injection into the patient, a deviation of $75.1{\pm}4.30%$ prior to the shielding and a deviation of $82.1{\pm}2.35%$ after the shielding were displayed, thereby illustrating differences in the kidney intake rates in the order of direct injection, a Heparin cap and the 3-way methods. Conclusion Since a substantially minute quantity of radiopharmaceuticals is injected for infants in comparison to adults, the cases of having shielded the bladder by removing radiation of the bladder displayed kidney intake rates that are improved from those of the cases of not having shielded the bladder. Although there are difficulties in securing blood vessels, it is deemed that the method of direct injection would be more helpful in acquisition of better images since it displays improved kidney intake rate in comparison to other methods.

  • PDF

The Effect of Common Features on Consumer Preference for a No-Choice Option: The Moderating Role of Regulatory Focus (재몰유선택적정황하공동특성대우고객희호적영향(在没有选择的情况下共同特性对于顾客喜好的影响): 조절초점적조절작용(调节焦点的调节作用))

  • Park, Jong-Chul;Kim, Kyung-Jin
    • Journal of Global Scholars of Marketing Science
    • /
    • v.20 no.1
    • /
    • pp.89-97
    • /
    • 2010
  • This study researches the effects of common features on a no-choice option with respect to regulatory focus theory. The primary interest is in three factors and their interrelationship: common features, no-choice option, and regulatory focus. Prior studies have compiled vast body of research in these areas. First, the "common features effect" has been observed bymany noted marketing researchers. Tversky (1972) proposed the seminal theory, the EBA model: elimination by aspect. According to this theory, consumers are prone to focus only on unique features during comparison processing, thereby dismissing any common features as redundant information. Recently, however, more provocative ideas have attacked the EBA model by asserting that common features really do affect consumer judgment. Chernev (1997) first reported that adding common features mitigates the choice gap because of the increasing perception of similarity among alternatives. Later, however, Chernev (2001) published a critically developed study against his prior perspective with the proposition that common features may be a cognitive load to consumers, and thus consumers are possible that they are prone to prefer the heuristic processing to the systematic processing. This tends to bring one question to the forefront: Do "common features" affect consumer choice? If so, what are the concrete effects? This study tries to answer the question with respect to the "no-choice" option and regulatory focus. Second, some researchers hold that the no-choice option is another best alternative of consumers, who are likely to avoid having to choose in the context of knotty trade-off settings or mental conflicts. Hope for the future also may increase the no-choice option in the context of optimism or the expectancy of a more satisfactory alternative appearing later. Other issues reported in this domain are time pressure, consumer confidence, and alternative numbers (Dhar and Nowlis 1999; Lin and Wu 2005; Zakay and Tsal 1993). This study casts the no-choice option in yet another perspective: the interactive effects between common features and regulatory focus. Third, "regulatory focus theory" is a very popular theme in recent marketing research. It suggests that consumers have two focal goals facing each other: promotion vs. prevention. A promotion focus deals with the concepts of hope, inspiration, achievement, or gain, whereas prevention focus involves duty, responsibility, safety, or loss-aversion. Thus, while consumers with a promotion focus tend to take risks for gain, the same does not hold true for a prevention focus. Regulatory focus theory predicts consumers' emotions, creativity, attitudes, memory, performance, and judgment, as documented in a vast field of marketing and psychology articles. The perspective of the current study in exploring consumer choice and common features is a somewhat creative viewpoint in the area of regulatory focus. These reviews inspire this study of the interaction possibility between regulatory focus and common features with a no-choice option. Specifically, adding common features rather than omitting them may increase the no-choice option ratio in the choice setting only to prevention-focused consumers, but vice versa to promotion-focused consumers. The reasoning is that when prevention-focused consumers come in contact with common features, they may perceive higher similarity among the alternatives. This conflict among similar options would increase the no-choice ratio. Promotion-focused consumers, however, are possible that they perceive common features as a cue of confirmation bias. And thus their confirmation processing would make their prior preference more robust, then the no-choice ratio may shrink. This logic is verified in two experiments. The first is a $2{\times}2$ between-subject design (whether common features or not X regulatory focus) using a digital cameras as the relevant stimulus-a product very familiar to young subjects. Specifically, the regulatory focus variable is median split through a measure of eleven items. Common features included zoom, weight, memory, and battery, whereas the other two attributes (pixel and price) were unique features. Results supported our hypothesis that adding common features enhanced the no-choice ratio only to prevention-focus consumers, not to those with a promotion focus. These results confirm our hypothesis - the interactive effects between a regulatory focus and the common features. Prior research had suggested that including common features had a effect on consumer choice, but this study shows that common features affect choice by consumer segmentation. The second experiment was used to replicate the results of the first experiment. This experimental study is equal to the prior except only two - priming manipulation and another stimulus. For the promotion focus condition, subjects had to write an essay using words such as profit, inspiration, pleasure, achievement, development, hedonic, change, pursuit, etc. For prevention, however, they had to use the words persistence, safety, protection, aversion, loss, responsibility, stability etc. The room for rent had common features (sunshine, facility, ventilation) and unique features (distance time and building state). These attributes implied various levels and valence for replication of the prior experiment. Our hypothesis was supported repeatedly in the results, and the interaction effects were significant between regulatory focus and common features. Thus, these studies showed the dual effects of common features on consumer choice for a no-choice option. Adding common features may enhance or mitigate no-choice, contradictory as it may sound. Under a prevention focus, adding common features is likely to enhance the no-choice ratio because of increasing mental conflict; under the promotion focus, it is prone to shrink the ratio perhaps because of a "confirmation bias." The research has practical and theoretical implications for marketers, who may need to consider common features carefully in a practical display context according to consumer segmentation (i.e., promotion vs. prevention focus.) Theoretically, the results suggest some meaningful moderator variable between common features and no-choice in that the effect on no-choice option is partly dependent on a regulatory focus. This variable corresponds not only to a chronic perspective but also a situational perspective in our hypothesis domain. Finally, in light of some shortcomings in the research, such as overlooked attribute importance, low ratio of no-choice, or the external validity issue, we hope it influences future studies to explore the little-known world of the "no-choice option."

A New Approach to Automatic Keyword Generation Using Inverse Vector Space Model (키워드 자동 생성에 대한 새로운 접근법: 역 벡터공간모델을 이용한 키워드 할당 방법)

  • Cho, Won-Chin;Rho, Sang-Kyu;Yun, Ji-Young Agnes;Park, Jin-Soo
    • Asia pacific journal of information systems
    • /
    • v.21 no.1
    • /
    • pp.103-122
    • /
    • 2011
  • Recently, numerous documents have been made available electronically. Internet search engines and digital libraries commonly return query results containing hundreds or even thousands of documents. In this situation, it is virtually impossible for users to examine complete documents to determine whether they might be useful for them. For this reason, some on-line documents are accompanied by a list of keywords specified by the authors in an effort to guide the users by facilitating the filtering process. In this way, a set of keywords is often considered a condensed version of the whole document and therefore plays an important role for document retrieval, Web page retrieval, document clustering, summarization, text mining, and so on. Since many academic journals ask the authors to provide a list of five or six keywords on the first page of an article, keywords are most familiar in the context of journal articles. However, many other types of documents could not benefit from the use of keywords, including Web pages, email messages, news reports, magazine articles, and business papers. Although the potential benefit is large, the implementation itself is the obstacle; manually assigning keywords to all documents is a daunting task, or even impractical in that it is extremely tedious and time-consuming requiring a certain level of domain knowledge. Therefore, it is highly desirable to automate the keyword generation process. There are mainly two approaches to achieving this aim: keyword assignment approach and keyword extraction approach. Both approaches use machine learning methods and require, for training purposes, a set of documents with keywords already attached. In the former approach, there is a given set of vocabulary, and the aim is to match them to the texts. In other words, the keywords assignment approach seeks to select the words from a controlled vocabulary that best describes a document. Although this approach is domain dependent and is not easy to transfer and expand, it can generate implicit keywords that do not appear in a document. On the other hand, in the latter approach, the aim is to extract keywords with respect to their relevance in the text without prior vocabulary. In this approach, automatic keyword generation is treated as a classification task, and keywords are commonly extracted based on supervised learning techniques. Thus, keyword extraction algorithms classify candidate keywords in a document into positive or negative examples. Several systems such as Extractor and Kea were developed using keyword extraction approach. Most indicative words in a document are selected as keywords for that document and as a result, keywords extraction is limited to terms that appear in the document. Therefore, keywords extraction cannot generate implicit keywords that are not included in a document. According to the experiment results of Turney, about 64% to 90% of keywords assigned by the authors can be found in the full text of an article. Inversely, it also means that 10% to 36% of the keywords assigned by the authors do not appear in the article, which cannot be generated through keyword extraction algorithms. Our preliminary experiment result also shows that 37% of keywords assigned by the authors are not included in the full text. This is the reason why we have decided to adopt the keyword assignment approach. In this paper, we propose a new approach for automatic keyword assignment namely IVSM(Inverse Vector Space Model). The model is based on a vector space model. which is a conventional information retrieval model that represents documents and queries by vectors in a multidimensional space. IVSM generates an appropriate keyword set for a specific document by measuring the distance between the document and the keyword sets. The keyword assignment process of IVSM is as follows: (1) calculating the vector length of each keyword set based on each keyword weight; (2) preprocessing and parsing a target document that does not have keywords; (3) calculating the vector length of the target document based on the term frequency; (4) measuring the cosine similarity between each keyword set and the target document; and (5) generating keywords that have high similarity scores. Two keyword generation systems were implemented applying IVSM: IVSM system for Web-based community service and stand-alone IVSM system. Firstly, the IVSM system is implemented in a community service for sharing knowledge and opinions on current trends such as fashion, movies, social problems, and health information. The stand-alone IVSM system is dedicated to generating keywords for academic papers, and, indeed, it has been tested through a number of academic papers including those published by the Korean Association of Shipping and Logistics, the Korea Research Academy of Distribution Information, the Korea Logistics Society, the Korea Logistics Research Association, and the Korea Port Economic Association. We measured the performance of IVSM by the number of matches between the IVSM-generated keywords and the author-assigned keywords. According to our experiment, the precisions of IVSM applied to Web-based community service and academic journals were 0.75 and 0.71, respectively. The performance of both systems is much better than that of baseline systems that generate keywords based on simple probability. Also, IVSM shows comparable performance to Extractor that is a representative system of keyword extraction approach developed by Turney. As electronic documents increase, we expect that IVSM proposed in this paper can be applied to many electronic documents in Web-based community and digital library.

A Mobile Landmarks Guide : Outdoor Augmented Reality based on LOD and Contextual Device (모바일 랜드마크 가이드 : LOD와 문맥적 장치 기반의 실외 증강현실)

  • Zhao, Bi-Cheng;Rosli, Ahmad Nurzid;Jang, Chol-Hee;Lee, Kee-Sung;Jo, Geun-Sik
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.1
    • /
    • pp.1-21
    • /
    • 2012
  • In recent years, mobile phone has experienced an extremely fast evolution. It is equipped with high-quality color displays, high resolution cameras, and real-time accelerated 3D graphics. In addition, some other features are includes GPS sensor and Digital Compass, etc. This evolution advent significantly helps the application developers to use the power of smart-phones, to create a rich environment that offers a wide range of services and exciting possibilities. To date mobile AR in outdoor research there are many popular location-based AR services, such Layar and Wikitude. These systems have big limitation the AR contents hardly overlaid on the real target. Another research is context-based AR services using image recognition and tracking. The AR contents are precisely overlaid on the real target. But the real-time performance is restricted by the retrieval time and hardly implement in large scale area. In our work, we exploit to combine advantages of location-based AR with context-based AR. The system can easily find out surrounding landmarks first and then do the recognition and tracking with them. The proposed system mainly consists of two major parts-landmark browsing module and annotation module. In landmark browsing module, user can view an augmented virtual information (information media), such as text, picture and video on their smart-phone viewfinder, when they pointing out their smart-phone to a certain building or landmark. For this, landmark recognition technique is applied in this work. SURF point-based features are used in the matching process due to their robustness. To ensure the image retrieval and matching processes is fast enough for real time tracking, we exploit the contextual device (GPS and digital compass) information. This is necessary to select the nearest and pointed orientation landmarks from the database. The queried image is only matched with this selected data. Therefore, the speed for matching will be significantly increased. Secondly is the annotation module. Instead of viewing only the augmented information media, user can create virtual annotation based on linked data. Having to know a full knowledge about the landmark, are not necessary required. They can simply look for the appropriate topic by searching it with a keyword in linked data. With this, it helps the system to find out target URI in order to generate correct AR contents. On the other hand, in order to recognize target landmarks, images of selected building or landmark are captured from different angle and distance. This procedure looks like a similar processing of building a connection between the real building and the virtual information existed in the Linked Open Data. In our experiments, search range in the database is reduced by clustering images into groups according to their coordinates. A Grid-base clustering method and user location information are used to restrict the retrieval range. Comparing the existed research using cluster and GPS information the retrieval time is around 70~80ms. Experiment results show our approach the retrieval time reduces to around 18~20ms in average. Therefore the totally processing time is reduced from 490~540ms to 438~480ms. The performance improvement will be more obvious when the database growing. It demonstrates the proposed system is efficient and robust in many cases.

Spatial effect on the diffusion of discount stores (대형할인점 확산에 대한 공간적 영향)

  • Joo, Young-Jin;Kim, Mi-Ae
    • Journal of Distribution Research
    • /
    • v.15 no.4
    • /
    • pp.61-85
    • /
    • 2010
  • Introduction: Diffusion is process by which an innovation is communicated through certain channel overtime among the members of a social system(Rogers 1983). Bass(1969) suggested the Bass model describing diffusion process. The Bass model assumes potential adopters of innovation are influenced by mass-media and word-of-mouth from communication with previous adopters. Various expansions of the Bass model have been conducted. Some of them proposed a third factor affecting diffusion. Others proposed multinational diffusion model and it stressed interactive effect on diffusion among several countries. We add a spatial factor in the Bass model as a third communication factor. Because of situation where we can not control the interaction between markets, we need to consider that diffusion within certain market can be influenced by diffusion in contiguous market. The process that certain type of retail extends is a result that particular market can be described by the retail life cycle. Diffusion of retail has pattern following three phases of spatial diffusion: adoption of innovation happens in near the diffusion center first, spreads to the vicinity of the diffusing center and then adoption of innovation is completed in peripheral areas in saturation stage. So we expect spatial effect to be important to describe diffusion of domestic discount store. We define a spatial diffusion model using multinational diffusion model and apply it to the diffusion of discount store. Modeling: In this paper, we define a spatial diffusion model and apply it to the diffusion of discount store. To define a spatial diffusion model, we expand learning model(Kumar and Krishnan 2002) and separate diffusion process in diffusion center(market A) from diffusion process in the vicinity of the diffusing center(market B). The proposed spatial diffusion model is shown in equation (1a) and (1b). Equation (1a) is the diffusion process in diffusion center and equation (1b) is one in the vicinity of the diffusing center. $$\array{{S_{i,t}=(p_i+q_i{\frac{Y_{i,t-1}}{m_i}})(m_i-Y_{i,t-1})\;i{\in}\{1,{\cdots},I\}\;(1a)}\\{S_{j,t}=(p_j+q_j{\frac{Y_{j,t-1}}{m_i}}+{\sum\limits_{i=1}^I}{\gamma}_{ij}{\frac{Y_{i,t-1}}{m_i}})(m_j-Y_{j,t-1})\;i{\in}\{1,{\cdots},I\},\;j{\in}\{I+1,{\cdots},I+J\}\;(1b)}}$$ We rise two research questions. (1) The proposed spatial diffusion model is more effective than the Bass model to describe the diffusion of discount stores. (2) The more similar retail environment of diffusing center with that of the vicinity of the contiguous market is, the larger spatial effect of diffusing center on diffusion of the vicinity of the contiguous market is. To examine above two questions, we adopt the Bass model to estimate diffusion of discount store first. Next spatial diffusion model where spatial factor is added to the Bass model is used to estimate it. Finally by comparing Bass model with spatial diffusion model, we try to find out which model describes diffusion of discount store better. In addition, we investigate the relationship between similarity of retail environment(conceptual distance) and spatial factor impact with correlation analysis. Result and Implication: We suggest spatial diffusion model to describe diffusion of discount stores. To examine the proposed spatial diffusion model, 347 domestic discount stores are used and we divide nation into 5 districts, Seoul-Gyeongin(SG), Busan-Gyeongnam(BG), Daegu-Gyeongbuk(DG), Gwan- gju-Jeonla(GJ), Daejeon-Chungcheong(DC), and the result is shown

    . In a result of the Bass model(I), the estimates of innovation coefficient(p) and imitation coefficient(q) are 0.017 and 0.323 respectively. While the estimate of market potential is 384. A result of the Bass model(II) for each district shows the estimates of innovation coefficient(p) in SG is 0.019 and the lowest among 5 areas. This is because SG is the diffusion center. The estimates of imitation coefficient(q) in BG is 0.353 and the highest. The imitation coefficient in the vicinity of the diffusing center such as BG is higher than that in the diffusing center because much information flows through various paths more as diffusion is progressing. A result of the Bass model(II) shows the estimates of innovation coefficient(p) in SG is 0.019 and the lowest among 5 areas. This is because SG is the diffusion center. The estimates of imitation coefficient(q) in BG is 0.353 and the highest. The imitation coefficient in the vicinity of the diffusing center such as BG is higher than that in the diffusing center because much information flows through various paths more as diffusion is progressing. In a result of spatial diffusion model(IV), we can notice the changes between coefficients of the bass model and those of the spatial diffusion model. Except for GJ, the estimates of innovation and imitation coefficients in Model IV are lower than those in Model II. The changes of innovation and imitation coefficients are reflected to spatial coefficient(${\gamma}$). From spatial coefficient(${\gamma}$) we can infer that when the diffusion in the vicinity of the diffusing center occurs, the diffusion is influenced by one in the diffusing center. The difference between the Bass model(II) and the spatial diffusion model(IV) is statistically significant with the ${\chi}^2$-distributed likelihood ratio statistic is 16.598(p=0.0023). Which implies that the spatial diffusion model is more effective than the Bass model to describe diffusion of discount stores. So the research question (1) is supported. In addition, we found that there are statistically significant relationship between similarity of retail environment and spatial effect by using correlation analysis. So the research question (2) is also supported.

  • PDF
  • The study on the scattering ratio at the edge of the block according to the increasing block thickness in electron therapy (전자선 치료 시 차폐블록 두께 변화에 따른 블록 주변 선량에 관한 연구)

    • Park, Zi On;Gwak, Geun Tak;Park, Ju Kyeong;Lee, Seung Hun;Kim, Yang Su;Kim, Jung Soo;Kwon, Hyoung Cheol;Lee, Sun Young
      • The Journal of Korean Society for Radiation Therapy
      • /
      • v.31 no.1
      • /
      • pp.57-65
      • /
      • 2019
    • Purpose: The purpose is to clarify the effect of additional scattering ratio on the edge of the block according to the increasing block thickness with low melting point lead alloy and pure lead in electron beam therapy. Methods and materials: $10{\times}10cm^2$ Shielding blocks made of low melting point lead alloy and pure lead were fabricated to shield mold frame half of applicator. Block thickness was 3, 5, 10, 15, 20 (mm) for each material. The common irradiation conditions were set at 6 MeV energy, 300 MU / Min dose rate, gantry angle of $0^{\circ}$, and dose of 100 MU. The relative scattering ratio with increasing block thickness was measured with a parallel plate type ion chamber(Exradin P11) and phantom(RW3) by varying the position of the shielding block(cone and on the phantom), the position of the measuring point(surface ans depth of $D_{max}$), and the block material(lead alloy and pure lead). Results : When (depth of measurement / block position / block material) was (surface / applicator / pure lead), the relative value(scattering ratio) was 15.33 nC(+0.33 %), 15.28 nC(0 %), 15.08 nC(-1.31 %), 15.05 nC(-1.51 %), 15.07 nC(-1.37 %) as the block thickness increased in order of 3, 5, 10, 15, 20 (mm) respectively. When it was (surface / applicator / alloy lead), the relative value(scattering ratio) was 15.19 nC(-0.59 %), 15.25 nC(-0.20 %), 15.15 nC(-0.85 %), 14.96 nC(-2.09 %), 15.15 nC(-0.85 %) respectively. When it was (surface / phantom / pure lead), the relative value(scattering ratio) was 15.62 nC(+2.23 %), 15.59 nC(+2.03 %), 15.53 nC(+1.67 %), 15.48 nC(+1.31 %), 15.34 nC(+0.39 %) respectively. When it was (surface / phantom / alloy lead), the relative value(scattering ratio) was 15.56 nC(+1.83 %), 15.55 nC(+1.77 %), 15.51 nC(+1.51 %), 15.42 nC(+0.92 %), 15.39 nC(+0.72 %) respectively. When it was (depth of $D_{max}$ / applicator / pure lead), the relative value(scattering ratio) was 16.70 nC(-10.87 %), 16.84 nC(-10.12 %), 16.72 nC(-10.78 %), 16.88 nC(-9.93 %), 16.90 nC(-9.82 %) respectively. When it was (depth of $D_{max}$ / applicator / alloy lead), the relative value(scattering ratio) was 16.83 nC(-10.19 %), 17.12 nC(-8.64 %), 16.89 nC(-9.87 %), 16.77 nC(-10.51 %), 16.52 nC(-11.85 %) respectively. When it was (depth of $D_{max}$ / phantom / pure lead), the relative value(scattering ratio) was 17.41 nC(-7.10 %), 17.45 nC(-6.88 %), 17.34 nC(-7.47 %), 17.42 nC(-7.04 %), 17.25 nC(-7.95 %) respectively. When it was (depth of $D_{max}$ / phantom / alloy lead), the relative value(scattering ratio) was 17.45 nC(-6.88 %), 17.44 nC(-6.94 %), 17.47 nC(-6.78 %), 17.43 nC(-6.99 %), 17.35 nC(-7.42 %) respectively. Conclusions: When performing electron therapy using a shielding block, the block position should be inserted applicator rather than the patient's body surface. The block thickness should be made to the minimum appropriate shielding thickness of each corresponding using energy. Also it is useful that the treatment should be performed considering the influence of scattering dose varying with distance from the edge of block.


    (34141) Korea Institute of Science and Technology Information, 245, Daehak-ro, Yuseong-gu, Daejeon
    Copyright (C) KISTI. All Rights Reserved.