• Title/Summary/Keyword: iterative method

Search Result 2,058, Processing Time 0.038 seconds

Clinical Usefulness of PET-MRI in Lymph Node Metastasis Evaluation of Head and Neck Cancer (두경부암 림프절 전이 평가에서 PET-MRI의 임상적 유용성)

  • Kim, Jung-Soo;Lee, Hong-Jae;Kim, Jin-Eui
    • The Korean Journal of Nuclear Medicine Technology
    • /
    • v.18 no.1
    • /
    • pp.26-32
    • /
    • 2014
  • Purpose: As PET-MRI which has excellent soft tissue contrast is developed as integration system, many researches about clinical application are being conducted by comparing with existing display equipments. Because PET-MRI is actively used for head and neck cancer diagnosis in our hospital, lymph node metastasis before the patient's surgery was diagnosed and clinical usefulness of head and neck cancer PET-MRI scan was evaluated using pathological opinions and idiopathy surrounding tissue metastasis evaluation method. Materials and Methods: Targeting 100 head and neck cancer patients in SNUH from January to August in 2013. $^{18}F-FDG$ (5.18 MBq/kg) was intravenous injected and after 60 min of rest, torso (body TIM coil, Vibe-Dixon) and dedication (head-neck TIM coil, UTE, Dotarem injection) scans were conducted using $Bio-graph^{TM}$ mMR 3T (SIEMENS, Munich). Data were reorganized using iterative reconstruction and lymph node metastasis was read with Syngo.Via workstation. Subsequently, pathological observations and diagnosis before-and-after surgery were examined with integrated medical information system (EMR, best-care) in SNUH. Patient's diagnostic information was entered in each category of $2{\times}2$ decision matrix and was classified into true positive (TP), true negative (TN), false positive (FP) and false negative (FN). Based on these classified test results, sensitivity, specificity, accuracy, false negative and false positive rate were calculated. Results: In PET-MRI scan results of head and neck cancer patients, positive and negative cases of lymph node metastasis were 49 and 51 cases respectively and positive and negative lymph node metastasis through before-and-after surgery pathological results were 46 and 54 cases respectively. In both tests, TP which received positive lymph node metastasis were analyzed as 34 cases, FP which received positive lymph node metastasis in PET-MRI scan but received negative lymph node metastasis in pathological test were 4 cases, FN which received negative lymph node metastasis but received positive lymph node metastasis in pathological test was 1 case, and TN which received negative lymph node metastasis in both two tests were 50 cases. Based on these data, sensitivity in PET-MRI scan of head and neck cancer patient was identified to be 97.8%, specificity was 92.5%, accuracy was 95%, FN rate was 2.1% and FP rate was 7.00% respectively. Conclusion: PET-MRI which can apply the acquired functional information using high tissue contrast and various sequences was considered to be useful in determining the weapons before-and-after surgery in head and neck cancer diagnosis or in the evaluation of recurrence and remote detection of metastasis and uncertain idiopathy cervical lymph node metastasis. Additionally, clinical usefulness of PET-MRI through pathological test and integrated diagnosis and follow-up scan was considered to be sufficient as a standard diagnosis scan of head and neck cancer, and additional researches about the development of optimum MR sequence and clinical application are required.

  • PDF

Evaluation of Image Quality in Micro-CT System Using Constrained Total Variation (TV) Minimization (Micro-CT 시스템에서 제한된 조건의 Total Variation (TV) Minimization을 이용한 영상화질 평가)

  • Jo, Byung-Du;Choi, Jong-Hwa;Kim, Yun-Hwan;Lee, Kyung-Ho;Kim, Dae-Hong;Kim, Hee-Joung
    • Progress in Medical Physics
    • /
    • v.23 no.4
    • /
    • pp.252-260
    • /
    • 2012
  • The reduction of radiation dose from x-ray is a main concern in computed tomography (CT) imaging due to the side-effect of the dose on human body. Recently, the various methods for dose reduction have been studied in CT and one of the method is a iterative reconstruction based on total variation (TV) minimization at few-views data. In this paper, we evaluated the image quality between total variation (TV) minimization algorithm and Feldkam-Davis-kress (FDK) algorithm in micro computed tomography (CT). To evaluate the effect of TV minimization algorithm, we produced a cylindrical phantom including contrast media, water, air inserts. We can acquire maximum 400 projection views per rotation of the x-ray tube and detector. 20, 50, 90, 180 projection data were chosen for evaluating the level of image restoration by TV minimization. The phantom and mouse image reconstructed with FDK algorithm at 400 projection data used as a reference image for comparing with TV minimization and FDK algorithm at few-views. Contrast-to-noise ratio (CNR), Universal quality index (UQI) were used as a image evaluation metric. When projection data are not insufficient, our results show that the image quality of reconstructed with TV minimization is similar to reconstructed image with FDK at 400 view. In the cylindrical phantom study, the CNR of TV image was 5.86, FDK image was 5.65 and FDK-reference was 5.98 at 90-views. The CNR of TV image 0.21 higher than FDK image CNR at 90-views. UQI of TV image was 0.99 and FDK image was 0.81 at 90-views. where, the number of projection is 90, the UQI of TV image 0.18 higher than FDK image at 90-views. In the mouse study UQI of TV image was 0.91, FDK was 0.83 at 90-views. the UQI of TV image 0.08 higher than FDK image at 90-views. In cylindrical phantom image and mouse image study, TV minimization algorithm shows the best performance in artifact reduction and preserving edges at few view data. Therefore, TV minimization can potentially be expected to reduce patient dose in clinics.

Estimate on the Crustal Thickness from Using Multi-geophysical Data Sets and Its Comparison to Heat Flow Distribution of Korean Peninsula (다양한 지구물리 자료를 통해 얻은 한반도의 지각두께 예측과 지열류량과의 비교)

  • Choi, Soon-Young;Kim, Hyung-Rae;Kim, Chang-Hwan;Park, Chan-Hong;Suh, Man-Chul
    • Economic and Environmental Geology
    • /
    • v.44 no.6
    • /
    • pp.493-502
    • /
    • 2011
  • We study the deep structure of Korean Peninsula by estimating Moho depth and crustal thickness from using land and oceanic topography and free-air gravity anomaly data. Based on Airy-Heiskanen isostatic hypothesis, the correlated components between the terrain gravity effects and free-air gravity anomalies by wavenumber correlation analysis(WCA) are extracted to estimate the gravity effects that will be resulted from isostatic compensation for the area. With the resulting compensated gravity estimates, Moho depth that is a subsurface between the crust and mantle is estimated by the inversion in an iterative method with the constraints of 20 seismic depth estimates by the receiver function analysis, to minimize the uncertainty of non-uniqueness. Consequently, the average of the resulting crustal thickness estimate of Korean Peninsula is 32.15 km and the standard deviation is 3.12 km. Moho depth of South Korea estimated from this study is compared with the ones from the previous studies, showing they are approximately consistent. And the aspects of Moho undulation from the respective study are in common deep along Taebaek Mountains and Sobaek Mountains and low depth in Gyeongsang Basin relatively. Also, it is discussed that the terrain decorrelated free-air gravity anomalies inferring from the intracrustal characteristics of the crust are compared to the heat flow distributions of South Korea. The low-frequency components of terrain decorrelated Free-air gravity anomalies are highly correlated with the heat flow data, especially in the area of Gyeongsang basin where high heat flow causes to decrease the density of the rocks in the lower crust resulting in lowering the Moho depth by compensation. This result confirms that the high heat sources in this area coming from the upper mantle by Kim et al. (2008).

Assessment of Attenuation Correction Techniques with a $^{137}Cs$ Point Source ($^{137}Cs$ 점선원을 이용한 감쇠 보정기법들의 평가)

  • Bong, Jung-Kyun;Kim, Hee-Joung;Son, Hye-Kyoung;Park, Yun-Young;Park, Hae-Joung;Yun, Mi-Jin;Lee, Jong-Doo;Jung, Hae-Jo
    • The Korean Journal of Nuclear Medicine
    • /
    • v.39 no.1
    • /
    • pp.57-68
    • /
    • 2005
  • Purpose: The objective of this study was to assess attenuation correction algorithms with the $^{137}Cs$ point source for the brain positron omission tomography (PET) imaging process. Materials & Methods: Four different types of phantoms were used in this study for testing various types of the attenuation correction techniques. Transmission data of a $^{137}Cs$ point source were acquired after infusing the emission source into phantoms and then the emission data were subsequently acquired in 3D acquisition mode. Scatter corrections were performed with a background tail-fitting algorithm. Emission data were then reconstructed using iterative reconstruction method with a measured (MAC), elliptical (ELAC), segmented (SAC) and remapping (RAC) attenuation correction, respectively. Reconstructed images were then both qualitatively and quantitatively assessed. In addition, reconstructed images of a normal subject were assessed by nuclear medicine physicians. Subtracted images were also compared. Results: ELEC, SAC, and RAC provided a uniform phantom image with less noise for a cylindrical phantom. In contrast, a decrease in intensity at the central portion of the attenuation map was noticed at the result of the MAC. Reconstructed images of Jaszack and Hoffan phantoms presented better quality with RAC and SAC. The attenuation of a skull on images of the normal subject was clearly noticed and the attenuation correction without considering the attenuation of the skull resulted in artificial defects on images of the brain. Conclusion: the complicated and improved attenuation correction methods were needed to obtain the better accuracy of the quantitative brain PET images.

A Study on Interactions of Competitive Promotions Between the New and Used Cars (신차와 중고차간 프로모션의 상호작용에 대한 연구)

  • Chang, Kwangpil
    • Asia Marketing Journal
    • /
    • v.14 no.1
    • /
    • pp.83-98
    • /
    • 2012
  • In a market where new and used cars are competing with each other, we would run the risk of obtaining biased estimates of cross elasticity between them if we focus on only new cars or on only used cars. Unfortunately, most of previous studies on the automobile industry have focused on only new car models without taking into account the effect of used cars' pricing policy on new cars' market shares and vice versa, resulting in inadequate prediction of reactive pricing in response to competitors' rebate or price discount. However, there are some exceptions. Purohit (1992) and Sullivan (1990) looked into both new and used car markets at the same time to examine the effect of new car model launching on the used car prices. But their studies have some limitations in that they employed the average used car prices reported in NADA Used Car Guide instead of actual transaction prices. Some of the conflicting results may be due to this problem in the data. Park (1998) recognized this problem and used the actual prices in his study. His work is notable in that he investigated the qualitative effect of new car model launching on the pricing policy of the used car in terms of reinforcement of brand equity. The current work also used the actual price like Park (1998) but the quantitative aspect of competitive price promotion between new and used cars of the same model was explored. In this study, I develop a model that assumes that the cross elasticity between new and used cars of the same model is higher than those amongst new cars and used cars of the different model. Specifically, I apply the nested logit model that assumes the car model choice at the first stage and the choice between new and used cars at the second stage. This proposed model is compared to the IIA (Independence of Irrelevant Alternatives) model that assumes that there is no decision hierarchy but that new and used cars of the different model are all substitutable at the first stage. The data for this study are drawn from Power Information Network (PIN), an affiliate of J.D. Power and Associates. PIN collects sales transaction data from a sample of dealerships in the major metropolitan areas in the U.S. These are retail transactions, i.e., sales or leases to final consumers, excluding fleet sales and including both new car and used car sales. Each observation in the PIN database contains the transaction date, the manufacturer, model year, make, model, trim and other car information, the transaction price, consumer rebates, the interest rate, term, amount financed (when the vehicle is financed or leased), etc. I used data for the compact cars sold during the period January 2009- June 2009. The new and used cars of the top nine selling models are included in the study: Mazda 3, Honda Civic, Chevrolet Cobalt, Toyota Corolla, Hyundai Elantra, Ford Focus, Volkswagen Jetta, Nissan Sentra, and Kia Spectra. These models in the study accounted for 87% of category unit sales. Empirical application of the nested logit model showed that the proposed model outperformed the IIA (Independence of Irrelevant Alternatives) model in both calibration and holdout samples. The other comparison model that assumes choice between new and used cars at the first stage and car model choice at the second stage turned out to be mis-specfied since the dissimilarity parameter (i.e., inclusive or categroy value parameter) was estimated to be greater than 1. Post hoc analysis based on estimated parameters was conducted employing the modified Lanczo's iterative method. This method is intuitively appealing. For example, suppose a new car offers a certain amount of rebate and gains market share at first. In response to this rebate, a used car of the same model keeps decreasing price until it regains the lost market share to maintain the status quo. The new car settle down to a lowered market share due to the used car's reaction. The method enables us to find the amount of price discount to main the status quo and equilibrium market shares of the new and used cars. In the first simulation, I used Jetta as a focal brand to see how its new and used cars set prices, rebates or APR interactively assuming that reactive cars respond to price promotion to maintain the status quo. The simulation results showed that the IIA model underestimates cross elasticities, resulting in suggesting less aggressive used car price discount in response to new cars' rebate than the proposed nested logit model. In the second simulation, I used Elantra to reconfirm the result for Jetta and came to the same conclusion. In the third simulation, I had Corolla offer $1,000 rebate to see what could be the best response for Elantra's new and used cars. Interestingly, Elantra's used car could maintain the status quo by offering lower price discount ($160) than the new car ($205). In the future research, we might want to explore the plausibility of the alternative nested logit model. For example, the NUB model that assumes choice between new and used cars at the first stage and brand choice at the second stage could be a possibility even though it was rejected in the current study because of mis-specification (A dissimilarity parameter turned out to be higher than 1). The NUB model may have been rejected due to true mis-specification or data structure transmitted from a typical car dealership. In a typical car dealership, both new and used cars of the same model are displayed. Because of this fact, the BNU model that assumes brand choice at the first stage and choice between new and used cars at the second stage may have been favored in the current study since customers first choose a dealership (brand) then choose between new and used cars given this market environment. However, suppose there are dealerships that carry both new and used cars of various models, then the NUB model might fit the data as well as the BNU model. Which model is a better description of the data is an empirical question. In addition, it would be interesting to test a probabilistic mixture model of the BNU and NUB on a new data set.

  • PDF

A Comparative Study of Subset Construction Methods in OSEM Algorithms using Simulated Projection Data of Compton Camera (모사된 컴프턴 카메라 투사데이터의 재구성을 위한 OSEM 알고리즘의 부분집합 구성법 비교 연구)

  • Kim, Soo-Mee;Lee, Jae-Sung;Lee, Mi-No;Lee, Ju-Hahn;Kim, Joong-Hyun;Kim, Chan-Hyeong;Lee, Chun-Sik;Lee, Dong-Soo;Lee, Soo-Jin
    • Nuclear Medicine and Molecular Imaging
    • /
    • v.41 no.3
    • /
    • pp.234-240
    • /
    • 2007
  • Purpose: In this study we propose a block-iterative method for reconstructing Compton scattered data. This study shows that the well-known expectation maximization (EM) approach along with its accelerated version based on the ordered subsets principle can be applied to the problem of image reconstruction for Compton camera. This study also compares several methods of constructing subsets for optimal performance of our algorithms. Materials and Methods: Three reconstruction algorithms were implemented; simple backprojection (SBP), EM, and ordered subset EM (OSEM). For OSEM, the projection data were grouped into subsets in a predefined order. Three different schemes for choosing nonoverlapping subsets were considered; scatter angle-based subsets, detector position-based subsets, and both scatter angle- and detector position-based subsets. EM and OSEM with 16 subsets were performed with 64 and 4 iterations, respectively. The performance of each algorithm was evaluated in terms of computation time and normalized mean-squared error. Results: Both EM and OSEM clearly outperformed SBP in all aspects of accuracy. The OSEM with 16 subsets and 4 iterations, which is equivalent to the standard EM with 64 iterations, was approximately 14 times faster in computation time than the standard EM. In OSEM, all of the three schemes for choosing subsets yielded similar results in computation time as well as normalized mean-squared error. Conclusion: Our results show that the OSEM algorithm, which have proven useful in emission tomography, can also be applied to the problem of image reconstruction for Compton camera. With properly chosen subset construction methods and moderate numbers of subsets, our OSEM algorithm significantly improves the computational efficiency while keeping the original quality of the standard EM reconstruction. The OSEM algorithm with scatter angle- and detector position-based subsets is most available.

The Evaluation of Attenuation Difference and SUV According to Arm Position in Whole Body PET/CT (전신 PET/CT 검사에서 팔의 위치에 따른 감약 정도와 SUV 변화 평가)

  • Kwak, In-Suk;Lee, Hyuk;Choi, Sung-Wook;Suk, Jae-Dong
    • The Korean Journal of Nuclear Medicine Technology
    • /
    • v.14 no.2
    • /
    • pp.21-25
    • /
    • 2010
  • Purpose: For better PET imaging with accuracy the transmission scanning is inevitably required for attenuation correction. The attenuation is affected by condition of acquisition and patient position, consequently quantitative accuracy may be decreased in emission scan imaging. In this paper, the present study aims at providing the measurement for attenuation varying with the positions of the patient's arm in whole body PET/CT, further performing the comparative analysis over its SUV changes. Materials and Methods: NEMA 1994 PET phantom was filled with $^{18}F$-FDG and the concentration ratio of insert cylinder and background water fit to 4:1. Phantom images were acquired through emission scanning for 4min after conducting transmission scanning by using CT. In an attempt to acquire image at the state that the arm of the patient was positioned at the lower of ahead, image was acquired in away that two pieces of Teflon inserts were used additionally by fixing phantoms at both sides of phantom. The acquired imaged at a were reconstructed by applying the iterative reconstruction method (iteration: 2, subset: 28) as well as attenuation correction using the CT, and then VOI was drawn on each image plane so as to measure CT number and SUV and comparatively analyze axial uniformity (A.U=Standard deviation/Average SUV) of PET images. Results: It was found from the above phantom test that, when comparing two cases of whether Teflon insert was fixed or removed, the CT number of cylinder increased from -5.76 HU to 0 HU, while SUV decreased from 24.64 to 24.29 and A.U from 0.064 to 0.052. And the CT number of background water was identified to increase from -6.14 HU to -0.43 HU, whereas SUV decreased from 6.3 to 5.6 and A.U also decreased from 0.12 to 0.10. In addition, as for the patient image, CT number was verified to increase from 53.09 HU to 58.31 HU and SUV decreased from 24.96 to 21.81 when the patient's arm was positioned over the head rather than when it was lowered. Conclusion: When arms up protocol was applied, the SUV of phantom and patient image was decreased by 1.4% and 9.2% respectively. With the present study it was concluded that in case of PET/CT scanning against the whole body of a patient the position of patient's arm was not so much significant. Especially, the scanning under the condition that the arm is raised over to the head gives rise to more probability that the patient is likely to move due to long scanning time that causes the increase of uptake of $^{18}F$-FDG of brown fat at the shoulder part together with increased pain imposing to the shoulder and discomfort to a patient. As regarding consideration all of such factors, it could be rationally drawn that PET/CT scanning could be made with the arm of the subject lowered.

  • PDF

Performance Characteristics of 3D GSO PET/CT Scanner (Philips GEMINI PET/DT) (3차원 GSO PET/CT 스캐너(Philips GEMINI PET/CT의 특성 평가)

  • Kim, Jin-Su;Lee, Jae-Sung;Lee, Byeong-Il;Lee, Dong-Soo;Chung, June-Key;Lee, Myung-Chul
    • The Korean Journal of Nuclear Medicine
    • /
    • v.38 no.4
    • /
    • pp.318-324
    • /
    • 2004
  • Purpose: Philips GEMINI is a newly introduced whole-body GSO PET/CT scanner. In this study, performance of the scanner including spatial resolution, sensitivity, scatter fraction, noise equivalent count ratio (NECR) was measured utilizing NEMA NU2-2001 standard protocol and compared with performance of LSO, BGO crystal scanner. Methods: GEMINI is composed of the Philips ALLEGRO PET and MX8000 D multi-slice CT scanners. The PET scanner has 28 detector segments which have an array of 29 by 22 GSO crystals ($4{\times}6{\times}20$ mm), covering axial FOV of 18 cm. PET data to measure spatial resolution, sensitivity, scatter fraction, and NECR were acquired in 3D mode according to the NEMA NU2 protocols (coincidence window: 8 ns, energy window: $409[\sim}664$ keV). For the measurement of spatial resolution, images were reconstructed with FBP using ramp filter and an iterative reconstruction algorithm, 3D RAMLA. Data for sensitivity measurement were acquired using NEMA sensitivity phantom filled with F-18 solution and surrounded by $1{\sim}5$ aluminum sleeves after we confirmed that dead time loss did not exceed 1%. To measure NECR and scatter fraction, 1110 MBq of F-18 solution was injected into a NEMA scatter phantom with a length of 70 cm and dynamic scan with 20-min frame duration was acquired for 7 half-lives. Oblique sinograms were collapsed into transaxial slices using single slice rebinning method, and true to background (scatter+random) ratio for each slice and frame was estimated. Scatter fraction was determined by averaging the true to background ratio of last 3 frames in which the dead time loss was below 1%. Results: Transverse and axial resolutions at 1cm radius were (1) 5.3 and 6.5 mm (FBP), (2) 5.1 and 5.9 mm (3D RAMLA). Transverse radial, transverse tangential, and axial resolution at 10 cm were (1) 5.7, 5.7, and 7.0 mm (FBP), (2) 5.4, 5.4, and 6.4 mm (3D RAMLA). Attenuation free values of sensitivity were 3,620 counts/sec/MBq at the center of transaxial FOV and 4,324 counts/sec/MBq at 10 cm offset from the center. Scatter fraction was 40.6%, and peak true count rate and NECR were 88.9 kcps @ 12.9 kBq/mL and 34.3 kcps @ 8.84 kBq/mL. These characteristics are better than that of ECAT EXACT PET scanner with BGO crystal. Conclusion: The results of this field test demonstrate high resolution, sensitivity and count rate performance of the 3D PET/CT scanner with GSO crystal. The data provided here will be useful for the comparative study with other 3D PET/CT scanners using BGO or LSO crystals.