• Title/Summary/Keyword: Image Normalization

Search Result 245, Processing Time 0.036 seconds

Deep Learning Based Digital Staining Method in Fourier Ptychographic Microscopy Image (Fourier Ptychographic Microscopy 영상에서의 딥러닝 기반 디지털 염색 방법 연구)

  • Seok-Min Hwang;Dong-Bum Kim;Yu-Jeong Kim;Yeo-Rin Kim;Jong-Ha Lee
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.23 no.2
    • /
    • pp.97-106
    • /
    • 2022
  • In this study, H&E staining is necessary to distinguish cells. However, dyeing directly requires a lot of money and time. The purpose is to convert the phase image of unstained cells to the amplitude image of stained cells. Image data taken with FPM was created with Phase image and Amplitude image using Matlab's parameters. Through normalization, a visually identifiable image was obtained. Through normalization, a visually distinguishable image was obtained. Using the GAN algorithm, a Fake Amplitude image similar to the Real Amplitude image was created based on the Phase image, and cells were distinguished by objectification using MASK R-CNN with the Fake Amplitude image As a result of the study, D loss max is 3.3e-1, min is 6.8e-2, G loss max is 6.9e-2, min is 2.9e-2, A loss max is 5.8e-1, min is 1.2e-1, Mask R-CNN max is 1.9e0, and min is 3.2e-1.

A Study on Utilization of Vision Transformer for CTR Prediction (CTR 예측을 위한 비전 트랜스포머 활용에 관한 연구)

  • Kim, Tae-Suk;Kim, Seokhun;Im, Kwang Hyuk
    • Knowledge Management Research
    • /
    • v.22 no.4
    • /
    • pp.27-40
    • /
    • 2021
  • Click-Through Rate (CTR) prediction is a key function that determines the ranking of candidate items in the recommendation system and recommends high-ranking items to reduce customer information overload and achieve profit maximization through sales promotion. The fields of natural language processing and image classification are achieving remarkable growth through the use of deep neural networks. Recently, a transformer model based on an attention mechanism, differentiated from the mainstream models in the fields of natural language processing and image classification, has been proposed to achieve state-of-the-art in this field. In this study, we present a method for improving the performance of a transformer model for CTR prediction. In order to analyze the effect of discrete and categorical CTR data characteristics different from natural language and image data on performance, experiments on embedding regularization and transformer normalization are performed. According to the experimental results, it was confirmed that the prediction performance of the transformer was significantly improved when the L2 generalization was applied in the embedding process for CTR data input processing and when batch normalization was applied instead of layer normalization, which is the default regularization method, to the transformer model.

Adaptive Optimal Thresholding for the Segmentation of Individual Tooth from CT Images (CT영상에서 개별 치아 분리를 위한 적응 최적 임계화 방안)

  • Heo, Hoon;Chae, Ok-Sam
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.41 no.3
    • /
    • pp.163-174
    • /
    • 2004
  • The 3D tooth model in which each tooth can be manipulated individualy is essential component for the orthodontic simulation and implant simulation in dental field. For the reconstruction of such a tooth model, we need an image segmentation algorithm capable of separating individual tooth from neighboring teeth and alveolar bone. In this paper we propose a CT image normalization method and adaptive optimal thresholding algorithm for the segmenation of tooth region in CT image slices. The proposed segmentation algorithm is based on the fact that the shape and intensity of tooth change gradually among CT image slices. It generates temporary boundary of a tooth by using the threshold value estimated in the previous imge slice, and compute histograms for the inner region and the outer region seperated by the temporary boundary. The optimal threshold value generating the finnal tooth region is computed based on these two histogram.

Image Enhancement for Western Epigraphy Using Local Statistics (국부 통계치를 활용한 서양금석문 영상향상)

  • Hwang, Jae-Ho
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.44 no.3
    • /
    • pp.80-87
    • /
    • 2007
  • In this paper, we investigate an enhancement method for Western epigraphic images, which is based on local statistics. Image data is partitioned into two regions, background and information. Statistical and functional analyses are proceeded for image modeling. The Western epigraphic images, for the most part, have shown the Gaussian distribution. It is clarified that each region can be differentiated statistically. The local normalization process algorithm is designed on this model. The parameter is extracted and it‘s properties are verified with the size of moving window. The spatial gray-level distribution is modified and regions are differentiated by adjusting parameter and the size of moving window. Local statistics are utilized for realization of the enhancement, so that difference between regions can be enhanced and noise or speckles of region can be smoothed. Experimental results are presented to show the superiority of the proposed algorithm over the conventional methods.

Enhanced Vein Detection Method by Using Image Scaler Based on Poly Phase Filter (Poly Phase Filter 기반의 영상 스케일러를 이용한 개선 된 정맥 영역 추출 방법)

  • Kim, HeeKyung;Lee, Seungmin;Kang, Bongsoon
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.22 no.5
    • /
    • pp.734-739
    • /
    • 2018
  • Fingerprint recognition and iris recognition, which are one of the biometric methods, are easily influenced by external factors such as sunlight. Recently, finger vein recognition is used as a method utilizing internal features. However, for accurate finger vein recognition, it is important to clearly separate vein and background regions. However, it is difficult to separate the vein region and background region due to the abnormalized illumination, and a method of separating the vein region and the background region after normalized the illumination of the input image has been proposed. In this paper, we proposed a method to enhance the quality improvement and improve the processing time compared to the existing finger vein recognition system binarization and labeling method of the image including the image stretching process based on the existing illumination normalization method.

A Rotation Resistant Logo Embedding Watermark on Frequency Domain (회전 변환에 강인한 주파수 영역 로고 삽입 워터마크 방법)

  • Lee, In-Jung;Lee, Hyoung;Min, Joon-Young
    • Proceedings of the Korea Society of Information Technology Applications Conference
    • /
    • 2006.06a
    • /
    • pp.730-736
    • /
    • 2006
  • In this paper, we propose a rotation resistant robust logo embedding watermarking technique. Geometric manipulations make the detection process very complex and difficult. Watermark embedding ill the normalized image directly suffers from smoothing effect due to the interpolation during the image normalization. This can be avoided by estimating the transform parameters using image normalization technique, instead of embedding in the normalized image. Conventional rotation resistant schemes that use full frame transform. In this paper, we adopt $8{\times}8$ block DCT and calculate masking using a spatio-frequency localization of the $8{\times}8$ block DCT coefficients. Experimental results show that the proposed algorithm is robust against rotation process.

  • PDF

Image Similarity Retrieval using an Scale and Rotation Invariant Region Feature (크기 및 회전 불변 영역 특징을 이용한 이미지 유사성 검색)

  • Yu, Seung-Hoon;Kim, Hyun-Soo;Lee, Seok-Lyong;Lim, Myung-Kwan;Kim, Deok-Hwan
    • Journal of KIISE:Databases
    • /
    • v.36 no.6
    • /
    • pp.446-454
    • /
    • 2009
  • Among various region detector and shape feature extraction method, MSER(Maximally Stable Extremal Region) and SIFT and its variant methods are popularly used in computer vision application. However, since SIFT is sensitive to the illumination change and MSER is sensitive to the scale change, it is not easy to apply the image similarity retrieval. In this paper, we present a Scale and Rotation Invariant Region Feature(SRIRF) descriptor using scale pyramid, MSER and affine normalization. The proposed SRIRF method is robust to scale, rotation, illumination change of image since it uses the affine normalization and the scale pyramid. We have tested the SRIRF method on various images. Experimental results demonstrate that the retrieval performance of the SRIRF method is about 20%, 38%, 11%, 24% better than those of traditional SIFT, PCA-SIFT, CE-SIFT and SURF, respectively.

Block based Normalized Numeric Image Descriptor (블록기반 정규화 된 이미지 수 표현자)

  • Park, Yu-Yung;Cho, Sang-Bock;Lee, Jong-Hwa
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.49 no.2
    • /
    • pp.61-68
    • /
    • 2012
  • This paper describes a normalized numeric image descriptor used to assess the luminance and contrast of the image. The proposed image descriptor used the each pixel data as weighted value of the probability density function (PDF) and defined by normalization in order to objective represent. The proposed image numeric descriptor can be used to the adaptive gamma process because it suggests the objective basis of the gamma value selection.

Semantic Segmentation of Drone Imagery Using Deep Learning for Seagrass Habitat Monitoring (잘피 서식지 모니터링을 위한 딥러닝 기반의 드론 영상 의미론적 분할)

  • Jeon, Eui-Ik;Kim, Seong-Hak;Kim, Byoung-Sub;Park, Kyung-Hyun;Choi, Ock-In
    • Korean Journal of Remote Sensing
    • /
    • v.36 no.2_1
    • /
    • pp.199-215
    • /
    • 2020
  • A seagrass that is marine vascular plants plays an important role in the marine ecosystem, so periodic monitoring ofseagrass habitatsis being performed. Recently, the use of dronesthat can easily acquire very high-resolution imagery is increasing to efficiently monitor seagrass habitats. And deep learning based on a convolutional neural network has shown excellent performance in semantic segmentation. So, studies applied to deep learning models have been actively conducted in remote sensing. However, the segmentation accuracy was different due to the hyperparameter, various deep learning models and imagery. And the normalization of the image and the tile and batch size are also not standardized. So,seagrass habitats were segmented from drone-borne imagery using a deep learning that shows excellent performance in this study. And it compared and analyzed the results focused on normalization and tile size. For comparison of the results according to the normalization, tile and batch size, a grayscale image and grayscale imagery converted to Z-score and Min-Max normalization methods were used. And the tile size isincreased at a specific interval while the batch size is allowed the memory size to be used as much as possible. As a result, IoU was 0.26 ~ 0.4 higher than that of Z-score normalized imagery than other imagery. Also, it wasfound that the difference to 0.09 depending on the tile and batch size. The results were different according to the normalization, tile and batch. Therefore, this experiment found that these factors should have a suitable decision process.

Truncation Artifact Reduction Using Weighted Normalization Method in Prototype R/F Chest Digital Tomosynthesis (CDT) System (프로토타입 R/F 흉부 디지털 단층영상합성장치 시스템에서 잘림 아티팩트 감소를 위한 가중 정규화 접근법에 대한 연구)

  • Son, Junyoung;Choi, Sunghoon;Lee, Donghoon;Kim, Hee-Joung
    • Journal of the Korean Society of Radiology
    • /
    • v.13 no.1
    • /
    • pp.111-118
    • /
    • 2019
  • Chest digital tomosynthesis has become a practical imaging modality because it can solve the problem of anatomy overlapping in conventional chest radiography. However, because of both limited scan angle and finite-size detector, a portion of chest cannot be represented in some or all of the projection. These bring a discontinuity in intensity across the field of view boundaries in the reconstructed slices, which we refer to as the truncation artifacts. The purpose of this study was to reduce truncation artifacts using a weighted normalization approach and to investigate the performance of this approach for our prototype chest digital tomosynthesis system. The system source-to-image distance was 1100 mm, and the center of rotation of X-ray source was located on 100 mm above the detector surface. After obtaining 41 projection views with ${\pm}20^{\circ}$ degrees, tomosynthesis slices were reconstructed with the filtered back projection algorithm. For quantitative evaluation, peak signal to noise ratio and structure similarity index values were evaluated after reconstructing reference image using simulation, and mean value of specific direction values was evaluated using real data. Simulation results showed that the peak signal to noise ratio and structure similarity index was improved respectively. In the case of the experimental results showed that the effect of artifact in the mean value of specific direction of the reconstructed image was reduced. In conclusion, the weighted normalization method improves the quality of image by reducing truncation artifacts. These results suggested that weighted normalization method could improve the image quality of chest digital tomosynthesis.