• Title/Summary/Keyword: visual model

Search Result 2,021, Processing Time 0.031 seconds

A Model-Based Image Steganography Method Using Watson's Visual Model

  • Fakhredanesh, Mohammad;Safabakhsh, Reza;Rahmati, Mohammad
    • ETRI Journal
    • /
    • v.36 no.3
    • /
    • pp.479-489
    • /
    • 2014
  • This paper presents a model-based image steganography method based on Watson's visual model. Model-based steganography assumes a model for cover image statistics. This approach, however, has some weaknesses, including perceptual detectability. We propose to use Watson's visual model to improve perceptual undetectability of model-based steganography. The proposed method prevents visually perceptible changes during embedding. First, the maximum acceptable change in each discrete cosine transform coefficient is extracted based on Watson's visual model. Then, a model is fitted to a low-precision histogram of such coefficients and the message bits are encoded to this model. Finally, the encoded message bits are embedded in those coefficients whose maximum possible changes are visually imperceptible. Experimental results show that changes resulting from the proposed method are perceptually undetectable, whereas model-based steganography retains perceptually detectable changes. This perceptual undetectability is achieved while the perceptual quality - based on the structural similarity measure - and the security - based on two steganalysis methods - do not show any significant changes.

Predicting Human Performance of Multiple-Target Search Using a Visual Lobe (비쥬얼 롭을 사용한 다수표적 탐색의 수행도 예측)

  • Hong, Seung-Kweon
    • Journal of the Ergonomics Society of Korea
    • /
    • v.28 no.3
    • /
    • pp.55-62
    • /
    • 2009
  • This study is concerned with predicting human search performance using a visual lobe. The most previous studies on human performance in visual search have been limited to a single-target search. This study extended the visual search research to multiple-target search including targets of different types as well as targets of same types. A model for predicting visual search performance was proposed and the model was validated by human search data. Additionally, this study found that human subjects always did not use a constant ratio of the whole visual lobe size for each type of targets in visual search process. The more conspicuous the target is, the more ratio of the whole visual lobe size human subjects use. The model that can predict human performance in multiple-target search may facilitate visual inspection plan in manufacturing.

A study on the model of measuring visual accessibility to urban green spaces (도시녹지의 시각적 접근성 측정모델에 관한 연구)

  • 임승빈;허윤정
    • Journal of the Korean Institute of Landscape Architecture
    • /
    • v.23 no.3
    • /
    • pp.1-14
    • /
    • 1995
  • The aspect of visual accessibility to urban green spaces is an important factor because it contributes making pleasant environment by increasing the visual experience of nature in urban environment. But we have tried neither to consider nor to measure it. Since he concept of visual accessibility has not formally defined yet, it was operationally defined in this study. And then the model of measuring visual accessibility was suggested and verified through the case study on neighborhood parks in Seoul. The findings are as follows : 1) The concept of visual accessibility is defined as opportunity and potentiality to observe green spaces. 2) The model of measuring visual accessibility deals with not only adjacent area but also viewshed area. In adjacent area, considering factors are the area of road adjacent to green spaces and the area of exposed green spaces. In viewshed area, considering factors are the area of road located in viewshed area, the area of exposed green spaces, and the weight according to observing distance. 3) The final model of measuring visual accessibility suggested in this study is as follows.

  • PDF

Bag of Visual Words Method based on PLSA and Chi-Square Model for Object Category

  • Zhao, Yongwei;Peng, Tianqiang;Li, Bicheng;Ke, Shengcai
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.9 no.7
    • /
    • pp.2633-2648
    • /
    • 2015
  • The problem of visual words' synonymy and ambiguity always exist in the conventional bag of visual words (BoVW) model based object category methods. Besides, the noisy visual words, so-called "visual stop-words" will degrade the semantic resolution of visual dictionary. In view of this, a novel bag of visual words method based on PLSA and chi-square model for object category is proposed. Firstly, Probabilistic Latent Semantic Analysis (PLSA) is used to analyze the semantic co-occurrence probability of visual words, infer the latent semantic topics in images, and get the latent topic distributions induced by the words. Secondly, the KL divergence is adopt to measure the semantic distance between visual words, which can get semantically related homoionym. Then, adaptive soft-assignment strategy is combined to realize the soft mapping between SIFT features and some homoionym. Finally, the chi-square model is introduced to eliminate the "visual stop-words" and reconstruct the visual vocabulary histograms. Moreover, SVM (Support Vector Machine) is applied to accomplish object classification. Experimental results indicated that the synonymy and ambiguity problems of visual words can be overcome effectively. The distinguish ability of visual semantic resolution as well as the object classification performance are substantially boosted compared with the traditional methods.

Visual SLAM using Local Bundle Optimization in Unstructured Seafloor Environment (국소 집단 최적화 기법을 적용한 비정형 해저면 환경에서의 비주얼 SLAM)

  • Hong, Seonghun;Kim, Jinwhan
    • The Journal of Korea Robotics Society
    • /
    • v.9 no.4
    • /
    • pp.197-205
    • /
    • 2014
  • As computer vision algorithms are developed on a continuous basis, the visual information from vision sensors has been widely used in the context of simultaneous localization and mapping (SLAM), called visual SLAM, which utilizes relative motion information between images. This research addresses a visual SLAM framework for online localization and mapping in an unstructured seabed environment that can be applied to a low-cost unmanned underwater vehicle equipped with a single monocular camera as a major measurement sensor. Typically, an image motion model with a predefined dimensionality can be corrupted by errors due to the violation of the model assumptions, which may lead to performance degradation of the visual SLAM estimation. To deal with the erroneous image motion model, this study employs a local bundle optimization (LBO) scheme when a closed loop is detected. The results of comparison between visual SLAM estimation with LBO and the other case are presented to validate the effectiveness of the proposed methodology.

Modeling the Visual Target Search in Natural Scenes

  • Park, Daecheol;Myung, Rohae;Kim, Sang-Hyeob;Jang, Eun-Hye;Park, Byoung-Jun
    • Journal of the Ergonomics Society of Korea
    • /
    • v.31 no.6
    • /
    • pp.705-713
    • /
    • 2012
  • Objective: The aim of this study is to predict human visual target search using ACT-R cognitive architecture in real scene images. Background: Human uses both the method of bottom-up and top-down process at the same time using characteristics of image itself and knowledge about images. Modeling of human visual search also needs to include both processes. Method: In this study, visual target object search performance in real scene images was analyzed comparing experimental data and result of ACT-R model. 10 students participated in this experiment and the model was simulated ten times. This experiment was conducted in two conditions, indoor images and outdoor images. The ACT-R model considering the first saccade region through calculating the saliency map and spatial layout was established. Proposed model in this study used the guide of visual search and adopted visual search strategies according to the guide. Results: In the analysis results, no significant difference on performance time between model prediction and empirical data was found. Conclusion: The proposed ACT-R model is able to predict the human visual search process in real scene images using salience map and spatial layout. Application: This study is useful in conducting model-based evaluation in visual search, particularly in real images. Also, this study is able to adopt in diverse image processing program such as helper of the visually impaired.

A Bio-Inspired Modeling of Visual Information Processing for Action Recognition (생체 기반 시각정보처리 동작인식 모델링)

  • Kim, JinOk
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.3 no.8
    • /
    • pp.299-308
    • /
    • 2014
  • Various literatures related computing of information processing have been recently shown the researches inspired from the remarkably excellent human capabilities which recognize and categorize very complex visual patterns such as body motions and facial expressions. Applied from human's outstanding ability of perception, the classification function of visual sequences without context information is specially crucial task for computer vision to understand both the coding and the retrieval of spatio-temporal patterns. This paper presents a biological process based action recognition model of computer vision, which is inspired from visual information processing of human brain for action recognition of visual sequences. Proposed model employs the structure of neural fields of bio-inspired visual perception on detecting motion sequences and discriminating visual patterns in human brain. Experimental results show that proposed recognition model takes not only into account several biological properties of visual information processing, but also is tolerant of time-warping. Furthermore, the model allows robust temporal evolution of classification compared to researches of action recognition. Presented model contributes to implement bio-inspired visual processing system such as intelligent robot agent, etc.

A Study on the influence of Visual Mental model in human to percept product form (제품형태 지각에 있어서 시각적 멘탈모델의 영향에 관한 연구)

  • 오해춘
    • Archives of design research
    • /
    • v.15 no.1
    • /
    • pp.407-414
    • /
    • 2002
  • Human does information processing more efficiently using mental model at his understanding process. if We know what mental model is, we will know how to percept that. Similarly is it used visual mental model to percept visual object\ulcorner If it is true, we will analysis how man to understand. In this research, we have one experimentation to timid out that is true. So we did experimentation to know how man to understand new products by visual mental model. So we use 2700cc car's side view as stimuli. To A Subjects, we give them to see re\ulcorner size view before 120% ratio. To B Subject, we give them to see 120% ratio. After a experimentation, It is true that we make hypothesis that A Subjects percept this stimuli too long then B Subjects. Accordingly it proves human to use visual mental model to percept visual object. The result of experimentation, a testee use preliminary stimulus as basis of evaluation. It means that he use mental model as basis of evaluation of main stimulus. Accordingly development of new design can afford more strategic approach. Industrial designer who want to strategic approach can forced on high-weighted form attribute as consumer gives.

  • PDF

Visual Search Model based on Saliency and Scene-Context in Real-World Images (실제 이미지에서 현저성과 맥락 정보의 영향을 고려한 시각 탐색 모델)

  • Choi, Yoonhyung;Oh, Hyungseok;Myung, Rohae
    • Journal of Korean Institute of Industrial Engineers
    • /
    • v.41 no.4
    • /
    • pp.389-395
    • /
    • 2015
  • According to much research on cognitive science, the impact of the scene-context on human visual search in real-world images could be as important as the saliency. Therefore, this study proposed a method of Adaptive Control of Thought-Rational (ACT-R) modeling of visual search in real-world images, based on saliency and scene-context. The modeling method was developed by using the utility system of ACT-R to describe influences of saliency and scene-context in real-world images. Then, the validation of the model was performed, by comparing the data of the model and eye-tracking data from experiments in simple task in which subjects search some targets in indoor bedroom images. Results show that model data was quite well fit with eye-tracking data. In conclusion, the method of modeling human visual search proposed in this study should be used, in order to provide an accurate model of human performance in visual search tasks in real-world images.

Object Classification based on Weakly Supervised E2LSH and Saliency map Weighting

  • Zhao, Yongwei;Li, Bicheng;Liu, Xin;Ke, Shengcai
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.10 no.1
    • /
    • pp.364-380
    • /
    • 2016
  • The most popular approach in object classification is based on the bag of visual-words model, which has several fundamental problems that restricting the performance of this method, such as low time efficiency, the synonym and polysemy of visual words, and the lack of spatial information between visual words. In view of this, an object classification based on weakly supervised E2LSH and saliency map weighting is proposed. Firstly, E2LSH (Exact Euclidean Locality Sensitive Hashing) is employed to generate a group of weakly randomized visual dictionary by clustering SIFT features of the training dataset, and the selecting process of hash functions is effectively supervised inspired by the random forest ideas to reduce the randomcity of E2LSH. Secondly, graph-based visual saliency (GBVS) algorithm is applied to detect the saliency map of different images and weight the visual words according to the saliency prior. Finally, saliency map weighted visual language model is carried out to accomplish object classification. Experimental results datasets of Pascal 2007 and Caltech-256 indicate that the distinguishability of objects is effectively improved and our method is superior to the state-of-the-art object classification methods.