Iron objects become corroded at fast speed from the moment when they are excavated, so it is needed to control corrosion through processes of conservation treatment. However, re-corrosion mostly takes place in excavate iron objects, although they have already gone through the process of conservation treatment, and it is more difficult to carry out the second conservation treatment of re-corroded excavated iron objects than the first conservation treatment, and it requires a longer period of time to treat them as well. In this study, aims to discover factors of re-corrosion by scientifically analyzing corrosion products generated during the process of storage after the process of conservation treatment. The finished on conservation treatment of the iron artifacts, which were unearthed from three ancient site in Gyeongju by using the same conservation method between 2002 and 2009, re-corrosion condition observed on the packaging-iron artifacts. Focused on 9 target forged iron artifacts among them, this study analyzed the physical changes by mass measurement, naked-eye and microscopic observations and the chemical changes by SEM-EDS, XRD, IC and ICP analysis. The results show that the yellowish brown corrosion products formed on the facing surface of part dropped from the artifacts had different associated forms but acicular shape. In addition, the acicular shape became clearer as the color changed from red to yellowish brown. According to the process when the conservation treatment was completed, the mass of the artifacts increased in proportion to the corrosion products and the chloride ion ($Cl^-$) concentration had a tendency to increase relatively. ${\beta}$-FeOOH (akaganeite) was confirmed in the XRD analysis for the corrosion products of all the collected samples. As a result of ICP analysis, $Na^+$ and $Ca^{2+}$ components were confirmed.
Journal of Korean Society for Geospatial Information Science
/
v.24
no.4
/
pp.59-66
/
2016
For the performance experiments of databases systems with moving object databases, we need moving object trajectory data sets. For example, benchmark data sets of moving object trajectories are required for experiments on query processing of moving object databases. For those reasons, several tools have been developed for generating moving objects in Euclidean spaces or road network spaces. Indoor space differs from outdoor spaces in many aspects and moving object generator for indoor space should reflect these differences. Even some tools were developed to produce virtual moving object trajectories in indoor space, the movements generated by them are not realistic. In this paper, we present a moving object generation tool for indoor space. First, this tool generates trajectories for pedestrians in an indoor space. And it provides a parametric generation of trajectories considering not only speed, number of pedestrians, minimum distance between pedestrians but also type of spaces, time constraints, and type of pedestrians. We try to reflect the patterns of pedestrians in indoor space as realistic as possible. For the reason of interoperability, several geospatial standards are used in the development of the tool.
Due to recent increase in applications requiring huge amount of data such as spatial data analysis and image analysis, clustering on large databases has been actively studied. In a hierarchical clustering method, a tree representing hierarchical decomposition of the database is first created, and then, used for efficient clustering. Existing hierarchical clustering methods mainly adopted the bottom-up approach, which creates a tree from the bottom to the topmost level of the hierarchy. These bottom-up methods require at least one scan over the entire database in order to build the tree and need to search most nodes of the tree since the clustering algorithm starts from the leaf level. In this paper, we propose a novel top-down hierarchical clustering method that uses multidimensional indexes that are already maintained in most database applications. Generally, multidimensional indexes have the clustering property storing similar objects in the same (or adjacent) data pares. Using this property we can find adjacent objects without calculating distances among them. We first formally define the cluster based on the density of objects. For the definition, we propose the concept of the region contrast partition based on the density of the region. To speed up the clustering algorithm, we use the branch-and-bound algorithm. We propose the bounds and formally prove their correctness. Experimental results show that the proposed method is at least as effective in quality of clustering as BIRCH, a bottom-up hierarchical clustering method, while reducing the number of page accesses by up to 26~187 times depending on the size of the database. As a result, we believe that the proposed method significantly improves the clustering performance in large databases and is practically usable in various database applications.
This study is based on Freudenthal's mathmatising process and the didactical phenomenology of linear function concept, I have described and examined the process in which students represent the constant rate of change into tables, graphs and equations and, in this way, how they construct mental objects and essence of the linear function concept. The students used the proportionality as composite units, when they represented the phenomenon with constant rate of change into tables. When representing in graphs, all but one student represented it into a line. There were differences among the students in the level they were using the given conditions, co-variation perspective, and corresponding rules when formulating equations. The students compared the relationship between two variables in a multiplicative way, and under the guidance of teachers they reached to the understanding that its relationship becomes a constant. Moreover, they could construct mental objects of a constant rate of change, understanding the situation where the relationship between time difference and distance difference becomes one value, namely speed. The students had difficulties in connecting the rate of change with the inclination of a line. The students constructed the essence (concept) of linear functions, after building and organizing the image that the rate of change is constant, the graph is linear, and the equation is formulated as y=ax+b (a: inclination, b: intercept).
Journal of the Korea Institute of Information and Communication Engineering
/
v.24
no.10
/
pp.1331-1340
/
2020
Object detection in maritime means that the captain detects floating objects that has a risk of colliding with the ship using the computer automatically and as accurately as human eyes. In conventional ships, the presence and distance of objects are determined through radar waves. However, it cannot identify the shape and type. In contrast, with the development of AI, cameras help accurately identify obstacles on the sea route with excellent performance in detecting or recognizing objects. The computer must calculate high-volume pixels to analyze digital images. However, the CPU is specialized for sequential processing; the processing speed is very slow, and smooth service support or security is not guaranteed. Accordingly, this study developed maritime object detection software and implemented it with FPGA to accelerate the processing of large-scale computations. Additionally, the system implementation was improved through embedded boards and FPGA interface, achieving 30 times faster performance than the existing algorithm and a three-times faster entire system.
The Journal of Korean Institute of Communications and Information Sciences
/
v.36
no.7C
/
pp.421-427
/
2011
The importance of the real-time object recognition and tracking field has been growing steadily due to rapid advancement in the computer vision applications industry. As is well known, the mean-shift algorithm is widely used in robust real-time object tracking systems. Since the mentioned algorithm is easy to implement and efficient in object tracking computation, many say it is suitable to be applied to real-time object tracking systems. However, one of the major drawbacks of this algorithm is that it always converges to a local mode, failing to perform well in a cluttered environment. In this paper, an Optical Flow-based algorithm which fits for real-time recognition of multiple moving objects is proposed. Also in the tests, the newly proposed method contributed to raising the similarity of multiple moving objects, the similarity was as high as 0.96, up 13.4% over that of the mean-shift algorithm. Meanwhile, the level of pixel errors from using the new method keenly decreased by more than 50% over that from applying the mean-shift algorithm. If the data processing speed in the video surveillance systems can be reduced further, owing to improved algorithms for faster moving object recognition and tracking functions, we will be able to expect much more efficient intelligent systems in this industrial arena.
Through photogrammetry techniques, what current researches can achieve at present is rough 3D mesh and color map of objects, rather than usable photo-realistic 3D assets. This research aims to propose a new method to create photo-realistic 3D assets that can be used in the field of visualization applications. The new method combines photogrammetry with computer graphics modeling. Through the description of the production process of three objects in the real world - "Bullet Box", "Gun" and "Metal Beverage Bottle," it introduces in details the concept, functions, operating skills and software packages used in the steps including the photograph object, white balance, reconstruction, cleanup reconstruction, retopology, UV unwrapping, projection, texture baking, De-Lighting and Create Material Maps. In order to increase the flexibility of the method, alternatives to the software packages are also recommended for each step. In this research, 3D assets are produced that are accurate in shape, correct in color, easy to render and can be physically interacted with dynamic lighting in texture. The new method can obtain more realistic visual effects at a faster speed. It does not require large-scale teams, expensive equipment and software packages, therefore it is suitable for small studios and independent artists and educational institutions.
Al-Sharari, Waad;Mahmood, Mahmood A.;Abd El-Aziz, A.A.;Azim, Nesrine A.
International Journal of Computer Science & Network Security
/
v.22
no.6
/
pp.131-138
/
2022
Novel Coronavirus (COVID-19) is viewed as one of the main general wellbeing theaters on the worldwide level all over the planet. Because of the abrupt idea of the flare-up and the irresistible force of the infection, it causes individuals tension, melancholy, and other pressure responses. The avoidance and control of the novel Covid pneumonia have moved into an imperative stage. It is fundamental to early foresee and figure of infection episode during this troublesome opportunity to control of its grimness and mortality. The entire world is investing unimaginable amounts of energy to fight against the spread of this lethal infection. In this paper, we utilized machine learning and deep learning techniques for analyzing what is going on utilizing countries shared information and for detecting the climate factors that effect on spreading Covid-19, such as humidity, sunny hours, temperature and wind speed for understanding its regular dramatic way of behaving alongside the forecast of future reachability of the COVID-2019 around the world. We utilized data collected and produced by Kaggle and the Johns Hopkins Center for Systems Science. The dataset has 25 attributes and 9566 objects. Our Experiment consists of two phases. In phase one, we preprocessed dataset for DL model and features were decreased to four features humidity, sunny hours, temperature and wind speed by utilized the Pearson Correlation Coefficient technique (correlation attributes feature selection). In phase two, we utilized the traditional famous six machine learning techniques for numerical datasets, and Dense Net deep learning model to predict and detect the climatic factor that aide to disease outbreak. We validated the model by using confusion matrix (CM) and measured the performance by four different metrics: accuracy, f-measure, recall, and precision.
Background: Stroke is one of the causes affecting gait and balance. Taping is considered an effective method for improving balance and gait in stroke patients. Numerous studies have confirmed the functional effects of taping in stroke patients. However, there is still no consensus regarding the use of taping to improve gait and balance. Objects: The purpose of this review was to investigate the effects of taping on the balance and gait of patients with stroke through meta-analysis of studies. Methods: PubMed, Medline, Embase, Web of Science, Cochrane Review, RISS, DBPia, and Science on were used to collect articles on Kinesio and non-elastic taping. The key terms were "Stroke", "Hemiplegia", "Taping", "Tape", "Balance", and "Gait" with cut-off of October, 2022. Taping group was compared with control groups with sham, placebo, and no taping. The outcome measures included the Berg Balance Scale (BBS), Timed Up and Go (TUG) test, and gait speed (cm/s). Eighteen studies (524 patients) were selected for the meta-analysis. Results: Overall, taping improved balance and gait in stroke patients, and Kinesio and non-elastic taping had similar effect sizes. Taping improved the BBS and TUG, and was most effective on gait speed. Contrary to the expectation that a longer duration of taping would be more affective, taping was most effective when the total taping duration was shorter than 500 minutes. In addition, the effect size of taping was greater when it was simultaneously attached to multiple locations. Conclusion: This meta-analysis supports the use of taping to improve gait and balance in stroke patients, and provides guidelines for the location, duration, and type of tape to increase taping efficiency.
In the traditional buffeting response analysis method, the spanwise incomplete correlation of buffeting forces is always assumed to be same as that of the incident wind turbulence and the action of the signature turbulence is ignored. In this paper, three typical bridge decks usually adopted in the real bridge engineering, a single flat box deck, a central slotted box deck and a two-separated paralleled box deck, were employed as the investigated objects. The wind induced pressure on these bridge decks were measured via a series of wind tunnel pressure tests of the sectional models. The influences of the wind speed in the tests, the angle of attack, the turbulence intensity and the characteristic distance were taken into account and discussed. The spanwise root coherence of buffeting forces was also compared with that of the incidence turbulence. The signature turbulence effect on the spanwise root coherence function was decomposed and explained by a new empirical method with a double-variable model. Finally, the formula of a sum of rational fractions that accounted for the signature turbulence effect was proposed in order to fit the results of the spanwise root coherence function. The results show that, the spanwise root coherence of the drag force agrees with that of incidence turbulence in some range of the reduced frequency but disagree in the mostly reduced frequency. The spanwise root coherence of the lift force and the torsional moment is much larger than that of the incidence turbulence. The influences of the wind speed and the angle of attack are slight, and they can be ignored in the wind tunnel test. The spanwise coherence function often involves several narrow peaks due to the signature turbulence effect in the high reduced frequency zone. The spanwise coherence function is related to the spanwise separation distance and the spanwise integral length scales, and the signature turbulence effect is related to the deck-width-related reduced frequency.
본 웹사이트에 게시된 이메일 주소가 전자우편 수집 프로그램이나
그 밖의 기술적 장치를 이용하여 무단으로 수집되는 것을 거부하며,
이를 위반시 정보통신망법에 의해 형사 처벌됨을 유념하시기 바랍니다.
[게시일 2004년 10월 1일]
이용약관
제 1 장 총칙
제 1 조 (목적)
이 이용약관은 KoreaScience 홈페이지(이하 “당 사이트”)에서 제공하는 인터넷 서비스(이하 '서비스')의 가입조건 및 이용에 관한 제반 사항과 기타 필요한 사항을 구체적으로 규정함을 목적으로 합니다.
제 2 조 (용어의 정의)
① "이용자"라 함은 당 사이트에 접속하여 이 약관에 따라 당 사이트가 제공하는 서비스를 받는 회원 및 비회원을
말합니다.
② "회원"이라 함은 서비스를 이용하기 위하여 당 사이트에 개인정보를 제공하여 아이디(ID)와 비밀번호를 부여
받은 자를 말합니다.
③ "회원 아이디(ID)"라 함은 회원의 식별 및 서비스 이용을 위하여 자신이 선정한 문자 및 숫자의 조합을
말합니다.
④ "비밀번호(패스워드)"라 함은 회원이 자신의 비밀보호를 위하여 선정한 문자 및 숫자의 조합을 말합니다.
제 3 조 (이용약관의 효력 및 변경)
① 이 약관은 당 사이트에 게시하거나 기타의 방법으로 회원에게 공지함으로써 효력이 발생합니다.
② 당 사이트는 이 약관을 개정할 경우에 적용일자 및 개정사유를 명시하여 현행 약관과 함께 당 사이트의
초기화면에 그 적용일자 7일 이전부터 적용일자 전일까지 공지합니다. 다만, 회원에게 불리하게 약관내용을
변경하는 경우에는 최소한 30일 이상의 사전 유예기간을 두고 공지합니다. 이 경우 당 사이트는 개정 전
내용과 개정 후 내용을 명확하게 비교하여 이용자가 알기 쉽도록 표시합니다.
제 4 조(약관 외 준칙)
① 이 약관은 당 사이트가 제공하는 서비스에 관한 이용안내와 함께 적용됩니다.
② 이 약관에 명시되지 아니한 사항은 관계법령의 규정이 적용됩니다.
제 2 장 이용계약의 체결
제 5 조 (이용계약의 성립 등)
① 이용계약은 이용고객이 당 사이트가 정한 약관에 「동의합니다」를 선택하고, 당 사이트가 정한
온라인신청양식을 작성하여 서비스 이용을 신청한 후, 당 사이트가 이를 승낙함으로써 성립합니다.
② 제1항의 승낙은 당 사이트가 제공하는 과학기술정보검색, 맞춤정보, 서지정보 등 다른 서비스의 이용승낙을
포함합니다.
제 6 조 (회원가입)
서비스를 이용하고자 하는 고객은 당 사이트에서 정한 회원가입양식에 개인정보를 기재하여 가입을 하여야 합니다.
제 7 조 (개인정보의 보호 및 사용)
당 사이트는 관계법령이 정하는 바에 따라 회원 등록정보를 포함한 회원의 개인정보를 보호하기 위해 노력합니다. 회원 개인정보의 보호 및 사용에 대해서는 관련법령 및 당 사이트의 개인정보 보호정책이 적용됩니다.
제 8 조 (이용 신청의 승낙과 제한)
① 당 사이트는 제6조의 규정에 의한 이용신청고객에 대하여 서비스 이용을 승낙합니다.
② 당 사이트는 아래사항에 해당하는 경우에 대해서 승낙하지 아니 합니다.
- 이용계약 신청서의 내용을 허위로 기재한 경우
- 기타 규정한 제반사항을 위반하며 신청하는 경우
제 9 조 (회원 ID 부여 및 변경 등)
① 당 사이트는 이용고객에 대하여 약관에 정하는 바에 따라 자신이 선정한 회원 ID를 부여합니다.
② 회원 ID는 원칙적으로 변경이 불가하며 부득이한 사유로 인하여 변경 하고자 하는 경우에는 해당 ID를
해지하고 재가입해야 합니다.
③ 기타 회원 개인정보 관리 및 변경 등에 관한 사항은 서비스별 안내에 정하는 바에 의합니다.
제 3 장 계약 당사자의 의무
제 10 조 (KISTI의 의무)
① 당 사이트는 이용고객이 희망한 서비스 제공 개시일에 특별한 사정이 없는 한 서비스를 이용할 수 있도록
하여야 합니다.
② 당 사이트는 개인정보 보호를 위해 보안시스템을 구축하며 개인정보 보호정책을 공시하고 준수합니다.
③ 당 사이트는 회원으로부터 제기되는 의견이나 불만이 정당하다고 객관적으로 인정될 경우에는 적절한 절차를
거쳐 즉시 처리하여야 합니다. 다만, 즉시 처리가 곤란한 경우는 회원에게 그 사유와 처리일정을 통보하여야
합니다.
제 11 조 (회원의 의무)
① 이용자는 회원가입 신청 또는 회원정보 변경 시 실명으로 모든 사항을 사실에 근거하여 작성하여야 하며,
허위 또는 타인의 정보를 등록할 경우 일체의 권리를 주장할 수 없습니다.
② 당 사이트가 관계법령 및 개인정보 보호정책에 의거하여 그 책임을 지는 경우를 제외하고 회원에게 부여된
ID의 비밀번호 관리소홀, 부정사용에 의하여 발생하는 모든 결과에 대한 책임은 회원에게 있습니다.
③ 회원은 당 사이트 및 제 3자의 지적 재산권을 침해해서는 안 됩니다.
제 4 장 서비스의 이용
제 12 조 (서비스 이용 시간)
① 서비스 이용은 당 사이트의 업무상 또는 기술상 특별한 지장이 없는 한 연중무휴, 1일 24시간 운영을
원칙으로 합니다. 단, 당 사이트는 시스템 정기점검, 증설 및 교체를 위해 당 사이트가 정한 날이나 시간에
서비스를 일시 중단할 수 있으며, 예정되어 있는 작업으로 인한 서비스 일시중단은 당 사이트 홈페이지를
통해 사전에 공지합니다.
② 당 사이트는 서비스를 특정범위로 분할하여 각 범위별로 이용가능시간을 별도로 지정할 수 있습니다. 다만
이 경우 그 내용을 공지합니다.
제 13 조 (홈페이지 저작권)
① NDSL에서 제공하는 모든 저작물의 저작권은 원저작자에게 있으며, KISTI는 복제/배포/전송권을 확보하고
있습니다.
② NDSL에서 제공하는 콘텐츠를 상업적 및 기타 영리목적으로 복제/배포/전송할 경우 사전에 KISTI의 허락을
받아야 합니다.
③ NDSL에서 제공하는 콘텐츠를 보도, 비평, 교육, 연구 등을 위하여 정당한 범위 안에서 공정한 관행에
합치되게 인용할 수 있습니다.
④ NDSL에서 제공하는 콘텐츠를 무단 복제, 전송, 배포 기타 저작권법에 위반되는 방법으로 이용할 경우
저작권법 제136조에 따라 5년 이하의 징역 또는 5천만 원 이하의 벌금에 처해질 수 있습니다.
제 14 조 (유료서비스)
① 당 사이트 및 협력기관이 정한 유료서비스(원문복사 등)는 별도로 정해진 바에 따르며, 변경사항은 시행 전에
당 사이트 홈페이지를 통하여 회원에게 공지합니다.
② 유료서비스를 이용하려는 회원은 정해진 요금체계에 따라 요금을 납부해야 합니다.
제 5 장 계약 해지 및 이용 제한
제 15 조 (계약 해지)
회원이 이용계약을 해지하고자 하는 때에는 [가입해지] 메뉴를 이용해 직접 해지해야 합니다.
제 16 조 (서비스 이용제한)
① 당 사이트는 회원이 서비스 이용내용에 있어서 본 약관 제 11조 내용을 위반하거나, 다음 각 호에 해당하는
경우 서비스 이용을 제한할 수 있습니다.
- 2년 이상 서비스를 이용한 적이 없는 경우
- 기타 정상적인 서비스 운영에 방해가 될 경우
② 상기 이용제한 규정에 따라 서비스를 이용하는 회원에게 서비스 이용에 대하여 별도 공지 없이 서비스 이용의
일시정지, 이용계약 해지 할 수 있습니다.
제 17 조 (전자우편주소 수집 금지)
회원은 전자우편주소 추출기 등을 이용하여 전자우편주소를 수집 또는 제3자에게 제공할 수 없습니다.
제 6 장 손해배상 및 기타사항
제 18 조 (손해배상)
당 사이트는 무료로 제공되는 서비스와 관련하여 회원에게 어떠한 손해가 발생하더라도 당 사이트가 고의 또는 과실로 인한 손해발생을 제외하고는 이에 대하여 책임을 부담하지 아니합니다.
제 19 조 (관할 법원)
서비스 이용으로 발생한 분쟁에 대해 소송이 제기되는 경우 민사 소송법상의 관할 법원에 제기합니다.
[부 칙]
1. (시행일) 이 약관은 2016년 9월 5일부터 적용되며, 종전 약관은 본 약관으로 대체되며, 개정된 약관의 적용일 이전 가입자도 개정된 약관의 적용을 받습니다.