Low-Budget, Cost-Effective OCR: Optical Character Recognition for MS-DOS Micros.
ERIC Educational Resources Information Center
Perez, Ernest
1990-01-01
Discusses optical character recognition (OCR) for use with MS-DOS microcomputers. Cost effectiveness is considered, three types of software approaches to character recognition are explained, hardware and operation requirements are described, possible library applications are discussed, future OCR developments are suggested, and a list of OCR…
A Complete OCR System for Tamil Magazine Documents
NASA Astrophysics Data System (ADS)
Kokku, Aparna; Chakravarthy, Srinivasa
We present a complete optical character recognition (OCR) system for Tamil magazines/documents. All the standard elements of OCR process like de-skewing, preprocessing, segmentation, character recognition, and reconstruction are implemented. Experience with OCR problems teaches that for most subtasks of OCR, there is no single technique that gives perfect results for every type of document image. We exploit the ability of neural networks to learn from experience in solving the problems of segmentation and character recognition. Text segmentation of Tamil newsprint poses a new challenge owing to its italic-like font type; problems that arise in recognition of touching and close characters are discussed. Character recognition efficiency varied from 94 to 97% for this type of font. The grouping of blocks into logical units and the determination of reading order within each logical unit helped us in reconstructing automatically the document image in an editable format.
Degraded character recognition based on gradient pattern
NASA Astrophysics Data System (ADS)
Babu, D. R. Ramesh; Ravishankar, M.; Kumar, Manish; Wadera, Kevin; Raj, Aakash
2010-02-01
Degraded character recognition is a challenging problem in the field of Optical Character Recognition (OCR). The performance of an optical character recognition depends upon printed quality of the input documents. Many OCRs have been designed which correctly identifies the fine printed documents. But, very few reported work has been found on the recognition of the degraded documents. The efficiency of the OCRs system decreases if the input image is degraded. In this paper, a novel approach based on gradient pattern for recognizing degraded printed character is proposed. The approach makes use of gradient pattern of an individual character for recognition. Experiments were conducted on character image that is either digitally written or a degraded character extracted from historical documents and the results are found to be satisfactory.
Robust keyword retrieval method for OCRed text
NASA Astrophysics Data System (ADS)
Fujii, Yusaku; Takebe, Hiroaki; Tanaka, Hiroshi; Hotta, Yoshinobu
2011-01-01
Document management systems have become important because of the growing popularity of electronic filing of documents and scanning of books, magazines, manuals, etc., through a scanner or a digital camera, for storage or reading on a PC or an electronic book. Text information acquired by optical character recognition (OCR) is usually added to the electronic documents for document retrieval. Since texts generated by OCR generally include character recognition errors, robust retrieval methods have been introduced to overcome this problem. In this paper, we propose a retrieval method that is robust against both character segmentation and recognition errors. In the proposed method, the insertion of noise characters and dropping of characters in the keyword retrieval enables robustness against character segmentation errors, and character substitution in the keyword of the recognition candidate for each character in OCR or any other character enables robustness against character recognition errors. The recall rate of the proposed method was 15% higher than that of the conventional method. However, the precision rate was 64% lower.
Document Form and Character Recognition using SVM
NASA Astrophysics Data System (ADS)
Park, Sang-Sung; Shin, Young-Geun; Jung, Won-Kyo; Ahn, Dong-Kyu; Jang, Dong-Sik
2009-08-01
Because of development of computer and information communication, EDI (Electronic Data Interchange) has been developing. There is OCR (Optical Character Recognition) of Pattern recognition technology for EDI. OCR contributed to changing many manual in the past into automation. But for the more perfect database of document, much manual is needed for excluding unnecessary recognition. To resolve this problem, we propose document form based character recognition method in this study. Proposed method is divided into document form recognition part and character recognition part. Especially, in character recognition, change character into binarization by using SVM algorithm and extract more correct feature value.
Design and development of an ancient Chinese document recognition system
NASA Astrophysics Data System (ADS)
Peng, Liangrui; Xiu, Pingping; Ding, Xiaoqing
2003-12-01
The digitization of ancient Chinese documents presents new challenges to OCR (Optical Character Recognition) research field due to the large character set of ancient Chinese characters, variant font types, and versatile document layout styles, as these documents are historical reflections to the thousands of years of Chinese civilization. After analyzing the general characteristics of ancient Chinese documents, we present a solution for recognition of ancient Chinese documents with regular font-types and layout-styles. Based on the previous work on multilingual OCR in TH-OCR system, we focus on the design and development of two key technologies which include character recognition and page segmentation. Experimental results show that the developed character recognition kernel of 19,635 Chinese characters outperforms our original traditional Chinese recognition kernel; Benchmarked test on printed ancient Chinese books proves that the proposed system is effective for regular ancient Chinese documents.
Identification of Matra Region and Overlapping Characters for OCR of Printed Bengali Scripts
NASA Astrophysics Data System (ADS)
Goswami, Subhra Sundar
One of the important reasons for poor recognition rate in optical character recognition (OCR) system is the error in character segmentation. In case of Bangla scripts, the errors occur due to several reasons, which include incorrect detection of matra (headline), over-segmentation and under-segmentation. We have proposed a robust method for detecting the headline region. Existence of overlapping characters (in under-segmented parts) in scanned printed documents is a major problem in designing an effective character segmentation procedure for OCR systems. In this paper, a predictive algorithm is developed for effectively identifying overlapping characters and then selecting the cut-borders for segmentation. Our method can be successfully used in achieving high recognition result.
An Evaluation of PC-Based Optical Character Recognition Systems.
ERIC Educational Resources Information Center
Schreier, E. M.; Uslan, M. M.
1991-01-01
The review examines six personal computer-based optical character recognition (OCR) systems designed for use by blind and visually impaired people. Considered are OCR components and terms, documentation, scanning and reading, command structure, conversion, unique features, accuracy of recognition, scanning time, speed, and cost. (DB)
Multi-frame knowledge based text enhancement for mobile phone captured videos
NASA Astrophysics Data System (ADS)
Ozarslan, Suleyman; Eren, P. Erhan
2014-02-01
In this study, we explore automated text recognition and enhancement using mobile phone captured videos of store receipts. We propose a method which includes Optical Character Resolution (OCR) enhanced by our proposed Row Based Multiple Frame Integration (RB-MFI), and Knowledge Based Correction (KBC) algorithms. In this method, first, the trained OCR engine is used for recognition; then, the RB-MFI algorithm is applied to the output of the OCR. The RB-MFI algorithm determines and combines the most accurate rows of the text outputs extracted by using OCR from multiple frames of the video. After RB-MFI, KBC algorithm is applied to these rows to correct erroneous characters. Results of the experiments show that the proposed video-based approach which includes the RB-MFI and the KBC algorithm increases the word character recognition rate to 95%, and the character recognition rate to 98%.
A System for Mailpiece ZIP Code Assignment through Contextual Analysis. Phase 2
1991-03-01
Segmentation Address Block Interpretation Automatic Feature Generation Word Recognition Feature Detection Word Verification Optical Character Recognition Directory...in the Phase III effort. 1.1 Motivation The United States Postal Service (USPS) deploys large numbers of optical character recognition (OCR) machines...4):208-218, November 1986. [2] Gronmeyer, L. K., Ruffin, B. W., Lybanon, M. A., Neely, P. L., and Pierce, S. E. An Overview of Optical Character Recognition (OCR
Teach Your Computer to Read: Scanners and Optical Character Recognition.
ERIC Educational Resources Information Center
Marsden, Jim
1993-01-01
Desktop scanners can be used with a software technology called optical character recognition (OCR) to convert the text on virtually any paper document into an electronic form. OCR offers educators new flexibility in incorporating text into tests, lesson plans, and other materials. (MLF)
Optical character recognition reading aid for the visually impaired.
Grandin, Juan Carlos; Cremaschi, Fabian; Lombardo, Elva; Vitu, Ed; Dujovny, Manuel
2008-06-01
An optical character recognition (OCR) reading machine is a significant help for visually impaired patients. An OCR reading machine is used. This instrument can provide a significant help in order to improve the quality of life of patients with low vision or blindness.
Optical Character Recognition.
ERIC Educational Resources Information Center
Converso, L.; Hocek, S.
1990-01-01
This paper describes computer-based optical character recognition (OCR) systems, focusing on their components (the computer, the scanner, the OCR, and the output device); how the systems work; and features to consider in selecting a system. A list of 26 questions to ask to evaluate systems for potential purchase is included. (JDD)
ERIC Educational Resources Information Center
Kichuk, Diana
2015-01-01
The electronic conversion of scanned image files to readable text using optical character recognition (OCR) software and the subsequent migration of raw OCR text to e-book text file formats are key remediation or media conversion technologies used in digital repository e-book production. Despite real progress, the OCR problem of reliability and…
Practical vision based degraded text recognition system
NASA Astrophysics Data System (ADS)
Mohammad, Khader; Agaian, Sos; Saleh, Hani
2011-02-01
Rapid growth and progress in the medical, industrial, security and technology fields means more and more consideration for the use of camera based optical character recognition (OCR) Applying OCR to scanned documents is quite mature, and there are many commercial and research products available on this topic. These products achieve acceptable recognition accuracy and reasonable processing times especially with trained software, and constrained text characteristics. Even though the application space for OCR is huge, it is quite challenging to design a single system that is capable of performing automatic OCR for text embedded in an image irrespective of the application. Challenges for OCR systems include; images are taken under natural real world conditions, Surface curvature, text orientation, font, size, lighting conditions, and noise. These and many other conditions make it extremely difficult to achieve reasonable character recognition. Performance for conventional OCR systems drops dramatically as the degradation level of the text image quality increases. In this paper, a new recognition method is proposed to recognize solid or dotted line degraded characters. The degraded text string is localized and segmented using a new algorithm. The new method was implemented and tested using a development framework system that is capable of performing OCR on camera captured images. The framework allows parameter tuning of the image-processing algorithm based on a training set of camera-captured text images. Novel methods were used for enhancement, text localization and the segmentation algorithm which enables building a custom system that is capable of performing automatic OCR which can be used for different applications. The developed framework system includes: new image enhancement, filtering, and segmentation techniques which enabled higher recognition accuracies, faster processing time, and lower energy consumption, compared with the best state of the art published techniques. The system successfully produced impressive OCR accuracies (90% -to- 93%) using customized systems generated by our development framework in two industrial OCR applications: water bottle label text recognition and concrete slab plate text recognition. The system was also trained for the Arabic language alphabet, and demonstrated extremely high recognition accuracy (99%) for Arabic license name plate text recognition with processing times of 10 seconds. The accuracy and run times of the system were compared to conventional and many states of art methods, the proposed system shows excellent results.
Iterative cross section sequence graph for handwritten character segmentation.
Dawoud, Amer
2007-08-01
The iterative cross section sequence graph (ICSSG) is an algorithm for handwritten character segmentation. It expands the cross section sequence graph concept by applying it iteratively at equally spaced thresholds. The iterative thresholding reduces the effect of information loss associated with image binarization. ICSSG preserves the characters' skeletal structure by preventing the interference of pixels that causes flooding of adjacent characters' segments. Improving the structural quality of the characters' skeleton facilitates better feature extraction and classification, which improves the overall performance of optical character recognition (OCR). Experimental results showed significant improvements in OCR recognition rates compared to other well-established segmentation algorithms.
Optical character recognition of handwritten Arabic using hidden Markov models
NASA Astrophysics Data System (ADS)
Aulama, Mohannad M.; Natsheh, Asem M.; Abandah, Gheith A.; Olama, Mohammed M.
2011-04-01
The problem of optical character recognition (OCR) of handwritten Arabic has not received a satisfactory solution yet. In this paper, an Arabic OCR algorithm is developed based on Hidden Markov Models (HMMs) combined with the Viterbi algorithm, which results in an improved and more robust recognition of characters at the sub-word level. Integrating the HMMs represents another step of the overall OCR trends being currently researched in the literature. The proposed approach exploits the structure of characters in the Arabic language in addition to their extracted features to achieve improved recognition rates. Useful statistical information of the Arabic language is initially extracted and then used to estimate the probabilistic parameters of the mathematical HMM. A new custom implementation of the HMM is developed in this study, where the transition matrix is built based on the collected large corpus, and the emission matrix is built based on the results obtained via the extracted character features. The recognition process is triggered using the Viterbi algorithm which employs the most probable sequence of sub-words. The model was implemented to recognize the sub-word unit of Arabic text raising the recognition rate from being linked to the worst recognition rate for any character to the overall structure of the Arabic language. Numerical results show that there is a potentially large recognition improvement by using the proposed algorithms.
Optical character recognition of handwritten Arabic using hidden Markov models
DOE Office of Scientific and Technical Information (OSTI.GOV)
Aulama, Mohannad M.; Natsheh, Asem M.; Abandah, Gheith A.
2011-01-01
The problem of optical character recognition (OCR) of handwritten Arabic has not received a satisfactory solution yet. In this paper, an Arabic OCR algorithm is developed based on Hidden Markov Models (HMMs) combined with the Viterbi algorithm, which results in an improved and more robust recognition of characters at the sub-word level. Integrating the HMMs represents another step of the overall OCR trends being currently researched in the literature. The proposed approach exploits the structure of characters in the Arabic language in addition to their extracted features to achieve improved recognition rates. Useful statistical information of the Arabic language ismore » initially extracted and then used to estimate the probabilistic parameters of the mathematical HMM. A new custom implementation of the HMM is developed in this study, where the transition matrix is built based on the collected large corpus, and the emission matrix is built based on the results obtained via the extracted character features. The recognition process is triggered using the Viterbi algorithm which employs the most probable sequence of sub-words. The model was implemented to recognize the sub-word unit of Arabic text raising the recognition rate from being linked to the worst recognition rate for any character to the overall structure of the Arabic language. Numerical results show that there is a potentially large recognition improvement by using the proposed algorithms.« less
NASA Technical Reports Server (NTRS)
Kiang, Richard K.
1992-01-01
Neural networks have been applied to classifications of remotely sensed data with some success. To improve the performance of this approach, an examination was made of how neural networks are applied to the optical character recognition (OCR) of handwritten digits and letters. A three-layer, feedforward network, along with techniques adopted from OCR, was used to classify Landsat-4 Thematic Mapper data. Good results were obtained. To overcome the difficulties that are characteristic of remote sensing applications and to attain significant improvements in classification accuracy, a special network architecture may be required.
U.S. Army Research Laboratory (ARL) Corporate Dari Document Transcription and Translation Guidelines
2012-10-01
text file format. 15. SUBJECT TERMS Transcription, Translation, guidelines, ground truth, Optical character recognition , OCR, Machine Translation, MT...foreign language into a target language in order to train, test, and evaluate optical character recognition (OCR) and machine translation (MT) embedded...graphic element and should not be transcribed. Elements that are not part of the primary text such as handwritten annotations or stamps should not be
A New Experiment on Bengali Character Recognition
NASA Astrophysics Data System (ADS)
Barman, Sumana; Bhattacharyya, Debnath; Jeon, Seung-Whan; Kim, Tai-Hoon; Kim, Haeng-Kon
This paper presents a method to use View based approach in Bangla Optical Character Recognition (OCR) system providing reduced data set to the ANN classification engine rather than the traditional OCR methods. It describes how Bangla characters are processed, trained and then recognized with the use of a Backpropagation Artificial neural network. This is the first published account of using a segmentation-free optical character recognition system for Bangla using a view based approach. The methodology presented here assumes that the OCR pre-processor has presented the input images to the classification engine described here. The size and the font face used to render the characters are also significant in both training and classification. The images are first converted into greyscale and then to binary images; these images are then scaled to a fit a pre-determined area with a fixed but significant number of pixels. The feature vectors are then formed extracting the characteristics points, which in this case is simply a series of 0s and 1s of fixed length. Finally, an artificial neural network is chosen for the training and classification process.
OCR Scanners Facilitate WP Training in Business Schools and Colleges.
ERIC Educational Resources Information Center
School Business Affairs, 1983
1983-01-01
Optical Character Recognition Scanners (OCR) scan typed text and feed it directly into word processing systems, saving input time. OCRs are valuable in word processing training programs because they allow more students access to classes and more time for skill training. (MD)
ERIC Educational Resources Information Center
Sun, Wei; And Others
1992-01-01
Identifies types and distributions of errors in text produced by optical character recognition (OCR) and proposes a process using machine learning techniques to recognize and correct errors in OCR texts. Results of experiments indicating that this strategy can reduce human interaction required for error correction are reported. (25 references)…
The Effects of Noisy Data on Text Retrieval.
ERIC Educational Resources Information Center
Taghva, Kazem; And Others
1994-01-01
Discusses the use of optical character recognition (OCR) for inputting documents in an information retrieval system and describes a study that used an OCR-generated database and its corresponding corrected version to examine query evaluation in the presence of noisy data. Scanning technology, recognition technology, and retrieval technology are…
Keyless Entry: Building a Text Database Using OCR Technology.
ERIC Educational Resources Information Center
Grotophorst, Clyde W.
1989-01-01
Discusses the use of optical character recognition (OCR) technology to produce an ASCII text database. A tutorial on digital scanning and OCR is provided, and a systems integration project which used the Calera CDP-3000XF scanner and text retrieval software to construct a database of dissertations at George Mason University is described. (four…
Study of the Effectiveness of OCR for Decentralized Data Capture and Conversion. Final Report.
ERIC Educational Resources Information Center
Liston, David M.; And Others
The ERIC network conversion to an OCR (Optical Character Recognition) mode of data entry was studied to analyze the potential effectiveness of OCR data entry for future EPC/s (Editorial Processing Centers). Study results are also applicable to any other system involving decentralized bibliographic data capture and conversion functions. The report…
Development of OCR system for portable passport and visa reader
NASA Astrophysics Data System (ADS)
Visilter, Yury V.; Zheltov, Sergey Y.; Lukin, Anton A.
1999-01-01
The modern passport and visa documents include special machine-readable zones satisfied the ICAO standards. This allows to develop the special passport and visa automatic readers. However, there are some special problems in such OCR systems: low resolution of character images captured by CCD-camera (down to 150 dpi), essential shifts and slopes (up to 10 degrees), rich paper texture under the character symbols, non-homogeneous illumination. This paper presents the structure and some special aspects of OCR system for portable passport and visa reader. In our approach the binarization procedure is performed after the segmentation step, and it is applied to the each character site separately. Character recognition procedure uses the structural information of machine-readable zone. Special algorithms are developed for machine-readable zone extraction and character segmentation.
Lee, Young Han; Song, Ho-Taek; Suh, Jin-Suck
2012-12-01
The objectives are (1) to introduce a new concept of making a quantitative computed tomography (QCT) reporting system by using optical character recognition (OCR) and macro program and (2) to illustrate the practical usages of the QCT reporting system in radiology reading environment. This reporting system was created as a development tool by using an open-source OCR software and an open-source macro program. The main module was designed for OCR to report QCT images in radiology reading process. The principal processes are as follows: (1) to save a QCT report as a graphic file, (2) to recognize the characters from an image as a text, (3) to extract the T scores from the text, (4) to perform error correction, (5) to reformat the values into QCT radiology reporting template, and (6) to paste the reports into the electronic medical record (EMR) or picture archiving and communicating system (PACS). The accuracy test of OCR was performed on randomly selected QCTs. QCT as a radiology reporting tool successfully acted as OCR of QCT. The diagnosis of normal, osteopenia, or osteoporosis is also determined. Error correction of OCR is done with AutoHotkey-coded module. The results of T scores of femoral neck and lumbar vertebrae had an accuracy of 100 and 95.4 %, respectively. A convenient QCT reporting system could be established by utilizing open-source OCR software and open-source macro program. This method can be easily adapted for other QCT applications and PACS/EMR.
Neural network-based systems for handprint OCR applications.
Ganis, M D; Wilson, C L; Blue, J L
1998-01-01
Over the last five years or so, neural network (NN)-based approaches have been steadily gaining performance and popularity for a wide range of optical character recognition (OCR) problems, from isolated digit recognition to handprint recognition. We present an NN classification scheme based on an enhanced multilayer perceptron (MLP) and describe an end-to-end system for form-based handprint OCR applications designed by the National Institute of Standards and Technology (NIST) Visual Image Processing Group. The enhancements to the MLP are based on (i) neuron activations functions that reduce the occurrences of singular Jacobians; (ii) successive regularization to constrain the volume of the weight space; and (iii) Boltzmann pruning to constrain the dimension of the weight space. Performance characterization studies of NN systems evaluated at the first OCR systems conference and the NIST form-based handprint recognition system are also summarized.
Boost OCR accuracy using iVector based system combination approach
NASA Astrophysics Data System (ADS)
Peng, Xujun; Cao, Huaigu; Natarajan, Prem
2015-01-01
Optical character recognition (OCR) is a challenging task because most existing preprocessing approaches are sensitive to writing style, writing material, noises and image resolution. Thus, a single recognition system cannot address all factors of real document images. In this paper, we describe an approach to combine diverse recognition systems by using iVector based features, which is a newly developed method in the field of speaker verification. Prior to system combination, document images are preprocessed and text line images are extracted with different approaches for each system, where iVector is transformed from a high-dimensional supervector of each text line and is used to predict the accuracy of OCR. We merge hypotheses from multiple recognition systems according to the overlap ratio and the predicted OCR score of text line images. We present evaluation results on an Arabic document database where the proposed method is compared against the single best OCR system using word error rate (WER) metric.
Guideline for Optical Character Recognition Forms.
ERIC Educational Resources Information Center
National Bureau of Standards (DOC), Washington, DC.
This publication provides materials relating to the design, preparation, acquisition, inspection, and application of Optical Character Recognition (OCR) forms in data entry systems. Since the materials are advisory and tutorial in nature, this publication has been issued as a guideline rather than as a standard in the Federal Information…
A comparison study between MLP and convolutional neural network models for character recognition
NASA Astrophysics Data System (ADS)
Ben Driss, S.; Soua, M.; Kachouri, R.; Akil, M.
2017-05-01
Optical Character Recognition (OCR) systems have been designed to operate on text contained in scanned documents and images. They include text detection and character recognition in which characters are described then classified. In the classification step, characters are identified according to their features or template descriptions. Then, a given classifier is employed to identify characters. In this context, we have proposed the unified character descriptor (UCD) to represent characters based on their features. Then, matching was employed to ensure the classification. This recognition scheme performs a good OCR Accuracy on homogeneous scanned documents, however it cannot discriminate characters with high font variation and distortion.3 To improve recognition, classifiers based on neural networks can be used. The multilayer perceptron (MLP) ensures high recognition accuracy when performing a robust training. Moreover, the convolutional neural network (CNN), is gaining nowadays a lot of popularity for its high performance. Furthermore, both CNN and MLP may suffer from the large amount of computation in the training phase. In this paper, we establish a comparison between MLP and CNN. We provide MLP with the UCD descriptor and the appropriate network configuration. For CNN, we employ the convolutional network designed for handwritten and machine-printed character recognition (Lenet-5) and we adapt it to support 62 classes, including both digits and characters. In addition, GPU parallelization is studied to speed up both of MLP and CNN classifiers. Based on our experimentations, we demonstrate that the used real-time CNN is 2x more relevant than MLP when classifying characters.
Study of style effects on OCR errors in the MEDLINE database
NASA Astrophysics Data System (ADS)
Garrison, Penny; Davis, Diane L.; Andersen, Tim L.; Barney Smith, Elisa H.
2005-01-01
The National Library of Medicine has developed a system for the automatic extraction of data from scanned journal articles to populate the MEDLINE database. Although the 5-engine OCR system used in this process exhibits good performance overall, it does make errors in character recognition that must be corrected in order for the process to achieve the requisite accuracy. The correction process works by feeding words that have characters with less than 100% confidence (as determined automatically by the OCR engine) to a human operator who then must manually verify the word or correct the error. The majority of these errors are contained in the affiliation information zone where the characters are in italics or small fonts. Therefore only affiliation information data is used in this research. This paper examines the correlation between OCR errors and various character attributes in the MEDLINE database, such as font size, italics, bold, etc. and OCR confidence levels. The motivation for this research is that if a correlation between the character style and types of errors exists it should be possible to use this information to improve operator productivity by increasing the probability that the correct word option is presented to the human editor. We have determined that this correlation exists, in particular for the case of characters with diacritics.
Benchmark for license plate character segmentation
NASA Astrophysics Data System (ADS)
Gonçalves, Gabriel Resende; da Silva, Sirlene Pio Gomes; Menotti, David; Shwartz, William Robson
2016-09-01
Automatic license plate recognition (ALPR) has been the focus of many researches in the past years. In general, ALPR is divided into the following problems: detection of on-track vehicles, license plate detection, segmentation of license plate characters, and optical character recognition (OCR). Even though commercial solutions are available for controlled acquisition conditions, e.g., the entrance of a parking lot, ALPR is still an open problem when dealing with data acquired from uncontrolled environments, such as roads and highways when relying only on imaging sensors. Due to the multiple orientations and scales of the license plates captured by the camera, a very challenging task of the ALPR is the license plate character segmentation (LPCS) step, because its effectiveness is required to be (near) optimal to achieve a high recognition rate by the OCR. To tackle the LPCS problem, this work proposes a benchmark composed of a dataset designed to focus specifically on the character segmentation step of the ALPR within an evaluation protocol. Furthermore, we propose the Jaccard-centroid coefficient, an evaluation measure more suitable than the Jaccard coefficient regarding the location of the bounding box within the ground-truth annotation. The dataset is composed of 2000 Brazilian license plates consisting of 14000 alphanumeric symbols and their corresponding bounding box annotations. We also present a straightforward approach to perform LPCS efficiently. Finally, we provide an experimental evaluation for the dataset based on five LPCS approaches and demonstrate the importance of character segmentation for achieving an accurate OCR.
Kansas State University Libraries' OCR Labeling Project.
ERIC Educational Resources Information Center
Thierer, Joyce; Bower, Merry
This publication describes the planning and implementation of an optical character recognition (OCR) labeling project, the first stage of Kansas State University (KSU) Libraries' program of conversion from a manual to an automated circulation system. It is noted that a telephone survey of libraries with automated circulation systems and…
The impact of OCR accuracy on automated cancer classification of pathology reports.
Zuccon, Guido; Nguyen, Anthony N; Bergheim, Anton; Wickman, Sandra; Grayson, Narelle
2012-01-01
To evaluate the effects of Optical Character Recognition (OCR) on the automatic cancer classification of pathology reports. Scanned images of pathology reports were converted to electronic free-text using a commercial OCR system. A state-of-the-art cancer classification system, the Medical Text Extraction (MEDTEX) system, was used to automatically classify the OCR reports. Classifications produced by MEDTEX on the OCR versions of the reports were compared with the classification from a human amended version of the OCR reports. The employed OCR system was found to recognise scanned pathology reports with up to 99.12% character accuracy and up to 98.95% word accuracy. Errors in the OCR processing were found to minimally impact on the automatic classification of scanned pathology reports into notifiable groups. However, the impact of OCR errors is not negligible when considering the extraction of cancer notification items, such as primary site, histological type, etc. The automatic cancer classification system used in this work, MEDTEX, has proven to be robust to errors produced by the acquisition of freetext pathology reports from scanned images through OCR software. However, issues emerge when considering the extraction of cancer notification items.
Automatic Cataloguing and Searching for Retrospective Data by Use of OCR Text.
ERIC Educational Resources Information Center
Tseng, Yuen-Hsien
2001-01-01
Describes efforts in supporting information retrieval from OCR (optical character recognition) degraded text. Reports on approaches used in an automatic cataloging and searching contest for books in multiple languages, including a vector space retrieval model, an n-gram indexing method, and a weighting scheme; and discusses problems of Asian…
Effects of OCR Errors on Ranking and Feedback Using the Vector Space Model.
ERIC Educational Resources Information Center
Taghva, Kazem; And Others
1996-01-01
Reports on the performance of the vector space model in the presence of OCR (optical character recognition) errors in information retrieval. Highlights include precision and recall, a full-text test collection, smart vector representation, impact of weighting parameters, ranking variability, and the effect of relevance feedback. (Author/LRW)
Biondich, Paul G; Overhage, J Marc; Dexter, Paul R; Downs, Stephen M; Lemmon, Larry; McDonald, Clement J
2002-01-01
Advances in optical character recognition (OCR) software and computer hardware have stimulated a reevaluation of the technology and its ability to capture structured clinical data from preexisting paper forms. In our pilot evaluation, we measured the accuracy and feasibility of capturing vitals data from a pediatric encounter form that has been in use for over twenty years. We found that the software had a digit recognition rate of 92.4% (95% confidence interval: 91.6 to 93.2) overall. More importantly, this system was approximately three times as fast as our existing method of data entry. These preliminary results suggest that with further refinements in the approach and additional development, we may be able to incorporate OCR as another method for capturing structured clinical data.
An evaluation of information retrieval accuracy with simulated OCR output
DOE Office of Scientific and Technical Information (OSTI.GOV)
Croft, W.B.; Harding, S.M.; Taghva, K.
Optical Character Recognition (OCR) is a critical part of many text-based applications. Although some commercial systems use the output from OCR devices to index documents without editing, there is very little quantitative data on the impact of OCR errors on the accuracy of a text retrieval system. Because of the difficulty of constructing test collections to obtain this data, we have carried out evaluation using simulated OCR output on a variety of databases. The results show that high quality OCR devices have little effect on the accuracy of retrieval, but low quality devices used with databases of short documents canmore » result in significant degradation.« less
ERIC Educational Resources Information Center
McClean, Clare M.
1998-01-01
Reviews strengths and weaknesses of five optical character recognition (OCR) software packages used to digitize paper documents before publishing on the Internet. Outlines options available and stages of the conversion process. Describes the learning experience of Eurotext, a United Kingdom-based electronic libraries project (eLib). (PEN)
Arabic Optical Character Recognition (OCR) Evaluation in Order to Develop a Post-OCR Module
2011-09-01
handwritten, and many more have some handwriting in the margins. Some images are blurred or faded to the point of illegibility. Others are mostly or...it is to English, because Arabic has more features such as agreement. We say that Arabic is more “morphologically rich” than English. We intend to
Biondich, Paul G.; Overhage, J. Marc; Dexter, Paul R.; Downs, Stephen M.; Lemmon, Larry; McDonald, Clement J.
2002-01-01
Advances in optical character recognition (OCR) software and computer hardware have stimulated a reevaluation of the technology and its ability to capture structured clinical data from preexisting paper forms. In our pilot evaluation, we measured the accuracy and feasibility of capturing vitals data from a pediatric encounter form that has been in use for over twenty years. We found that the software had a digit recognition rate of 92.4% (95% confidence interval: 91.6 to 93.2) overall. More importantly, this system was approximately three times as fast as our existing method of data entry. These preliminary results suggest that with further refinements in the approach and additional development, we may be able to incorporate OCR as another method for capturing structured clinical data. PMID:12463786
Improved document image segmentation algorithm using multiresolution morphology
NASA Astrophysics Data System (ADS)
Bukhari, Syed Saqib; Shafait, Faisal; Breuel, Thomas M.
2011-01-01
Page segmentation into text and non-text elements is an essential preprocessing step before optical character recognition (OCR) operation. In case of poor segmentation, an OCR classification engine produces garbage characters due to the presence of non-text elements. This paper describes modifications to the text/non-text segmentation algorithm presented by Bloomberg,1 which is also available in his open-source Leptonica library.2The modifications result in significant improvements and achieved better segmentation accuracy than the original algorithm for UW-III, UNLV, ICDAR 2009 page segmentation competition test images and circuit diagram datasets.
Neural Network--OCR/ICR Recognology: Theory and Applications.
ERIC Educational Resources Information Center
Schantz, Herbert F.
1993-01-01
Explains the value of neurocomputing as a unique and effective new technological concept for information processing and optical character recognition. Comparisons are made to digital computing and examples of applications such as recognizing handprinted characters are addressed. Products available from various companies are described. (Contains…
Optical character recognition: an illustrated guide to the frontier
NASA Astrophysics Data System (ADS)
Nagy, George; Nartker, Thomas A.; Rice, Stephen V.
1999-12-01
We offer a perspective on the performance of current OCR systems by illustrating and explaining actual OCR errors made by three commercial devices. After discussing briefly the character recognition abilities of humans and computers, we present illustrated examples of recognition errors. The top level of our taxonomy of the causes of errors consists of Imaging Defects, Similar Symbols, Punctuation, and Typography. The analysis of a series of 'snippets' from this perspective provides insight into the strengths and weaknesses of current systems, and perhaps a road map to future progress. The examples were drawn from the large-scale tests conducted by the authors at the Information Science Research Institute of the University of Nevada, Las Vegas. By way of conclusion, we point to possible approaches for improving the accuracy of today's systems. The talk is based on our eponymous monograph, recently published in The Kluwer International Series in Engineering and Computer Science, Kluwer Academic Publishers, 1999.
A super resolution framework for low resolution document image OCR
NASA Astrophysics Data System (ADS)
Ma, Di; Agam, Gady
2013-01-01
Optical character recognition is widely used for converting document images into digital media. Existing OCR algorithms and tools produce good results from high resolution, good quality, document images. In this paper, we propose a machine learning based super resolution framework for low resolution document image OCR. Two main techniques are used in our proposed approach: a document page segmentation algorithm and a modified K-means clustering algorithm. Using this approach, by exploiting coherence in the document, we reconstruct from a low resolution document image a better resolution image and improve OCR results. Experimental results show substantial gain in low resolution documents such as the ones captured from video.
Trigram-based algorithms for OCR result correction
NASA Astrophysics Data System (ADS)
Bulatov, Konstantin; Manzhikov, Temudzhin; Slavin, Oleg; Faradjev, Igor; Janiszewski, Igor
2017-03-01
In this paper we consider a task of improving optical character recognition (OCR) results of document fields on low-quality and average-quality images using N-gram models. Cyrillic fields of Russian Federation internal passport are analyzed as an example. Two approaches are presented: the first one is based on hypothesis of dependence of a symbol from two adjacent symbols and the second is based on calculation of marginal distributions and Bayesian networks computation. A comparison of the algorithms and experimental results within a real document OCR system are presented, it's showed that the document field OCR accuracy can be improved by more than 6% for low-quality images.
The use of Optical Character Recognition (OCR) in the digitisation of herbarium specimen labels.
Drinkwater, Robyn E; Cubey, Robert W N; Haston, Elspeth M
2014-01-01
At the Royal Botanic Garden Edinburgh (RBGE) the use of Optical Character Recognition (OCR) to aid the digitisation process has been investigated. This was tested using a herbarium specimen digitisation process with two stages of data entry. Records were initially batch-processed to add data extracted from the OCR text prior to being sorted based on Collector and/or Country. Using images of the specimens, a team of six digitisers then added data to the specimen records. To investigate whether the data from OCR aid the digitisation process, they completed a series of trials which compared the efficiency of data entry between sorted and unsorted batches of specimens. A survey was carried out to explore the opinion of the digitisation staff to the different sorting options. In total 7,200 specimens were processed. When compared to an unsorted, random set of specimens, those which were sorted based on data added from the OCR were quicker to digitise. Of the methods tested here, the most successful in terms of efficiency used a protocol which required entering data into a limited set of fields and where the records were filtered by Collector and Country. The survey and subsequent discussions with the digitisation staff highlighted their preference for working with sorted specimens, in which label layout, locations and handwriting are likely to be similar, and so a familiarity with the Collector or Country is rapidly established.
NASA Astrophysics Data System (ADS)
Ben Salah, Ahmed; Ragot, Nicolas; Paquet, Thierry
2013-01-01
The French National Library (BnF*) has launched many mass digitization projects in order to give access to its collection. The indexation of digital documents on Gallica (digital library of the BnF) is done through their textual content obtained thanks to service providers that use Optical Character Recognition softwares (OCR). OCR softwares have become increasingly complex systems composed of several subsystems dedicated to the analysis and the recognition of the elements in a page. However, the reliability of these systems is always an issue at stake. Indeed, in some cases, we can find errors in OCR outputs that occur because of an accumulation of several errors at different levels in the OCR process. One of the frequent errors in OCR outputs is the missed text components. The presence of such errors may lead to severe defects in digital libraries. In this paper, we investigate the detection of missed text components to control the OCR results from the collections of the French National Library. Our verification approach uses local information inside the pages based on Radon transform descriptors and Local Binary Patterns descriptors (LBP) coupled with OCR results to control their consistency. The experimental results show that our method detects 84.15% of the missed textual components, by comparing the OCR ALTO files outputs (produced by the service providers) to the images of the document.
Text vectorization based on character recognition and character stroke modeling
NASA Astrophysics Data System (ADS)
Fan, Zhigang; Zhou, Bingfeng; Tse, Francis; Mu, Yadong; He, Tao
2014-03-01
In this paper, a text vectorization method is proposed using OCR (Optical Character Recognition) and character stroke modeling. This is based on the observation that for a particular character, its font glyphs may have different shapes, but often share same stroke structures. Like many other methods, the proposed algorithm contains two procedures, dominant point determination and data fitting. The first one partitions the outlines into segments and second one fits a curve to each segment. In the proposed method, the dominant points are classified as "major" (specifying stroke structures) and "minor" (specifying serif shapes). A set of rules (parameters) are determined offline specifying for each character the number of major and minor dominant points and for each dominant point the detection and fitting parameters (projection directions, boundary conditions and smoothness). For minor points, multiple sets of parameters could be used for different fonts. During operation, OCR is performed and the parameters associated with the recognized character are selected. Both major and minor dominant points are detected as a maximization process as specified by the parameter set. For minor points, an additional step could be performed to test the competing hypothesis and detect degenerated cases.
Post processing of optically recognized text via second order hidden Markov model
NASA Astrophysics Data System (ADS)
Poudel, Srijana
In this thesis, we describe a postprocessing system on Optical Character Recognition(OCR) generated text. Second Order Hidden Markov Model (HMM) approach is used to detect and correct the OCR related errors. The reason for choosing the 2nd order HMM is to keep track of the bigrams so that the model can represent the system more accurately. Based on experiments with training data of 159,733 characters and testing of 5,688 characters, the model was able to correct 43.38 % of the errors with a precision of 75.34 %. However, the precision value indicates that the model introduced some new errors, decreasing the correction percentage to 26.4%.
Towards Mobile OCR: How To Take a Good Picture of a Document Without Sight.
Cutter, Michael; Manduchi, Roberto
The advent of mobile OCR (optical character recognition) applications on regular smartphones holds great promise for enabling blind people to access printed information. Unfortunately, these systems suffer from a problem: in order for OCR output to be meaningful, a well-framed image of the document needs to be taken, something that is difficult to do without sight. This contribution presents an experimental investigation of how blind people position and orient a camera phone while acquiring document images. We developed experimental software to investigate if verbal guidance aids in the acquisition of OCR-readable images without sight. We report on our participant's feedback and performance before and after assistance from our software.
Towards Mobile OCR: How To Take a Good Picture of a Document Without Sight
Cutter, Michael; Manduchi, Roberto
2015-01-01
The advent of mobile OCR (optical character recognition) applications on regular smartphones holds great promise for enabling blind people to access printed information. Unfortunately, these systems suffer from a problem: in order for OCR output to be meaningful, a well-framed image of the document needs to be taken, something that is difficult to do without sight. This contribution presents an experimental investigation of how blind people position and orient a camera phone while acquiring document images. We developed experimental software to investigate if verbal guidance aids in the acquisition of OCR-readable images without sight. We report on our participant's feedback and performance before and after assistance from our software. PMID:26677461
NASA Astrophysics Data System (ADS)
Lam, Meng Chun; Nizam, Siti Soleha Muhammad; Arshad, Haslina; A'isyah Ahmad Shukri, Saidatul; Hashim, Nurhazarifah Che; Putra, Haekal Mozzia; Abidin, Rimaniza Zainal
2017-10-01
This article discusses the usability of an interactive application for halal products using Optical Character Recognition (OCR) and Augmented Reality (AR) technologies. Among the problems that have been identified in this study is that consumers have little knowledge about the E-Code. Therefore, users often have doubts about the halal status of the product. Nowadays, the integrity of halal status can be doubtful due to the actions of some irresponsible people spreading false information about a product. Therefore, an application that uses OCR and AR technology developed in this study will help the users to identify the information content of a product by scanning the E-Code label and by scanning the product's brand to know the halal status of the product. In this application, E-Code on the label of a product is scanned using OCR technology to display information about the E-Code. The product's brand is scan using augmented reality technology to display halal status of the product. The findings reveal that users are satisfied with this application and it is useful and easy to use.
The use of Optical Character Recognition (OCR) in the digitisation of herbarium specimen labels
Drinkwater, Robyn E.; Cubey, Robert W. N.; Haston, Elspeth M.
2014-01-01
Abstract At the Royal Botanic Garden Edinburgh (RBGE) the use of Optical Character Recognition (OCR) to aid the digitisation process has been investigated. This was tested using a herbarium specimen digitisation process with two stages of data entry. Records were initially batch-processed to add data extracted from the OCR text prior to being sorted based on Collector and/or Country. Using images of the specimens, a team of six digitisers then added data to the specimen records. To investigate whether the data from OCR aid the digitisation process, they completed a series of trials which compared the efficiency of data entry between sorted and unsorted batches of specimens. A survey was carried out to explore the opinion of the digitisation staff to the different sorting options. In total 7,200 specimens were processed. When compared to an unsorted, random set of specimens, those which were sorted based on data added from the OCR were quicker to digitise. Of the methods tested here, the most successful in terms of efficiency used a protocol which required entering data into a limited set of fields and where the records were filtered by Collector and Country. The survey and subsequent discussions with the digitisation staff highlighted their preference for working with sorted specimens, in which label layout, locations and handwriting are likely to be similar, and so a familiarity with the Collector or Country is rapidly established. PMID:25009435
Automatic feature design for optical character recognition using an evolutionary search procedure.
Stentiford, F W
1985-03-01
An automatic evolutionary search is applied to the problem of feature extraction in an OCR application. A performance measure based on feature independence is used to generate features which do not appear to suffer from peaking effects [17]. Features are extracted from a training set of 30 600 machine printed 34 class alphanumeric characters derived from British mail. Classification results on the training set and a test set of 10 200 characters are reported for an increasing number of features. A 1.01 percent forced decision error rate is obtained on the test data using 316 features. The hardware implementation should be cheap and fast to operate. The performance compares favorably with current low cost OCR page readers.
Warped document image correction method based on heterogeneous registration strategies
NASA Astrophysics Data System (ADS)
Tong, Lijing; Zhan, Guoliang; Peng, Quanyao; Li, Yang; Li, Yifan
2013-03-01
With the popularity of digital camera and the application requirement of digitalized document images, using digital cameras to digitalize document images has become an irresistible trend. However, the warping of the document surface impacts on the quality of the Optical Character Recognition (OCR) system seriously. To improve the warped document image's vision quality and the OCR rate, this paper proposed a warped document image correction method based on heterogeneous registration strategies. This method mosaics two warped images of the same document from different viewpoints. Firstly, two feature points are selected from one image. Then the two feature points are registered in the other image base on heterogeneous registration strategies. At last, image mosaics are done for the two images, and the best mosaiced image is selected by OCR recognition results. As a result, for the best mosaiced image, the distortions are mostly removed and the OCR results are improved markedly. Experimental results show that the proposed method can resolve the issue of warped document image correction more effectively.
Printed Arabic optical character segmentation
NASA Astrophysics Data System (ADS)
Mohammad, Khader; Ayyesh, Muna; Qaroush, Aziz; Tumar, Iyad
2015-03-01
A considerable progress in recognition techniques for many non-Arabic characters has been achieved. In contrary, few efforts have been put on the research of Arabic characters. In any Optical Character Recognition (OCR) system the segmentation step is usually the essential stage in which an extensive portion of processing is devoted and a considerable share of recognition errors is attributed. In this research, a novel segmentation approach for machine Arabic printed text with diacritics is proposed. The proposed method reduces computation, errors, gives a clear description for the sub-word and has advantages over using the skeleton approach in which the data and information of the character can be lost. Both of initial evaluation and testing of the proposed method have been developed using MATLAB and shows 98.7% promising results.
Review of chart recognition in document images
NASA Astrophysics Data System (ADS)
Liu, Yan; Lu, Xiaoqing; Qin, Yeyang; Tang, Zhi; Xu, Jianbo
2013-01-01
As an effective information transmitting way, chart is widely used to represent scientific statistics datum in books, research papers, newspapers etc. Though textual information is still the major source of data, there has been an increasing trend of introducing graphs, pictures, and figures into the information pool. Text recognition techniques for documents have been accomplished using optical character recognition (OCR) software. Chart recognition techniques as a necessary supplement of OCR for document images are still an unsolved problem due to the great subjectiveness and variety of charts styles. This paper reviews the development process of chart recognition techniques in the past decades and presents the focuses of current researches. The whole process of chart recognition is presented systematically, which mainly includes three parts: chart segmentation, chart classification, and chart Interpretation. In each part, the latest research work is introduced. In the last, the paper concludes with a summary and promising future research direction.
Public domain optical character recognition
NASA Astrophysics Data System (ADS)
Garris, Michael D.; Blue, James L.; Candela, Gerald T.; Dimmick, Darrin L.; Geist, Jon C.; Grother, Patrick J.; Janet, Stanley A.; Wilson, Charles L.
1995-03-01
A public domain document processing system has been developed by the National Institute of Standards and Technology (NIST). The system is a standard reference form-based handprint recognition system for evaluating optical character recognition (OCR), and it is intended to provide a baseline of performance on an open application. The system's source code, training data, performance assessment tools, and type of forms processed are all publicly available. The system recognizes the handprint entered on handwriting sample forms like the ones distributed with NIST Special Database 1. From these forms, the system reads hand-printed numeric fields, upper and lowercase alphabetic fields, and unconstrained text paragraphs comprised of words from a limited-size dictionary. The modular design of the system makes it useful for component evaluation and comparison, training and testing set validation, and multiple system voting schemes. The system contains a number of significant contributions to OCR technology, including an optimized probabilistic neural network (PNN) classifier that operates a factor of 20 times faster than traditional software implementations of the algorithm. The source code for the recognition system is written in C and is organized into 11 libraries. In all, there are approximately 19,000 lines of code supporting more than 550 subroutines. Source code is provided for form registration, form removal, field isolation, field segmentation, character normalization, feature extraction, character classification, and dictionary-based postprocessing. The recognition system has been successfully compiled and tested on a host of UNIX workstations. This paper gives an overview of the recognition system's software architecture, including descriptions of the various system components along with timing and accuracy statistics.
ERIC Educational Resources Information Center
Lazzaro, Joseph J.
1993-01-01
Describes adaptive technology for personal computers that accommodate disabled users and may require special equipment including hardware, memory, expansion slots, and ports. Highlights include vision aids, including speech synthesizers, magnification, braille, and optical character recognition (OCR); hearing adaptations; motor-impaired…
Arabic OCR: toward a complete system
NASA Astrophysics Data System (ADS)
El-Bialy, Ahmed M.; Kandil, Ahmed H.; Hashish, Mohamed; Yamany, Sameh M.
1999-12-01
Latin and Chinese OCR systems have been studied extensively in the literature. Yet little work was performed for Arabic character recognition. This is due to the technical challenges found in the Arabic text. Due to its cursive nature, a powerful and stable text segmentation is needed. Also; features capturing the characteristics of the rich Arabic character representation are needed to build the Arabic OCR. In this paper a novel segmentation technique which is font and size independent is introduced. This technique can segment the cursive written text line even if the line suffers from small skewness. The technique is not sensitive to the location of the centerline of the text line and can segment different font sizes and type (for different character sets) occurring on the same line. Features extraction is considered one of the most important phases of the text reading system. Ideally, the features extracted from a character image should capture the essential characteristics of this character that are independent of the font type and size. In such ideal case, the classifier stores a single prototype per character. However, it is practically challenging to find such ideal set of features. In this paper, a set of features that reflect the topological aspects of Arabia characters is proposed. These proposed features integrated with a topological matching technique introduce an Arabic text reading system that is semi Omni.
Intelligent form removal with character stroke preservation
NASA Astrophysics Data System (ADS)
Garris, Michael D.
1996-03-01
A new technique for intelligent form removal has been developed along with a new method for evaluating its impact on optical character recognition (OCR). All the dominant lines in the image are automatically detected using the Hough line transform and intelligently erased while simultaneously preserving overlapping character strokes by computing line width statistics and keying off of certain visual cues. This new method of form removal operates on loosely defined zones with no image deskewing. Any field in which the writer is provided a horizontal line to enter a response can be processed by this method. Several examples of processed fields are provided, including a comparison of results between the new method and a commercially available forms removal package. Even if this new form removal method did not improve character recognition accuracy, it is still a significant improvement to the technology because the requirement of a priori knowledge of the form's geometric details has been greatly reduced. This relaxes the recognition system's dependence on rigid form design, printing, and reproduction by automatically detecting and removing some of the physical structures (lines) on the form. Using the National Institute of Standards and Technology (NIST) public domain form-based handprint recognition system, the technique was tested on a large number of fields containing randomly ordered handprinted lowercase alphabets, as these letters (especially those with descenders) frequently touch and extend through the line along which they are written. Preserving character strokes improves overall lowercase recognition performance by 3%, which is a net improvement, but a single performance number like this doesn't communicate how the recognition process was really influenced. There is expected to be trade- offs with the introduction of any new technique into a complex recognition system. To understand both the improvements and the trade-offs, a new analysis was designed to compare the statistical distributions of individual confusion pairs between two systems. As OCR technology continues to improve, sophisticated analyses like this are necessary to reduce the errors remaining in complex recognition problems.
Lee, Young Han; Park, Eun Hae; Suh, Jin-Suck
2015-01-01
The objectives are: 1) to introduce a simple and efficient method for extracting region of interest (ROI) values from a Picture Archiving and Communication System (PACS) viewer using optical character recognition (OCR) software and a macro program, and 2) to evaluate the accuracy of this method with a PACS workstation. This module was designed to extract the ROI values on the images of the PACS, and created as a development tool by using open-source OCR software and an open-source macro program. The principal processes are as follows: (1) capture a region of the ROI values as a graphic file for OCR, (2) recognize the text from the captured image by OCR software, (3) perform error-correction, (4) extract the values including area, average, standard deviation, max, and min values from the text, (5) reformat the values into temporary strings with tabs, and (6) paste the temporary strings into the spreadsheet. This principal process was repeated for the number of ROIs. The accuracy of this module was evaluated on 1040 recognitions from 280 randomly selected ROIs of the magnetic resonance images. The input times of ROIs were compared between conventional manual method and this extraction module-assisted input method. The module for extracting ROI values operated successfully using the OCR and macro programs. The values of the area, average, standard deviation, maximum, and minimum could be recognized and error-corrected with AutoHotkey-coded module. The average input times using the conventional method and the proposed module-assisted method were 34.97 seconds and 7.87 seconds, respectively. A simple and efficient method for ROI value extraction was developed with open-source OCR and a macro program. Accurate inputs of various numbers from ROIs can be extracted with this module. The proposed module could be applied to the next generation of PACS or existing PACS that have not yet been upgraded. Copyright © 2015 AUR. Published by Elsevier Inc. All rights reserved.
Document image cleanup and binarization
NASA Astrophysics Data System (ADS)
Wu, Victor; Manmatha, Raghaven
1998-04-01
Image binarization is a difficult task for documents with text over textured or shaded backgrounds, poor contrast, and/or considerable noise. Current optical character recognition (OCR) and document analysis technology do not handle such documents well. We have developed a simple yet effective algorithm for document image clean-up and binarization. The algorithm consists of two basic steps. In the first step, the input image is smoothed using a low-pass filter. The smoothing operation enhances the text relative to any background texture. This is because background texture normally has higher frequency than text does. The smoothing operation also removes speckle noise. In the second step, the intensity histogram of the smoothed image is computed and a threshold automatically selected as follows. For black text, the first peak of the histogram corresponds to text. Thresholding the image at the value of the valley between the first and second peaks of the histogram binarizes the image well. In order to reliably identify the valley, the histogram is smoothed by a low-pass filter before the threshold is computed. The algorithm has been applied to some 50 images from a wide variety of source: digitized video frames, photos, newspapers, advertisements in magazines or sales flyers, personal checks, etc. There are 21820 characters and 4406 words in these images. 91 percent of the characters and 86 percent of the words are successfully cleaned up and binarized. A commercial OCR was applied to the binarized text when it consisted of fonts which were OCR recognizable. The recognition rate was 84 percent for the characters and 77 percent for the words.
Image simulation for automatic license plate recognition
NASA Astrophysics Data System (ADS)
Bala, Raja; Zhao, Yonghui; Burry, Aaron; Kozitsky, Vladimir; Fillion, Claude; Saunders, Craig; Rodríguez-Serrano, José
2012-01-01
Automatic license plate recognition (ALPR) is an important capability for traffic surveillance applications, including toll monitoring and detection of different types of traffic violations. ALPR is a multi-stage process comprising plate localization, character segmentation, optical character recognition (OCR), and identification of originating jurisdiction (i.e. state or province). Training of an ALPR system for a new jurisdiction typically involves gathering vast amounts of license plate images and associated ground truth data, followed by iterative tuning and optimization of the ALPR algorithms. The substantial time and effort required to train and optimize the ALPR system can result in excessive operational cost and overhead. In this paper we propose a framework to create an artificial set of license plate images for accelerated training and optimization of ALPR algorithms. The framework comprises two steps: the synthesis of license plate images according to the design and layout for a jurisdiction of interest; and the modeling of imaging transformations and distortions typically encountered in the image capture process. Distortion parameters are estimated by measurements of real plate images. The simulation methodology is successfully demonstrated for training of OCR.
Improving the Accessibility of Mobile OCR Apps Via Interactive Modalities.
Cutter, Michael; Manduchi, Roberto
2017-10-01
Mobile optical character recognition (OCR) apps have come of age. Many blind individuals use them on a daily basis. The usability of such tools, however, is limited by the requirement that a good picture of the text to be read must be taken, something that is difficult to do without sight. Some mobile OCR apps already implement auto-shot and guidance mechanisms to facilitate this task. In this paper, we describe two experiments with blind participants, who tested these two interactive mechanisms on a customized iPhone implementation. These experiments bring to light a number of interesting aspects of accessing a printed document without sight, and enable a comparative analysis of the available interaction modalities.
Improving the Accessibility of Mobile OCR Apps Via Interactive Modalities
Cutter, Michael; Manduchi, Roberto
2017-01-01
Mobile optical character recognition (OCR) apps have come of age. Many blind individuals use them on a daily basis. The usability of such tools, however, is limited by the requirement that a good picture of the text to be read must be taken, something that is difficult to do without sight. Some mobile OCR apps already implement auto-shot and guidance mechanisms to facilitate this task. In this paper, we describe two experiments with blind participants, who tested these two interactive mechanisms on a customized iPhone implementation. These experiments bring to light a number of interesting aspects of accessing a printed document without sight, and enable a comparative analysis of the available interaction modalities. PMID:29270243
Multi-font printed Mongolian document recognition system
NASA Astrophysics Data System (ADS)
Peng, Liangrui; Liu, Changsong; Ding, Xiaoqing; Wang, Hua; Jin, Jianming
2009-01-01
Mongolian is one of the major ethnic languages in China. Large amount of Mongolian printed documents need to be digitized in digital library and various applications. Traditional Mongolian script has unique writing style and multi-font-type variations, which bring challenges to Mongolian OCR research. As traditional Mongolian script has some characteristics, for example, one character may be part of another character, we define the character set for recognition according to the segmented components, and the components are combined into characters by rule-based post-processing module. For character recognition, a method based on visual directional feature and multi-level classifiers is presented. For character segmentation, a scheme is used to find the segmentation point by analyzing the properties of projection and connected components. As Mongolian has different font-types which are categorized into two major groups, the parameter of segmentation is adjusted for each group. A font-type classification method for the two font-type group is introduced. For recognition of Mongolian text mixed with Chinese and English, language identification and relevant character recognition kernels are integrated. Experiments show that the presented methods are effective. The text recognition rate is 96.9% on the test samples from practical documents with multi-font-types and mixed scripts.
Applications of Optical Scanners in an Academic Center.
ERIC Educational Resources Information Center
Molinari, Carol; Tannenbaum, Robert S.
1995-01-01
Describes optical scanners, including how the technology works; applications in data management and research; development of instructional materials; and providing community services. Discussion includes the three basic types of optical scanners: optical character recognition (OCR), optical mark readers (OMR), and graphic scanners. A sidebar…
Counting OCR errors in typeset text
NASA Astrophysics Data System (ADS)
Sandberg, Jonathan S.
1995-03-01
Frequently object recognition accuracy is a key component in the performance analysis of pattern matching systems. In the past three years, the results of numerous excellent and rigorous studies of OCR system typeset-character accuracy (henceforth OCR accuracy) have been published, encouraging performance comparisons between a variety of OCR products and technologies. These published figures are important; OCR vendor advertisements in the popular trade magazines lead readers to believe that published OCR accuracy figures effect market share in the lucrative OCR market. Curiously, a detailed review of many of these OCR error occurrence counting results reveals that they are not reproducible as published and they are not strictly comparable due to larger variances in the counts than would be expected by the sampling variance. Naturally, since OCR accuracy is based on a ratio of the number of OCR errors over the size of the text searched for errors, imprecise OCR error accounting leads to similar imprecision in OCR accuracy. Some published papers use informal, non-automatic, or intuitively correct OCR error accounting. Still other published results present OCR error accounting methods based on string matching algorithms such as dynamic programming using Levenshtein (edit) distance but omit critical implementation details (such as the existence of suspect markers in the OCR generated output or the weights used in the dynamic programming minimization procedure). The problem with not specifically revealing the accounting method is that the number of errors found by different methods are significantly different. This paper identifies the basic accounting methods used to measure OCR errors in typeset text and offers an evaluation and comparison of the various accounting methods.
Data Input for Libraries: State-of-the-Art Report.
ERIC Educational Resources Information Center
Buckland, Lawrence F.
This brief overview of new manuscript preparation methods which allow authors and editors to set their own type discusses the advantages and disadvantages of optical character recognition (OCR), microcomputers and personal computers, minicomputers, and word processors for editing and database entry. Potential library applications are also…
Intelligent Classification in Huge Heterogeneous Data Sets
2015-06-01
Competencies DoD Department of Defense GMTI Ground Moving Target Indicator ISR Intelligence, Surveillance and Reconnaissance NCD Noncoherent Change...Detection OCR Optical Character Recognition PCA Principal Component Analysis SAR Synthetic Aperture Radar SVD Singular Value Decomponsition USPS United States Postal Service 8 Approved for Public Release; Distribution Unlimited.
Figure Text Extraction in Biomedical Literature
Kim, Daehyun; Yu, Hong
2011-01-01
Background Figures are ubiquitous in biomedical full-text articles, and they represent important biomedical knowledge. However, the sheer volume of biomedical publications has made it necessary to develop computational approaches for accessing figures. Therefore, we are developing the Biomedical Figure Search engine (http://figuresearch.askHERMES.org) to allow bioscientists to access figures efficiently. Since text frequently appears in figures, automatically extracting such text may assist the task of mining information from figures. Little research, however, has been conducted exploring text extraction from biomedical figures. Methodology We first evaluated an off-the-shelf Optical Character Recognition (OCR) tool on its ability to extract text from figures appearing in biomedical full-text articles. We then developed a Figure Text Extraction Tool (FigTExT) to improve the performance of the OCR tool for figure text extraction through the use of three innovative components: image preprocessing, character recognition, and text correction. We first developed image preprocessing to enhance image quality and to improve text localization. Then we adapted the off-the-shelf OCR tool on the improved text localization for character recognition. Finally, we developed and evaluated a novel text correction framework by taking advantage of figure-specific lexicons. Results/Conclusions The evaluation on 382 figures (9,643 figure texts in total) randomly selected from PubMed Central full-text articles shows that FigTExT performed with 84% precision, 98% recall, and 90% F1-score for text localization and with 62.5% precision, 51.0% recall and 56.2% F1-score for figure text extraction. When limiting figure texts to those judged by domain experts to be important content, FigTExT performed with 87.3% precision, 68.8% recall, and 77% F1-score. FigTExT significantly improved the performance of the off-the-shelf OCR tool we used, which on its own performed with 36.6% precision, 19.3% recall, and 25.3% F1-score for text extraction. In addition, our results show that FigTExT can extract texts that do not appear in figure captions or other associated text, further suggesting the potential utility of FigTExT for improving figure search. PMID:21249186
Figure text extraction in biomedical literature.
Kim, Daehyun; Yu, Hong
2011-01-13
Figures are ubiquitous in biomedical full-text articles, and they represent important biomedical knowledge. However, the sheer volume of biomedical publications has made it necessary to develop computational approaches for accessing figures. Therefore, we are developing the Biomedical Figure Search engine (http://figuresearch.askHERMES.org) to allow bioscientists to access figures efficiently. Since text frequently appears in figures, automatically extracting such text may assist the task of mining information from figures. Little research, however, has been conducted exploring text extraction from biomedical figures. We first evaluated an off-the-shelf Optical Character Recognition (OCR) tool on its ability to extract text from figures appearing in biomedical full-text articles. We then developed a Figure Text Extraction Tool (FigTExT) to improve the performance of the OCR tool for figure text extraction through the use of three innovative components: image preprocessing, character recognition, and text correction. We first developed image preprocessing to enhance image quality and to improve text localization. Then we adapted the off-the-shelf OCR tool on the improved text localization for character recognition. Finally, we developed and evaluated a novel text correction framework by taking advantage of figure-specific lexicons. The evaluation on 382 figures (9,643 figure texts in total) randomly selected from PubMed Central full-text articles shows that FigTExT performed with 84% precision, 98% recall, and 90% F1-score for text localization and with 62.5% precision, 51.0% recall and 56.2% F1-score for figure text extraction. When limiting figure texts to those judged by domain experts to be important content, FigTExT performed with 87.3% precision, 68.8% recall, and 77% F1-score. FigTExT significantly improved the performance of the off-the-shelf OCR tool we used, which on its own performed with 36.6% precision, 19.3% recall, and 25.3% F1-score for text extraction. In addition, our results show that FigTExT can extract texts that do not appear in figure captions or other associated text, further suggesting the potential utility of FigTExT for improving figure search.
Expert system for automatically correcting OCR output
NASA Astrophysics Data System (ADS)
Taghva, Kazem; Borsack, Julie; Condit, Allen
1994-03-01
This paper describes a new expert system for automatically correcting errors made by optical character recognition (OCR) devices. The system, which we call the post-processing system, is designed to improve the quality of text produced by an OCR device in preparation for subsequent retrieval from an information system. The system is composed of numerous parts: an information retrieval system, an English dictionary, a domain-specific dictionary, and a collection of algorithms and heuristics designed to correct as many OCR errors as possible. For the remaining errors that cannot be corrected, the system passes them on to a user-level editing program. This post-processing system can be viewed as part of a larger system that would streamline the steps of taking a document from its hard copy form to its usable electronic form, or it can be considered a stand alone system for OCR error correction. An earlier version of this system has been used to process approximately 10,000 pages of OCR generated text. Among the OCR errors discovered by this version, about 87% were corrected. We implement numerous new parts of the system, test this new version, and present the results.
A robust omnifont open-vocabulary Arabic OCR system using pseudo-2D-HMM
NASA Astrophysics Data System (ADS)
Rashwan, Abdullah M.; Rashwan, Mohsen A.; Abdel-Hameed, Ahmed; Abdou, Sherif; Khalil, A. H.
2012-01-01
Recognizing old documents is highly desirable since the demand for quickly searching millions of archived documents has recently increased. Using Hidden Markov Models (HMMs) has been proven to be a good solution to tackle the main problems of recognizing typewritten Arabic characters. These attempts however achieved a remarkable success for omnifont OCR under very favorable conditions, they didn't achieve the same performance in practical conditions, i.e. noisy documents. In this paper we present an omnifont, large-vocabulary Arabic OCR system using Pseudo Two Dimensional Hidden Markov Model (P2DHMM), which is a generalization of the HMM. P2DHMM offers a more efficient way to model the Arabic characters, such model offer both minimal dependency on the font size/style (omnifont), and high level of robustness against noise. The evaluation results of this system are very promising compared to a baseline HMM system and best OCRs available in the market (Sakhr and NovoDynamics). The recognition accuracy of the P2DHMM classifier is measured against the classic HMM classifier, the average word accuracy rates for P2DHMM and HMM classifiers are 79% and 66% respectively. The overall system accuracy is measured against Sakhr and NovoDynamics OCR systems, the average word accuracy rates for P2DHMM, NovoDynamics, and Sakhr are 74%, 71%, and 61% respectively.
TELLTALE: Experiments in a Dynamic Hypertext Environment for Degraded and Multilingual Data.
ERIC Educational Resources Information Center
Pearce, Claudia; Nicholas, Charles
1996-01-01
Presents experimentation results for the TELLTALE system, a dynamic hypertext environment that provides full-text search from a hypertext-style user interface for text corpora that may be garbled by OCR (optical character recognition) or transmission errors, and that may contain languages other than English. (Author/LRW)
Techniques of Document Management: A Review of Text Retrieval and Related Technologies.
ERIC Educational Resources Information Center
Veal, D. C.
2001-01-01
Reviews present and possible future developments in the techniques of electronic document management, the major ones being text retrieval and scanning and OCR (optical character recognition). Also addresses document acquisition, indexing and thesauri, publishing and dissemination standards, impact of the Internet, and the document management…
Fu, H C; Xu, Y Y; Chang, H Y
1999-12-01
Recognition of similar (confusion) characters is a difficult problem in optical character recognition (OCR). In this paper, we introduce a neural network solution that is capable of modeling minor differences among similar characters, and is robust to various personal handwriting styles. The Self-growing Probabilistic Decision-based Neural Network (SPDNN) is a probabilistic type neural network, which adopts a hierarchical network structure with nonlinear basis functions and a competitive credit-assignment scheme. Based on the SPDNN model, we have constructed a three-stage recognition system. First, a coarse classifier determines a character to be input to one of the pre-defined subclasses partitioned from a large character set, such as Chinese mixed with alphanumerics. Then a character recognizer determines the input image which best matches the reference character in the subclass. Lastly, the third module is a similar character recognizer, which can further enhance the recognition accuracy among similar or confusing characters. The prototype system has demonstrated a successful application of SPDNN to similar handwritten Chinese recognition for the public database CCL/HCCR1 (5401 characters x200 samples). Regarding performance, experiments on the CCL/HCCR1 database produced 90.12% recognition accuracy with no rejection, and 94.11% accuracy with 6.7% rejection, respectively. This recognition accuracy represents about 4% improvement on the previously announced performance. As to processing speed, processing before recognition (including image preprocessing, segmentation, and feature extraction) requires about one second for an A4 size character image, and recognition consumes approximately 0.27 second per character on a Pentium-100 based personal computer, without use of any hardware accelerator or co-processor.
A segmentation-free approach to Arabic and Urdu OCR
NASA Astrophysics Data System (ADS)
Sabbour, Nazly; Shafait, Faisal
2013-01-01
In this paper, we present a generic Optical Character Recognition system for Arabic script languages called Nabocr. Nabocr uses OCR approaches specific for Arabic script recognition. Performing recognition on Arabic script text is relatively more difficult than Latin text due to the nature of Arabic script, which is cursive and context sensitive. Moreover, Arabic script has different writing styles that vary in complexity. Nabocr is initially trained to recognize both Urdu Nastaleeq and Arabic Naskh fonts. However, it can be trained by users to be used for other Arabic script languages. We have evaluated our system's performance for both Urdu and Arabic. In order to evaluate Urdu recognition, we have generated a dataset of Urdu text called UPTI (Urdu Printed Text Image Database), which measures different aspects of a recognition system. The performance of our system for Urdu clean text is 91%. For Arabic clean text, the performance is 86%. Moreover, we have compared the performance of our system against Tesseract's newly released Arabic recognition, and the performance of both systems on clean images is almost the same.
NASA Astrophysics Data System (ADS)
Hassibi, Khosrow M.
1994-02-01
This paper presents a brief overview of our research in the development of an OCR system for recognition of machine-printed texts in languages that use the Arabic alphabet. The cursive nature of machine-printed Arabic makes the segmentation of words into letters a challenging problem. In our approach, through a novel preliminary segmentation technique, a word is broken into pieces where each piece may not represent a valid letter in general. Neural networks trained on a training sample set of about 500 Arabic text images are used for recognition of these pieces. The rules governing the alphabet and character-level contextual information are used for recombining these pieces into valid letters. Higher-level contextual analysis schemes including the use of an Arabic lexicon and n-grams is also under development and are expected to improve the word recognition accuracy. The segmentation, recognition, and contextual analysis processes are closely integrated using a feedback scheme. The details of preparation of the training set and some recent results on training of the networks will be presented.
intelligentCAPTURE 1.0 Adds Tables of Content to Library Catalogues and Improves Retrieval.
ERIC Educational Resources Information Center
Hauer, Manfred; Simedy, Walton
2002-01-01
Describes an online library catalog that was developed for an Austrian scientific library that includes table of contents in addition to the standard bibliographic information in order to increase relevance for searchers. Discusses the technology involved, including OCR (Optical Character Recognition) and automatic indexing techniques; weighted…
EDP Applications to Musical Bibliography: Input Considerations
ERIC Educational Resources Information Center
Robbins, Donald C.
1972-01-01
The application of Electronic Data Processing (EDP) has been a boon in the analysis and bibliographic control of music. However, an extra step of encoding must be undertaken for input of music. The best hope to facilitate musical input is the development of an Optical Character Recognition (OCR) music-reading machine. (29 references) (Author/NH)
Combining multiple thresholding binarization values to improve OCR output
NASA Astrophysics Data System (ADS)
Lund, William B.; Kennard, Douglas J.; Ringger, Eric K.
2013-01-01
For noisy, historical documents, a high optical character recognition (OCR) word error rate (WER) can render the OCR text unusable. Since image binarization is often the method used to identify foreground pixels, a body of research seeks to improve image-wide binarization directly. Instead of relying on any one imperfect binarization technique, our method incorporates information from multiple simple thresholding binarizations of the same image to improve text output. Using a new corpus of 19th century newspaper grayscale images for which the text transcription is known, we observe WERs of 13.8% and higher using current binarization techniques and a state-of-the-art OCR engine. Our novel approach combines the OCR outputs from multiple thresholded images by aligning the text output and producing a lattice of word alternatives from which a lattice word error rate (LWER) is calculated. Our results show a LWER of 7.6% when aligning two threshold images and a LWER of 6.8% when aligning five. From the word lattice we commit to one hypothesis by applying the methods of Lund et al. (2011) achieving an improvement over the original OCR output and a 8.41% WER result on this data set.
Hawker, Charles D; McCarthy, William; Cleveland, David; Messinger, Bonnie L
2014-03-01
Mislabeled samples are a serious problem in most clinical laboratories. Published error rates range from 0.39/1000 to as high as 1.12%. Standardization of bar codes and label formats has not yet achieved the needed improvement. The mislabel rate in our laboratory, although low compared with published rates, prompted us to seek a solution to achieve zero errors. To reduce or eliminate our mislabeled samples, we invented an automated device using 4 cameras to photograph the outside of a sample tube. The system uses optical character recognition (OCR) to look for discrepancies between the patient name in our laboratory information system (LIS) vs the patient name on the customer label. All discrepancies detected by the system's software then require human inspection. The system was installed on our automated track and validated with production samples. We obtained 1 009 830 images during the validation period, and every image was reviewed. OCR passed approximately 75% of the samples, and no mislabeled samples were passed. The 25% failed by the system included 121 samples actually mislabeled by patient name and 148 samples with spelling discrepancies between the patient name on the customer label and the patient name in our LIS. Only 71 of the 121 mislabeled samples detected by OCR were found through our normal quality assurance process. We have invented an automated camera system that uses OCR technology to identify potential mislabeled samples. We have validated this system using samples transported on our automated track. Full implementation of this technology offers the possibility of zero mislabeled samples in the preanalytic stage.
Recognition and defect detection of dot-matrix text via variation-model based learning
NASA Astrophysics Data System (ADS)
Ohyama, Wataru; Suzuki, Koushi; Wakabayashi, Tetsushi
2017-03-01
An algorithm for recognition and defect detection of dot-matrix text printed on products is proposed. Extraction and recognition of dot-matrix text contains several difficulties, which are not involved in standard camera-based OCR, that the appearance of dot-matrix characters is corrupted and broken by illumination, complex texture in the background and other standard characters printed on product packages. We propose a dot-matrix text extraction and recognition method which does not require any user interaction. The method employs detected location of corner points and classification score. The result of evaluation experiment using 250 images shows that recall and precision of extraction are 78.60% and 76.03%, respectively. Recognition accuracy of correctly extracted characters is 94.43%. Detecting printing defect of dot-matrix text is also important in the production scene to avoid illegal productions. We also propose a detection method for printing defect of dot-matrix characters. The method constructs a feature vector of which elements are classification scores of each character class and employs support vector machine to classify four types of printing defect. The detection accuracy of the proposed method is 96.68 %.
Recognizing characters of ancient manuscripts
NASA Astrophysics Data System (ADS)
Diem, Markus; Sablatnig, Robert
2010-02-01
Considering printed Latin text, the main issues of Optical Character Recognition (OCR) systems are solved. However, for degraded handwritten document images, basic preprocessing steps such as binarization, gain poor results with state-of-the-art methods. In this paper ancient Slavonic manuscripts from the 11th century are investigated. In order to minimize the consequences of false character segmentation, a binarization-free approach based on local descriptors is proposed. Additionally local information allows the recognition of partially visible or washed out characters. The proposed algorithm consists of two steps: character classification and character localization. Initially Scale Invariant Feature Transform (SIFT) features are extracted which are subsequently classified using Support Vector Machines (SVM). Afterwards, the interest points are clustered according to their spatial information. Thereby, characters are localized and finally recognized based on a weighted voting scheme of pre-classified local descriptors. Preliminary results show that the proposed system can handle highly degraded manuscript images with background clutter (e.g. stains, tears) and faded out characters.
Development of Portable Automatic Number Plate Recognition System on Android Mobile Phone
NASA Astrophysics Data System (ADS)
Mutholib, Abdul; Gunawan, Teddy S.; Chebil, Jalel; Kartiwi, Mira
2013-12-01
The Automatic Number Plate Recognition (ANPR) System has performed as the main role in various access control and security, such as: tracking of stolen vehicles, traffic violations (speed trap) and parking management system. In this paper, the portable ANPR implemented on android mobile phone is presented. The main challenges in mobile application are including higher coding efficiency, reduced computational complexity, and improved flexibility. Significance efforts are being explored to find suitable and adaptive algorithm for implementation of ANPR on mobile phone. ANPR system for mobile phone need to be optimize due to its limited CPU and memory resources, its ability for geo-tagging image captured using GPS coordinates and its ability to access online database to store the vehicle's information. In this paper, the design of portable ANPR on android mobile phone will be described as follows. First, the graphical user interface (GUI) for capturing image using built-in camera was developed to acquire vehicle plate number in Malaysia. Second, the preprocessing of raw image was done using contrast enhancement. Next, character segmentation using fixed pitch and an optical character recognition (OCR) using neural network were utilized to extract texts and numbers. Both character segmentation and OCR were using Tesseract library from Google Inc. The proposed portable ANPR algorithm was implemented and simulated using Android SDK on a computer. Based on the experimental results, the proposed system can effectively recognize the license plate number at 90.86%. The required processing time to recognize a license plate is only 2 seconds on average. The result is consider good in comparison with the results obtained from previous system that was processed in a desktop PC with the range of result from 91.59% to 98% recognition rate and 0.284 second to 1.5 seconds recognition time.
A neural network based artificial vision system for licence plate recognition.
Draghici, S
1997-02-01
This paper presents a neural network based artificial vision system able to analyze the image of a car given by a camera, locate the registration plate and recognize the registration number of the car. The paper describes in detail various practical problems encountered in implementing this particular application and the solutions used to solve them. The main features of the system presented are: controlled stability-plasticity behavior, controlled reliability threshold, both off-line and on-line learning, self assessment of the output reliability and high reliability based on high level multiple feedback. The system has been designed using a modular approach. Sub-modules can be upgraded and/or substituted independently, thus making the system potentially suitable in a large variety of vision applications. The OCR engine was designed as an interchangeable plug-in module. This allows the user to choose an OCR engine which is suited to the particular application and to upgrade it easily in the future. At present, there are several versions of this OCR engine. One of them is based on a fully connected feedforward artificial neural network with sigmoidal activation functions. This network can be trained with various training algorithms such as error backpropagation. An alternative OCR engine is based on the constraint based decomposition (CBD) training architecture. The system has showed the following performances (on average) on real-world data: successful plate location and segmentation about 99%, successful character recognition about 98% and successful recognition of complete registration plates about 80%.
ERIC Educational Resources Information Center
Galloway, Edward A.; Michalek, Gabrielle V.
1995-01-01
Discusses the conversion project of the congressional papers of Senator John Heinz into digital format and the provision of electronic access to these papers by Carnegie Mellon University. Topics include collection background, project team structure, document processing, scanning, use of optical character recognition software, verification…
Document image retrieval through word shape coding.
Lu, Shijian; Li, Linlin; Tan, Chew Lim
2008-11-01
This paper presents a document retrieval technique that is capable of searching document images without OCR (optical character recognition). The proposed technique retrieves document images by a new word shape coding scheme, which captures the document content through annotating each word image by a word shape code. In particular, we annotate word images by using a set of topological shape features including character ascenders/descenders, character holes, and character water reservoirs. With the annotated word shape codes, document images can be retrieved by either query keywords or a query document image. Experimental results show that the proposed document image retrieval technique is fast, efficient, and tolerant to various types of document degradation.
NASA Astrophysics Data System (ADS)
Nagy, George
2008-01-01
The fifteenth anniversary of the first SPIE symposium (titled Character Recognition Technologies) on Document Recognition and Retrieval provides an opportunity to examine DRR's contributions to the development of document technologies. Many of the tools taken for granted today, including workable general purpose OCR, large-scale, semi-automatic forms processing, inter-format table conversion, and text mining, followed research presented at this venue. This occasion also affords an opportunity to offer tribute to the conference organizers and proceedings editors and to the coterie of professionals who regularly participate in DRR.
Development of a Digitalized Child's Checkups Information System.
Ito, Yoshiya; Takimoto, Hidemi
2017-01-01
In Japan, health checkups for children take place from infancy through high school and play an important role in the maintenance and control of childhood/adolescent health. The anthropometric data obtained during these checkups are kept in health centers and schools and are also recorded in a mother's maternal and child health handbook, as well as on school health cards. These data are meaningful if they are utilized well and in an appropriate manner. They are particularly useful for the prevention of obesity-related conditions in adulthood, such as metabolic syndrome and diabetes mellitus. For this purpose, we have tried to establish a scanning system with an optical character recognition (OCR) function, which links data obtained during health checkups in infancy with that obtained in schools. In this system, handwritten characters on the records are scanned and processed using OCR. However, because many of the scanned characters are not read properly, we must wait for the improvement in the performance of the OCR function. In addition, we have developed Microsoft Excel spreadsheets, on which obesity-related indices, such as body mass index and relative body weight, are calculated. These sheets also provide functions that tabulate the frequencies of obesity in specific groups. Actively using these data and digitalized systems will not only contribute towards resolving physical health problems in children, but also decrease the risk of developing lifestyle-related diseases in adulthood.
Font adaptive word indexing of modern printed documents.
Marinai, Simone; Marino, Emanuele; Soda, Giovanni
2006-08-01
We propose an approach for the word-level indexing of modern printed documents which are difficult to recognize using current OCR engines. By means of word-level indexing, it is possible to retrieve the position of words in a document, enabling queries involving proximity of terms. Web search engines implement this kind of indexing, allowing users to retrieve Web pages on the basis of their textual content. Nowadays, digital libraries hold collections of digitized documents that can be retrieved either by browsing the document images or relying on appropriate metadata assembled by domain experts. Word indexing tools would therefore increase the access to these collections. The proposed system is designed to index homogeneous document collections by automatically adapting to different languages and font styles without relying on OCR engines for character recognition. The approach is based on three main ideas: the use of Self Organizing Maps (SOM) to perform unsupervised character clustering, the definition of one suitable vector-based word representation whose size depends on the word aspect-ratio, and the run-time alignment of the query word with indexed words to deal with broken and touching characters. The most appropriate applications are for processing modern printed documents (17th to 19th centuries) where current OCR engines are less accurate. Our experimental analysis addresses six data sets containing documents ranging from books of the 17th century to contemporary journals.
Image based book cover recognition and retrieval
NASA Astrophysics Data System (ADS)
Sukhadan, Kalyani; Vijayarajan, V.; Krishnamoorthi, A.; Bessie Amali, D. Geraldine
2017-11-01
In this we are developing a graphical user interface using MATLAB for the users to check the information related to books in real time. We are taking the photos of the book cover using GUI, then by using MSER algorithm it will automatically detect all the features from the input image, after this it will filter bifurcate non-text features which will be based on morphological difference between text and non-text regions. We implemented a text character alignment algorithm which will improve the accuracy of the original text detection. We will also have a look upon the built in MATLAB OCR recognition algorithm and an open source OCR which is commonly used to perform better detection results, post detection algorithm is implemented and natural language processing to perform word correction and false detection inhibition. Finally, the detection result will be linked to internet to perform online matching. More than 86% accuracy can be obtained by this algorithm.
Comparison of approaches for mobile document image analysis using server supported smartphones
NASA Astrophysics Data System (ADS)
Ozarslan, Suleyman; Eren, P. Erhan
2014-03-01
With the recent advances in mobile technologies, new capabilities are emerging, such as mobile document image analysis. However, mobile phones are still less powerful than servers, and they have some resource limitations. One approach to overcome these limitations is performing resource-intensive processes of the application on remote servers. In mobile document image analysis, the most resource consuming process is the Optical Character Recognition (OCR) process, which is used to extract text in mobile phone captured images. In this study, our goal is to compare the in-phone and the remote server processing approaches for mobile document image analysis in order to explore their trade-offs. For the inphone approach, all processes required for mobile document image analysis run on the mobile phone. On the other hand, in the remote-server approach, core OCR process runs on the remote server and other processes run on the mobile phone. Results of the experiments show that the remote server approach is considerably faster than the in-phone approach in terms of OCR time, but adds extra delays such as network delay. Since compression and downscaling of images significantly reduce file sizes and extra delays, the remote server approach overall outperforms the in-phone approach in terms of selected speed and correct recognition metrics, if the gain in OCR time compensates for the extra delays. According to the results of the experiments, using the most preferable settings, the remote server approach performs better than the in-phone approach in terms of speed and acceptable correct recognition metrics.
Grouin, Cyril; Zweigenbaum, Pierre
2013-01-01
In this paper, we present a comparison of two approaches to automatically de-identify medical records written in French: a rule-based system and a machine-learning based system using a conditional random fields (CRF) formalism. Both systems have been designed to process nine identifiers in a corpus of medical records in cardiology. We performed two evaluations: first, on 62 documents in cardiology, and on 10 documents in foetopathology - produced by optical character recognition (OCR) - to evaluate the robustness of our systems. We achieved a 0.843 (rule-based) and 0.883 (machine-learning) exact match overall F-measure in cardiology. While the rule-based system allowed us to achieve good results on nominative (first and last names) and numerical data (dates, phone numbers, and zip codes), the machine-learning approach performed best on more complex categories (postal addresses, hospital names, medical devices, and towns). On the foetopathology corpus, although our systems have not been designed for this corpus and despite OCR character recognition errors, we obtained promising results: a 0.681 (rule-based) and 0.638 (machine-learning) exact-match overall F-measure. This demonstrates that existing tools can be applied to process new documents of lower quality.
Beta Testing a Novel Smartphone Application to Improve Medication Adherence.
Sarzynski, Erin; Decker, Brian; Thul, Aaron; Weismantel, David; Melaragni, Ronald; Cholakis, Elizabeth; Tewari, Megha; Beckholt, Kristy; Zaroukian, Michael; Kennedy, Angie C; Given, Charles
2017-04-01
We developed and beta-tested a patient-centered medication management application, PresRx optical character recognition (OCR), a mobile health (m-health) tool that auto-populates drug name and dosing instructions directly from patients' medication labels by OCR. We employed a single-subject design study to evaluate PresRx OCR for three outcomes: (1) accuracy of auto-populated medication dosing instructions, (2) acceptability of the user interface, and (3) patients' adherence to chronic medications. Eight patients beta-tested PresRx OCR. Five patients used the software for ≥6 months, and four completed exit interviews (n = 4 completers). At baseline, patients used 3.4 chronic prescription medications and exhibited moderate-to-high adherence rates. Accuracy of auto-populated information by OCR was 95% for drug name, 98% for dose, and 96% for frequency. Study completers rated PresRx OCR 74 on the System Usability Scale, where scores ≥70 indicate an acceptable user interface (scale 0-100). Adherence rates measured by PresRx OCR were high during the first month of app use (93%), but waned midway through the 6-month testing period (78%). Compared with pharmacy fill rates, PresRx OCR underestimated adherence among completers by 3%, while it overestimated adherence among noncompleters by 8%. Results suggest smartphone applications supporting medication management are feasible and accurately assess adherence compared with objective measures. Future efforts to improve medication-taking behavior using m-health tools should target specific patient populations and leverage common application programming interfaces to promote generalizability. Our medication management application PresRx OCR is innovative, acceptable for patient use, and accurately tracks medication adherence.
Text recognition and correction for automated data collection by mobile devices
NASA Astrophysics Data System (ADS)
Ozarslan, Suleyman; Eren, P. Erhan
2014-03-01
Participatory sensing is an approach which allows mobile devices such as mobile phones to be used for data collection, analysis and sharing processes by individuals. Data collection is the first and most important part of a participatory sensing system, but it is time consuming for the participants. In this paper, we discuss automatic data collection approaches for reducing the time required for collection, and increasing the amount of collected data. In this context, we explore automated text recognition on images of store receipts which are captured by mobile phone cameras, and the correction of the recognized text. Accordingly, our first goal is to evaluate the performance of the Optical Character Recognition (OCR) method with respect to data collection from store receipt images. Images captured by mobile phones exhibit some typical problems, and common image processing methods cannot handle some of them. Consequently, the second goal is to address these types of problems through our proposed Knowledge Based Correction (KBC) method used in support of the OCR, and also to evaluate the KBC method with respect to the improvement on the accurate recognition rate. Results of the experiments show that the KBC method improves the accurate data recognition rate noticeably.
Partitioning of the degradation space for OCR training
NASA Astrophysics Data System (ADS)
Barney Smith, Elisa H.; Andersen, Tim
2006-01-01
Generally speaking optical character recognition algorithms tend to perform better when presented with homogeneous data. This paper studies a method that is designed to increase the homogeneity of training data, based on an understanding of the types of degradations that occur during the printing and scanning process, and how these degradations affect the homogeneity of the data. While it has been shown that dividing the degradation space by edge spread improves recognition accuracy over dividing the degradation space by threshold or point spread function width alone, the challenge is in deciding how many partitions and at what value of edge spread the divisions should be made. Clustering of different types of character features, fonts, sizes, resolutions and noise levels shows that edge spread is indeed shown to be a strong indicator of the homogeneity of character data clusters.
Intelligent bar chart plagiarism detection in documents.
Al-Dabbagh, Mohammed Mumtaz; Salim, Naomie; Rehman, Amjad; Alkawaz, Mohammed Hazim; Saba, Tanzila; Al-Rodhaan, Mznah; Al-Dhelaan, Abdullah
2014-01-01
This paper presents a novel features mining approach from documents that could not be mined via optical character recognition (OCR). By identifying the intimate relationship between the text and graphical components, the proposed technique pulls out the Start, End, and Exact values for each bar. Furthermore, the word 2-gram and Euclidean distance methods are used to accurately detect and determine plagiarism in bar charts.
Intelligent Bar Chart Plagiarism Detection in Documents
Al-Dabbagh, Mohammed Mumtaz; Salim, Naomie; Alkawaz, Mohammed Hazim; Saba, Tanzila; Al-Rodhaan, Mznah; Al-Dhelaan, Abdullah
2014-01-01
This paper presents a novel features mining approach from documents that could not be mined via optical character recognition (OCR). By identifying the intimate relationship between the text and graphical components, the proposed technique pulls out the Start, End, and Exact values for each bar. Furthermore, the word 2-gram and Euclidean distance methods are used to accurately detect and determine plagiarism in bar charts. PMID:25309952
Automatic detection and recognition of signs from natural scenes.
Chen, Xilin; Yang, Jie; Zhang, Jing; Waibel, Alex
2004-01-01
In this paper, we present an approach to automatic detection and recognition of signs from natural scenes, and its application to a sign translation task. The proposed approach embeds multiresolution and multiscale edge detection, adaptive searching, color analysis, and affine rectification in a hierarchical framework for sign detection, with different emphases at each phase to handle the text in different sizes, orientations, color distributions and backgrounds. We use affine rectification to recover deformation of the text regions caused by an inappropriate camera view angle. The procedure can significantly improve text detection rate and optical character recognition (OCR) accuracy. Instead of using binary information for OCR, we extract features from an intensity image directly. We propose a local intensity normalization method to effectively handle lighting variations, followed by a Gabor transform to obtain local features, and finally a linear discriminant analysis (LDA) method for feature selection. We have applied the approach in developing a Chinese sign translation system, which can automatically detect and recognize Chinese signs as input from a camera, and translate the recognized text into English.
OCR enhancement through neighbor embedding and fast approximate nearest neighbors
NASA Astrophysics Data System (ADS)
Smith, D. C.
2012-10-01
Generic optical character recognition (OCR) engines often perform very poorly in transcribing scanned low resolution (LR) text documents. To improve OCR performance, we apply the Neighbor Embedding (NE) single-image super-resolution (SISR) technique to LR scanned text documents to obtain high resolution (HR) versions, which we subsequently process with OCR. For comparison, we repeat this procedure using bicubic interpolation (BI). We demonstrate that mean-square errors (MSE) in NE HR estimates do not increase substantially when NE is trained in one Latin font style and tested in another, provided both styles belong to the same font category (serif or sans serif). This is very important in practice, since for each font size, the number of training sets required for each category may be reduced from dozens to just one. We also incorporate randomized k-d trees into our NE implementation to perform approximate nearest neighbor search, and obtain a 1000x speed up of our original NE implementation, with negligible MSE degradation. This acceleration also made it practical to combine all of our size-specific NE Latin models into a single Universal Latin Model (ULM). The ULM eliminates the need to determine the unknown font category and size of an input LR text document and match it to an appropriate model, a very challenging task, since the dpi (pixels per inch) of the input LR image is generally unknown. Our experiments show that OCR character error rates (CER) were over 90% when we applied the Tesseract OCR engine to LR text documents (scanned at 75 dpi and 100 dpi) in the 6-10 pt range. By contrast, using k-d trees and the ULM, CER after NE preprocessing averaged less than 7% at 3x (100 dpi LR scanning) and 4x (75 dpi LR scanning) magnification, over an order of magnitude improvement. Moreover, CER after NE preprocessing was more that 6 times lower on average than after BI preprocessing.
Page Recognition: Quantum Leap In Recognition Technology
NASA Astrophysics Data System (ADS)
Miller, Larry
1989-07-01
No milestone has proven as elusive as the always-approaching "year of the LAN," but the "year of the scanner" might claim the silver medal. Desktop scanners have been around almost as long as personal computers. And everyone thinks they are used for obvious desktop-publishing and business tasks like scanning business documents, magazine articles and other pages, and translating those words into files your computer understands. But, until now, the reality fell far short of the promise. Because it's true that scanners deliver an accurate image of the page to your computer, but the software to recognize this text has been woefully disappointing. Old optical-character recognition (OCR) software recognized such a limited range of pages as to be virtually useless to real users. (For example, one OCR vendor specified 12-point Courier font from an IBM Selectric typewriter: the same font in 10-point, or from a Diablo printer, was unrecognizable!) Computer dealers have told me the chasm between OCR expectations and reality is so broad and deep that nine out of ten prospects leave their stores in disgust when they learn the limitations. And this is a very important, very unfortunate gap. Because the promise of recognition -- what people want it to do -- carries with it tremendous improvements in our productivity and ability to get tons of written documents into our computers where we can do real work with it. The good news is that a revolutionary new development effort has led to the new technology of "page recognition," which actually does deliver the promise we've always wanted from OCR. I'm sure every reader appreciates the breakthrough represented by the laser printer and page-makeup software, a combination so powerful it created new reasons for buying a computer. A similar breakthrough is happening right now in page recognition: the Macintosh (and, I must admit, other personal computers) equipped with a moderately priced scanner and OmniPage software (from Caere Corporation) can recognize not only different fonts (omnifont recogniton) but different page (omnipage) formats, as well.
Figure mining for biomedical research.
Rodriguez-Esteban, Raul; Iossifov, Ivan
2009-08-15
Figures from biomedical articles contain valuable information difficult to reach without specialized tools. Currently, there is no search engine that can retrieve specific figure types. This study describes a retrieval method that takes advantage of principles in image understanding, text mining and optical character recognition (OCR) to retrieve figure types defined conceptually. A search engine was developed to retrieve tables and figure types to aid computational and experimental research. http://iossifovlab.cshl.edu/figurome/.
An Image Processing Approach to Linguistic Translation
NASA Astrophysics Data System (ADS)
Kubatur, Shruthi; Sreehari, Suhas; Hegde, Rajeshwari
2011-12-01
The art of translation is as old as written literature. Developments since the Industrial Revolution have influenced the practice of translation, nurturing schools, professional associations, and standard. In this paper, we propose a method of translation of typed Kannada text (taken as an image) into its equivalent English text. The National Instruments (NI) Vision Assistant (version 8.5) has been used for Optical character Recognition (OCR). We developed a new way of transliteration (which we call NIV transliteration) to simplify the training of characters. Also, we build a special type of dictionary for the purpose of translation.
Goal-oriented evaluation of binarization algorithms for historical document images
NASA Astrophysics Data System (ADS)
Obafemi-Ajayi, Tayo; Agam, Gady
2013-01-01
Binarization is of significant importance in document analysis systems. It is an essential first step, prior to further stages such as Optical Character Recognition (OCR), document segmentation, or enhancement of readability of the document after some restoration stages. Hence, proper evaluation of binarization methods to verify their effectiveness is of great value to the document analysis community. In this work, we perform a detailed goal-oriented evaluation of image quality assessment of the 18 binarization methods that participated in the DIBCO 2011 competition using the 16 historical document test images used in the contest. We are interested in the image quality assessment of the outputs generated by the different binarization algorithms as well as the OCR performance, where possible. We compare our evaluation of the algorithms based on human perception of quality to the DIBCO evaluation metrics. The results obtained provide an insight into the effectiveness of these methods with respect to human perception of image quality as well as OCR performance.
Image Segmentation of Historical Handwriting from Palm Leaf Manuscripts
NASA Astrophysics Data System (ADS)
Surinta, Olarik; Chamchong, Rapeeporn
Palm leaf manuscripts were one of the earliest forms of written media and were used in Southeast Asia to store early written knowledge about subjects such as medicine, Buddhist doctrine and astrology. Therefore, historical handwritten palm leaf manuscripts are important for people who like to learn about historical documents, because we can learn more experience from them. This paper presents an image segmentation of historical handwriting from palm leaf manuscripts. The process is composed of three steps: 1) background elimination to separate text and background by Otsu's algorithm 2) line segmentation and 3) character segmentation by histogram of image. The end result is the character's image. The results from this research may be applied to optical character recognition (OCR) in the future.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Katsumi Marukawa; Kazuki Nakashima; Masashi Koga
1994-12-31
This paper presents a paper form processing system with an error correcting function for reading handwritten kanji strings. In the paper form processing system, names and addresses are important key data, and especially this paper takes up an error correcting method for name and address recognition. The method automatically corrects errors of the kanji OCR (Optical Character Reader) with the help of word dictionaries and other knowledge. Moreover, it allows names and addresses to be written in any style. The method consists of word matching {open_quotes}furigana{close_quotes} verification for name strings, and address approval for address strings. For word matching, kanjimore » name candidates are extracted by automaton-type word matching. In {open_quotes}furigana{close_quotes} verification, kana candidate characters recognized by the kana OCR are compared with kana`s searched from the name dictionary based on kanji name candidates, given by the word matching. The correct name is selected from the results of word matching and furigana verification. Also, the address approval efficiently searches for the right address based on a bottom-up procedure which follows hierarchical relations from a lower placename to a upper one by using the positional condition among the placenames. We ascertained that the error correcting method substantially improves the recognition rate and processing speed in experiments on 5,032 forms.« less
Font group identification using reconstructed fonts
NASA Astrophysics Data System (ADS)
Cutter, Michael P.; van Beusekom, Joost; Shafait, Faisal; Breuel, Thomas M.
2011-01-01
Ideally, digital versions of scanned documents should be represented in a format that is searchable, compressed, highly readable, and faithful to the original. These goals can theoretically be achieved through OCR and font recognition, re-typesetting the document text with original fonts. However, OCR and font recognition remain hard problems, and many historical documents use fonts that are not available in digital forms. It is desirable to be able to reconstruct fonts with vector glyphs that approximate the shapes of the letters that form a font. In this work, we address the grouping of tokens in a token-compressed document into candidate fonts. This permits us to incorporate font information into token-compressed images even when the original fonts are unknown or unavailable in digital format. This paper extends previous work in font reconstruction by proposing and evaluating an algorithm to assign a font to every character within a document. This is a necessary step to represent a scanned document image with a reconstructed font. Through our evaluation method, we have measured a 98.4% accuracy for the assignment of letters to candidate fonts in multi-font documents.
A new pre-classification method based on associative matching method
NASA Astrophysics Data System (ADS)
Katsuyama, Yutaka; Minagawa, Akihiro; Hotta, Yoshinobu; Omachi, Shinichiro; Kato, Nei
2010-01-01
Reducing the time complexity of character matching is critical to the development of efficient Japanese Optical Character Recognition (OCR) systems. To shorten processing time, recognition is usually split into separate preclassification and recognition stages. For high overall recognition performance, the pre-classification stage must both have very high classification accuracy and return only a small number of putative character categories for further processing. Furthermore, for any practical system, the speed of the pre-classification stage is also critical. The associative matching (AM) method has often been used for fast pre-classification, because its use of a hash table and reliance solely on logical bit operations to select categories makes it highly efficient. However, redundant certain level of redundancy exists in the hash table because it is constructed using only the minimum and maximum values of the data on each axis and therefore does not take account of the distribution of the data. We propose a modified associative matching method that satisfies the performance criteria described above but in a fraction of the time by modifying the hash table to reflect the underlying distribution of training characters. Furthermore, we show that our approach outperforms pre-classification by clustering, ANN and conventional AM in terms of classification accuracy, discriminative power and speed. Compared to conventional associative matching, the proposed approach results in a 47% reduction in total processing time across an evaluation test set comprising 116,528 Japanese character images.
Slant rectification in Russian passport OCR system using fast Hough transform
NASA Astrophysics Data System (ADS)
Limonova, Elena; Bezmaternykh, Pavel; Nikolaev, Dmitry; Arlazarov, Vladimir
2017-03-01
In this paper, we introduce slant detection method based on Fast Hough Transform calculation and demonstrate its application in industrial system for Russian passports recognition. About 1.5% of this kind of documents appear to be slant or italic. This fact reduces recognition rate, because Optical Recognition Systems are normally designed to process normal fonts. Our method uses Fast Hough Transform to analyse vertical strokes of characters extracted with the help of x-derivative of a text line image. To improve the quality of detector we also introduce field grouping rules. The resulting algorithm allowed to reach high detection quality. Almost all errors of considered approach happen on passports of nonstandard fonts, while slant detector works in appropriate way.
Building Structured Personal Health Records from Photographs of Printed Medical Records.
Li, Xiang; Hu, Gang; Teng, Xiaofei; Xie, Guotong
2015-01-01
Personal health records (PHRs) provide patient-centric healthcare by making health records accessible to patients. In China, it is very difficult for individuals to access electronic health records. Instead, individuals can easily obtain the printed copies of their own medical records, such as prescriptions and lab test reports, from hospitals. In this paper, we propose a practical approach to extract structured data from printed medical records photographed by mobile phones. An optical character recognition (OCR) pipeline is performed to recognize text in a document photo, which addresses the problems of low image quality and content complexity by image pre-processing and multiple OCR engine synthesis. A series of annotation algorithms that support flexible layouts are then used to identify the document type, entities of interest, and entity correlations, from which a structured PHR document is built. The proposed approach was applied to real world medical records to demonstrate the effectiveness and applicability.
Building Structured Personal Health Records from Photographs of Printed Medical Records
Li, Xiang; Hu, Gang; Teng, Xiaofei; Xie, Guotong
2015-01-01
Personal health records (PHRs) provide patient-centric healthcare by making health records accessible to patients. In China, it is very difficult for individuals to access electronic health records. Instead, individuals can easily obtain the printed copies of their own medical records, such as prescriptions and lab test reports, from hospitals. In this paper, we propose a practical approach to extract structured data from printed medical records photographed by mobile phones. An optical character recognition (OCR) pipeline is performed to recognize text in a document photo, which addresses the problems of low image quality and content complexity by image pre-processing and multiple OCR engine synthesis. A series of annotation algorithms that support flexible layouts are then used to identify the document type, entities of interest, and entity correlations, from which a structured PHR document is built. The proposed approach was applied to real world medical records to demonstrate the effectiveness and applicability. PMID:26958219
Does the cost function matter in Bayes decision rule?
Schlü ter, Ralf; Nussbaum-Thom, Markus; Ney, Hermann
2012-02-01
In many tasks in pattern recognition, such as automatic speech recognition (ASR), optical character recognition (OCR), part-of-speech (POS) tagging, and other string recognition tasks, we are faced with a well-known inconsistency: The Bayes decision rule is usually used to minimize string (symbol sequence) error, whereas, in practice, we want to minimize symbol (word, character, tag, etc.) error. When comparing different recognition systems, we do indeed use symbol error rate as an evaluation measure. The topic of this work is to analyze the relation between string (i.e., 0-1) and symbol error (i.e., metric, integer valued) cost functions in the Bayes decision rule, for which fundamental analytic results are derived. Simple conditions are derived for which the Bayes decision rule with integer-valued metric cost function and with 0-1 cost gives the same decisions or leads to classes with limited cost. The corresponding conditions can be tested with complexity linear in the number of classes. The results obtained do not make any assumption w.r.t. the structure of the underlying distributions or the classification problem. Nevertheless, the general analytic results are analyzed via simulations of string recognition problems with Levenshtein (edit) distance cost function. The results support earlier findings that considerable improvements are to be expected when initial error rates are high.
Document image database indexing with pictorial dictionary
NASA Astrophysics Data System (ADS)
Akbari, Mohammad; Azimi, Reza
2010-02-01
In this paper we introduce a new approach for information retrieval from Persian document image database without using Optical Character Recognition (OCR).At first an attribute called subword upper contour label is defined then, a pictorial dictionary is constructed based on this attribute for the subwords. By this approach we address two issues in document image retrieval: keyword spotting and retrieval according to the document similarities. The proposed methods have been evaluated on a Persian document image database. The results have proved the ability of this approach in document image information retrieval.
Digital Archiving: Where the Past Lives Again
NASA Astrophysics Data System (ADS)
Paxson, K. B.
2012-06-01
The process of digital archiving for variable star data by manual entry with an Excel spreadsheet is described. Excel-based tools including a Step Magnitude Calculator and a Julian Date Calculator for variable star observations where magnitudes and Julian dates have not been reduced are presented. Variable star data in the literature and the AAVSO International Database prior to 1911 are presented and reviewed, with recent archiving work being highlighted. Digitization using optical character recognition software conversion is also demonstrated, with editing and formatting suggestions for the OCR-converted text.
Pattern matching techniques for correcting low-confidence OCR words in a known context
NASA Astrophysics Data System (ADS)
Ford, Glenn; Hauser, Susan E.; Le, Daniel X.; Thoma, George R.
2000-12-01
A commercial OCR system is a key component of a system developed at the National Library of Medicine for the automated extraction of bibliographic fields from biomedical journals. This 5-engine OCR system, while exhibiting high performance overall, does not reliably convert very small characters, especially those that are in italics. As a result, the 'affiliations' field that typically contains such characters in most journals, is not captured accurately, and requires a disproportionately high manual input. To correct this problem, dictionaries have been created from words occurring in this field (e.g., university, department, street addresses, names of cities, etc.) from 230,000 articles already processed. The OCR output corresponding to the affiliation field is then matched against these dictionary entries by approximate string-matching techniques, and the ranked matches are presented to operators for verification. This paper outlines the techniques employed and the results of a comparative evaluation.
Approximate string matching algorithms for limited-vocabulary OCR output correction
NASA Astrophysics Data System (ADS)
Lasko, Thomas A.; Hauser, Susan E.
2000-12-01
Five methods for matching words mistranslated by optical character recognition to their most likely match in a reference dictionary were tested on data from the archives of the National Library of Medicine. The methods, including an adaptation of the cross correlation algorithm, the generic edit distance algorithm, the edit distance algorithm with a probabilistic substitution matrix, Bayesian analysis, and Bayesian analysis on an actively thinned reference dictionary were implemented and their accuracy rates compared. Of the five, the Bayesian algorithm produced the most correct matches (87%), and had the advantage of producing scores that have a useful and practical interpretation.
ESARR: enhanced situational awareness via road sign recognition
NASA Astrophysics Data System (ADS)
Perlin, V. E.; Johnson, D. B.; Rohde, M. M.; Lupa, R. M.; Fiorani, G.; Mohammad, S.
2010-04-01
The enhanced situational awareness via road sign recognition (ESARR) system provides vehicle position estimates in the absence of GPS signal via automated processing of roadway fiducials (primarily directional road signs). Sign images are detected and extracted from vehicle-mounted camera system, and preprocessed and read via a custom optical character recognition (OCR) system specifically designed to cope with low quality input imagery. Vehicle motion and 3D scene geometry estimation enables efficient and robust sign detection with low false alarm rates. Multi-level text processing coupled with GIS database validation enables effective interpretation even of extremely low resolution low contrast sign images. In this paper, ESARR development progress will be reported on, including the design and architecture, image processing framework, localization methodologies, and results to date. Highlights of the real-time vehicle-based directional road-sign detection and interpretation system will be described along with the challenges and progress in overcoming them.
TreeRipper web application: towards a fully automated optical tree recognition software.
Hughes, Joseph
2011-05-20
Relationships between species, genes and genomes have been printed as trees for over a century. Whilst this may have been the best format for exchanging and sharing phylogenetic hypotheses during the 20th century, the worldwide web now provides faster and automated ways of transferring and sharing phylogenetic knowledge. However, novel software is needed to defrost these published phylogenies for the 21st century. TreeRipper is a simple website for the fully-automated recognition of multifurcating phylogenetic trees (http://linnaeus.zoology.gla.ac.uk/~jhughes/treeripper/). The program accepts a range of input image formats (PNG, JPG/JPEG or GIF). The underlying command line c++ program follows a number of cleaning steps to detect lines, remove node labels, patch-up broken lines and corners and detect line edges. The edge contour is then determined to detect the branch length, tip label positions and the topology of the tree. Optical Character Recognition (OCR) is used to convert the tip labels into text with the freely available tesseract-ocr software. 32% of images meeting the prerequisites for TreeRipper were successfully recognised, the largest tree had 115 leaves. Despite the diversity of ways phylogenies have been illustrated making the design of a fully automated tree recognition software difficult, TreeRipper is a step towards automating the digitization of past phylogenies. We also provide a dataset of 100 tree images and associated tree files for training and/or benchmarking future software. TreeRipper is an open source project licensed under the GNU General Public Licence v3.
A guide for digitising manuscript climate data
NASA Astrophysics Data System (ADS)
Brönnimann, S.; Annis, J.; Dann, W.; Ewen, T.; Grant, A. N.; Griesser, T.; Krähenmann, S.; Mohr, C.; Scherer, M.; Vogler, C.
2006-05-01
Hand-written or printed manuscript data are an important source for paleo-climatological studies, but bringing them into a suitable format can be a time consuming adventure with uncertain success. Before starting the digitising work, it is worthwhile spending a few thoughts on the characteristics of the data, the scientific requirements with respect to quality and coverage, and on the different digitising techniques. Here we briefly discuss the most important considerations and report our own experience. We describe different methods for digitising numeric or text data, i.e., optical character recognition (OCR), speech recognition, and key entry. Each technique has its advantages and disadvantages that may become important for certain applications. It is therefore crucial to thoroughly investigate beforehand the characteristics of the manuscript data, define the quality targets and develop validation strategies.
Con-Text: Text Detection for Fine-grained Object Classification.
Karaoglu, Sezer; Tao, Ran; van Gemert, Jan C; Gevers, Theo
2017-05-24
This work focuses on fine-grained object classification using recognized scene text in natural images. While the state-of-the-art relies on visual cues only, this paper is the first work which proposes to combine textual and visual cues. Another novelty is the textual cue extraction. Unlike the state-of-the-art text detection methods, we focus more on the background instead of text regions. Once text regions are detected, they are further processed by two methods to perform text recognition i.e. ABBYY commercial OCR engine and a state-of-the-art character recognition algorithm. Then, to perform textual cue encoding, bi- and trigrams are formed between the recognized characters by considering the proposed spatial pairwise constraints. Finally, extracted visual and textual cues are combined for fine-grained classification. The proposed method is validated on four publicly available datasets: ICDAR03, ICDAR13, Con-Text and Flickr-logo. We improve the state-of-the-art end-to-end character recognition by a large margin of 15% on ICDAR03. We show that textual cues are useful in addition to visual cues for fine-grained classification. We show that textual cues are also useful for logo retrieval. Adding textual cues outperforms visual- and textual-only in fine-grained classification (70.7% to 60.3%) and logo retrieval (57.4% to 54.8%).
Recognizable or Not: Towards Image Semantic Quality Assessment for Compression
NASA Astrophysics Data System (ADS)
Liu, Dong; Wang, Dandan; Li, Houqiang
2017-12-01
Traditionally, image compression was optimized for the pixel-wise fidelity or the perceptual quality of the compressed images given a bit-rate budget. But recently, compressed images are more and more utilized for automatic semantic analysis tasks such as recognition and retrieval. For these tasks, we argue that the optimization target of compression is no longer perceptual quality, but the utility of the compressed images in the given automatic semantic analysis task. Accordingly, we propose to evaluate the quality of the compressed images neither at pixel level nor at perceptual level, but at semantic level. In this paper, we make preliminary efforts towards image semantic quality assessment (ISQA), focusing on the task of optical character recognition (OCR) from compressed images. We propose a full-reference ISQA measure by comparing the features extracted from text regions of original and compressed images. We then propose to integrate the ISQA measure into an image compression scheme. Experimental results show that our proposed ISQA measure is much better than PSNR and SSIM in evaluating the semantic quality of compressed images; accordingly, adopting our ISQA measure to optimize compression for OCR leads to significant bit-rate saving compared to using PSNR or SSIM. Moreover, we perform subjective test about text recognition from compressed images, and observe that our ISQA measure has high consistency with subjective recognizability. Our work explores new dimensions in image quality assessment, and demonstrates promising direction to achieve higher compression ratio for specific semantic analysis tasks.
Optimal frame-by-frame result combination strategy for OCR in video stream
NASA Astrophysics Data System (ADS)
Bulatov, Konstantin; Lynchenko, Aleksander; Krivtsov, Valeriy
2018-04-01
This paper describes the problem of combining classification results of multiple observations of one object. This task can be regarded as a particular case of a decision-making using a combination of experts votes with calculated weights. The accuracy of various methods of combining the classification results depending on different models of input data is investigated on the example of frame-by-frame character recognition in a video stream. Experimentally it is shown that the strategy of choosing a single most competent expert in case of input data without irrelevant observations has an advantage (in this case irrelevant means with character localization and segmentation errors). At the same time this work demonstrates the advantage of combining several most competent experts according to multiplication rule or voting if irrelevant samples are present in the input data.
Office for Civil Rights Annual Report to Congress, Fiscal Year 2005
ERIC Educational Resources Information Center
US Department of Education, 2006
2006-01-01
In this report, the Office for Civil Rights (OCR) provides a summary of its FY 2005 accomplishments. Also, in recognition of the 25th anniversary, some of the significant cases that OCR has resolved over the years are highlighted. The last quarter century was eventful and challenging for OCR. Major U.S. Supreme Court decisions, such as "Grove…
Basic test framework for the evaluation of text line segmentation and text parameter extraction.
Brodić, Darko; Milivojević, Dragan R; Milivojević, Zoran
2010-01-01
Text line segmentation is an essential stage in off-line optical character recognition (OCR) systems. It is a key because inaccurately segmented text lines will lead to OCR failure. Text line segmentation of handwritten documents is a complex and diverse problem, complicated by the nature of handwriting. Hence, text line segmentation is a leading challenge in handwritten document image processing. Due to inconsistencies in measurement and evaluation of text segmentation algorithm quality, some basic set of measurement methods is required. Currently, there is no commonly accepted one and all algorithm evaluation is custom oriented. In this paper, a basic test framework for the evaluation of text feature extraction algorithms is proposed. This test framework consists of a few experiments primarily linked to text line segmentation, skew rate and reference text line evaluation. Although they are mutually independent, the results obtained are strongly cross linked. In the end, its suitability for different types of letters and languages as well as its adaptability are its main advantages. Thus, the paper presents an efficient evaluation method for text analysis algorithms.
Basic Test Framework for the Evaluation of Text Line Segmentation and Text Parameter Extraction
Brodić, Darko; Milivojević, Dragan R.; Milivojević, Zoran
2010-01-01
Text line segmentation is an essential stage in off-line optical character recognition (OCR) systems. It is a key because inaccurately segmented text lines will lead to OCR failure. Text line segmentation of handwritten documents is a complex and diverse problem, complicated by the nature of handwriting. Hence, text line segmentation is a leading challenge in handwritten document image processing. Due to inconsistencies in measurement and evaluation of text segmentation algorithm quality, some basic set of measurement methods is required. Currently, there is no commonly accepted one and all algorithm evaluation is custom oriented. In this paper, a basic test framework for the evaluation of text feature extraction algorithms is proposed. This test framework consists of a few experiments primarily linked to text line segmentation, skew rate and reference text line evaluation. Although they are mutually independent, the results obtained are strongly cross linked. In the end, its suitability for different types of letters and languages as well as its adaptability are its main advantages. Thus, the paper presents an efficient evaluation method for text analysis algorithms. PMID:22399932
Launch Control System Software Development System Automation Testing
NASA Technical Reports Server (NTRS)
Hwang, Andrew
2017-01-01
The Spaceport Command and Control System (SCCS) is the National Aeronautics and Space Administration's (NASA) launch control system for the Orion capsule and Space Launch System, the next generation manned rocket currently in development. This system requires high quality testing that will measure and test the capabilities of the system. For the past two years, the Exploration and Operations Division at Kennedy Space Center (KSC) has assigned a group including interns and full-time engineers to develop automated tests to save the project time and money. The team worked on automating the testing process for the SCCS GUI that would use streamed simulated data from the testing servers to produce data, plots, statuses, etc. to the GUI. The software used to develop automated tests included an automated testing framework and an automation library. The automated testing framework has a tabular-style syntax, which means the functionality of a line of code must have the appropriate number of tabs for the line to function as intended. The header section contains either paths to custom resources or the names of libraries being used. The automation library contains functionality to automate anything that appears on a desired screen with the use of image recognition software to detect and control GUI components. The data section contains any data values strictly created for the current testing file. The body section holds the tests that are being run. The function section can include any number of functions that may be used by the current testing file or any other file that resources it. The resources and body section are required for all test files; the data and function sections can be left empty if the data values and functions being used are from a resourced library or another file. To help equip the automation team with better tools, the Project Lead of the Automated Testing Team, Jason Kapusta, assigned the task to install and train an optical character recognition (OCR) tool to Brandon Echols, a fellow intern, and I. The purpose of the OCR tool is to analyze an image and find the coordinates of any group of text. Some issues that arose while installing the OCR tool included the absence of certain libraries needed to train the tool and an outdated software version. We eventually resolved the issues and successfully installed the OCR tool. Training the tool required many images and different fonts and sizes, but in the end the tool learned to accurately decipher the text in the images and their coordinates. The OCR tool produced a file that contained significant metadata for each section of text, but only the text and coordinates of the text was required for our purpose. The team made a script to parse the information we wanted from the OCR file to a different file that would be used by automation functions within the automated framework. Since a majority of development and testing for the automated test cases for the GUI in question has been done using live simulated data on the workstations at the Launch Control Center (LCC), a large amount of progress has been made. As of this writing, about 60% of all of automated testing has been implemented. Additionally, the OCR tool will help make our automated tests more robust due to the tool's text recognition being highly scalable to different text fonts and text sizes. Soon we will have the whole test system automated, allowing for more full-time engineers working on development projects.
The Precise and Efficient Identification of Medical Order Forms Using Shape Trees
NASA Astrophysics Data System (ADS)
Henker, Uwe; Petersohn, Uwe; Ultsch, Alfred
A powerful and flexible technique to identify, classify and process documents using images from a scanning process is presented. The types of documents can be described to the system as a set of differentiating features in a case base using shape trees. The features are filtered and abstracted from an extremely reduced scanner image of the document. Classification rules are stored with the cases to enable precise recognition and further mark reading and Optical Character Recognition (OCR) process. The method is implemented in a system which actually processes the majority of requests for medical lab procedures in Germany. A large practical experiment with data from practitioners was performed. An average of 97% of the forms were correctly identified; none were identified incorrectly. This meets the quality requirements for most medical applications. The modular description of the recognition process allows for a flexible adaptation of future changes to the form and content of the document’s structures.
Open source OCR framework using mobile devices
NASA Astrophysics Data System (ADS)
Zhou, Steven Zhiying; Gilani, Syed Omer; Winkler, Stefan
2008-02-01
Mobile phones have evolved from passive one-to-one communication device to powerful handheld computing device. Today most new mobile phones are capable of capturing images, recording video, and browsing internet and do much more. Exciting new social applications are emerging on mobile landscape, like, business card readers, sing detectors and translators. These applications help people quickly gather the information in digital format and interpret them without the need of carrying laptops or tablet PCs. However with all these advancements we find very few open source software available for mobile phones. For instance currently there are many open source OCR engines for desktop platform but, to our knowledge, none are available on mobile platform. Keeping this in perspective we propose a complete text detection and recognition system with speech synthesis ability, using existing desktop technology. In this work we developed a complete OCR framework with subsystems from open source desktop community. This includes a popular open source OCR engine named Tesseract for text detection & recognition and Flite speech synthesis module, for adding text-to-speech ability.
NASA Astrophysics Data System (ADS)
Hsu, Chih-Yu; Huang, Hsuan-Yu; Lee, Lin-Tsang
2010-12-01
The paper propose a new procedure including four stages in order to preserve the desired edges during the image processing of noise reduction. A denoised image can be obtained from a noisy image at the first stage of the procedure. At the second stage, an edge map can be obtained by the Canny edge detector to find the edges of the object contours. Manual modification of an edge map at the third stage is optional to capture all the desired edges of the object contours. At the final stage, a new method called Edge Preserved Inhomogeneous Diffusion Equation (EPIDE) is used to smooth the noisy images or the previously denoised image at the first stage for achieving the edge preservation. The Optical Character Recognition (OCR) results in the experiments show that the proposed procedure has the best recognition result because of the capability of edge preservation.
Roy, Swapnoneel; Thakur, Ashok Kumar
2008-01-01
Genome rearrangements have been modelled by a variety of primitives such as reversals, transpositions, block moves and block interchanges. We consider such a genome rearrangement primitive Strip Exchanges. Given a permutation, the challenge is to sort it by using minimum number of strip exchanges. A strip exchanging move interchanges the positions of two chosen strips so that they merge with other strips. The strip exchange problem is to sort a permutation using minimum number of strip exchanges. We present here the first non-trivial 2-approximation algorithm to this problem. We also observe that sorting by strip-exchanges is fixed-parameter-tractable. Lastly we discuss the application of strip exchanges in a different area Optical Character Recognition (OCR) with an example.
Tian, Yingli; Yang, Xiaodong; Yi, Chucai; Arditi, Aries
2013-04-01
Independent travel is a well known challenge for blind and visually impaired persons. In this paper, we propose a proof-of-concept computer vision-based wayfinding aid for blind people to independently access unfamiliar indoor environments. In order to find different rooms (e.g. an office, a lab, or a bathroom) and other building amenities (e.g. an exit or an elevator), we incorporate object detection with text recognition. First we develop a robust and efficient algorithm to detect doors, elevators, and cabinets based on their general geometric shape, by combining edges and corners. The algorithm is general enough to handle large intra-class variations of objects with different appearances among different indoor environments, as well as small inter-class differences between different objects such as doors and door-like cabinets. Next, in order to distinguish intra-class objects (e.g. an office door from a bathroom door), we extract and recognize text information associated with the detected objects. For text recognition, we first extract text regions from signs with multiple colors and possibly complex backgrounds, and then apply character localization and topological analysis to filter out background interference. The extracted text is recognized using off-the-shelf optical character recognition (OCR) software products. The object type, orientation, location, and text information are presented to the blind traveler as speech.
Tian, YingLi; Yang, Xiaodong; Yi, Chucai; Arditi, Aries
2012-01-01
Independent travel is a well known challenge for blind and visually impaired persons. In this paper, we propose a proof-of-concept computer vision-based wayfinding aid for blind people to independently access unfamiliar indoor environments. In order to find different rooms (e.g. an office, a lab, or a bathroom) and other building amenities (e.g. an exit or an elevator), we incorporate object detection with text recognition. First we develop a robust and efficient algorithm to detect doors, elevators, and cabinets based on their general geometric shape, by combining edges and corners. The algorithm is general enough to handle large intra-class variations of objects with different appearances among different indoor environments, as well as small inter-class differences between different objects such as doors and door-like cabinets. Next, in order to distinguish intra-class objects (e.g. an office door from a bathroom door), we extract and recognize text information associated with the detected objects. For text recognition, we first extract text regions from signs with multiple colors and possibly complex backgrounds, and then apply character localization and topological analysis to filter out background interference. The extracted text is recognized using off-the-shelf optical character recognition (OCR) software products. The object type, orientation, location, and text information are presented to the blind traveler as speech. PMID:23630409
Cost analysis of a project to digitize classic articles in neurosurgery*
Bauer, Kathleen
2002-01-01
In summer 2000, the Cushing/Whitney Medical Library at Yale University began a demonstration project to digitize classic articles in neurosurgery from the late 1800s and early 1900s. The objective of the first phase of the project was to measure the time and costs involved in digitization, and those results are reported here. In the second phase, metadata will be added to the digitized articles, and the project will be publicized. Thirteen articles were scanned using optical character recognition (OCR) software, and the resulting text files were carefully proofread. Time for photocopying, scanning, and proofreading were recorded. This project achieved an average cost per item (total pages plus images) of $4.12, a figure at the high end of average costs found in other studies. This project experienced high costs for two reasons. First, the articles contained many images, which required extra processing. Second, the older fonts and the poor condition of many of these articles complicated the OCR process. The average article cost $84.46 to digitize. Although costs were high, the selection of historically important articles maximized the benefit gained from the investment in digitization. PMID:11999182
How well does multiple OCR error correction generalize?
NASA Astrophysics Data System (ADS)
Lund, William B.; Ringger, Eric K.; Walker, Daniel D.
2013-12-01
As the digitization of historical documents, such as newspapers, becomes more common, the need of the archive patron for accurate digital text from those documents increases. Building on our earlier work, the contributions of this paper are: 1. in demonstrating the applicability of novel methods for correcting optical character recognition (OCR) on disparate data sets, including a new synthetic training set, 2. enhancing the correction algorithm with novel features, and 3. assessing the data requirements of the correction learning method. First, we correct errors using conditional random fields (CRF) trained on synthetic training data sets in order to demonstrate the applicability of the methodology to unrelated test sets. Second, we show the strength of lexical features from the training sets on two unrelated test sets, yielding a relative reduction in word error rate on the test sets of 6.52%. New features capture the recurrence of hypothesis tokens and yield an additional relative reduction in WER of 2.30%. Further, we show that only 2.0% of the full training corpus of over 500,000 feature cases is needed to achieve correction results comparable to those using the entire training corpus, effectively reducing both the complexity of the training process and the learned correction model.
Cost analysis of a project to digitize classic articles in neurosurgery.
Bauer, Kathleen
2002-04-01
In summer 2000, the Cushing/Whitney Medical Library at Yale University began a demonstration project to digitize classic articles in neurosurgery from the late 1800s and early 1900s. The objective of the first phase of the project was to measure the time and costs involved in digitization, and those results are reported here. In the second phase, metadata will be added to the digitized articles, and the project will be publicized. Thirteen articles were scanned using optical character recognition (OCR) software, and the resulting text files were carefully proofread. Time for photocopying, scanning, and proofreading were recorded. This project achieved an average cost per item (total pages plus images) of $4.12, a figure at the high end of average costs found in other studies. This project experienced high costs for two reasons. First, the articles contained many images, which required extra processing. Second, the older fonts and the poor condition of many of these articles complicated the OCR process. The average article cost $84.46 to digitize. Although costs were high, the selection of historically important articles maximized the benefit gained from the investment in digitization.
Texture for script identification.
Busch, Andrew; Boles, Wageeh W; Sridharan, Sridha
2005-11-01
The problem of determining the script and language of a document image has a number of important applications in the field of document analysis, such as indexing and sorting of large collections of such images, or as a precursor to optical character recognition (OCR). In this paper, we investigate the use of texture as a tool for determining the script of a document image, based on the observation that text has a distinct visual texture. An experimental evaluation of a number of commonly used texture features is conducted on a newly created script database, providing a qualitative measure of which features are most appropriate for this task. Strategies for improving classification results in situations with limited training data and multiple font types are also proposed.
A contour-based shape descriptor for biomedical image classification and retrieval
NASA Astrophysics Data System (ADS)
You, Daekeun; Antani, Sameer; Demner-Fushman, Dina; Thoma, George R.
2013-12-01
Contours, object blobs, and specific feature points are utilized to represent object shapes and extract shape descriptors that can then be used for object detection or image classification. In this research we develop a shape descriptor for biomedical image type (or, modality) classification. We adapt a feature extraction method used in optical character recognition (OCR) for character shape representation, and apply various image preprocessing methods to successfully adapt the method to our application. The proposed shape descriptor is applied to radiology images (e.g., MRI, CT, ultrasound, X-ray, etc.) to assess its usefulness for modality classification. In our experiment we compare our method with other visual descriptors such as CEDD, CLD, Tamura, and PHOG that extract color, texture, or shape information from images. The proposed method achieved the highest classification accuracy of 74.1% among all other individual descriptors in the test, and when combined with CSD (color structure descriptor) showed better performance (78.9%) than using the shape descriptor alone.
Evaluation of image deblurring methods via a classification metric
NASA Astrophysics Data System (ADS)
Perrone, Daniele; Humphreys, David; Lamb, Robert A.; Favaro, Paolo
2012-09-01
The performance of single image deblurring algorithms is typically evaluated via a certain discrepancy measure between the reconstructed image and the ideal sharp image. The choice of metric, however, has been a source of debate and has also led to alternative metrics based on human visual perception. While fixed metrics may fail to capture some small but visible artifacts, perception-based metrics may favor reconstructions with artifacts that are visually pleasant. To overcome these limitations, we propose to assess the quality of reconstructed images via a task-driven metric. In this paper we consider object classification as the task and therefore use the rate of classification as the metric to measure deblurring performance. In our evaluation we use data with different types of blur in two cases: Optical Character Recognition (OCR), where the goal is to recognise characters in a black and white image, and object classification with no restrictions on pose, illumination and orientation. Finally, we show how off-the-shelf classification algorithms benefit from working with deblurred images.
High-speed railway real-time localization auxiliary method based on deep neural network
NASA Astrophysics Data System (ADS)
Chen, Dongjie; Zhang, Wensheng; Yang, Yang
2017-11-01
High-speed railway intelligent monitoring and management system is composed of schedule integration, geographic information, location services, and data mining technology for integration of time and space data. Assistant localization is a significant submodule of the intelligent monitoring system. In practical application, the general access is to capture the image sequences of the components by using a high-definition camera, digital image processing technique and target detection, tracking and even behavior analysis method. In this paper, we present an end-to-end character recognition method based on a deep CNN network called YOLO-toc for high-speed railway pillar plate number. Different from other deep CNNs, YOLO-toc is an end-to-end multi-target detection framework, furthermore, it exhibits a state-of-art performance on real-time detection with a nearly 50fps achieved on GPU (GTX960). Finally, we realize a real-time but high-accuracy pillar plate number recognition system and integrate natural scene OCR into a dedicated classification YOLO-toc model.
Face recognition by applying wavelet subband representation and kernel associative memory.
Zhang, Bai-Ling; Zhang, Haihong; Ge, Shuzhi Sam
2004-01-01
In this paper, we propose an efficient face recognition scheme which has two features: 1) representation of face images by two-dimensional (2-D) wavelet subband coefficients and 2) recognition by a modular, personalised classification method based on kernel associative memory models. Compared to PCA projections and low resolution "thumb-nail" image representations, wavelet subband coefficients can efficiently capture substantial facial features while keeping computational complexity low. As there are usually very limited samples, we constructed an associative memory (AM) model for each person and proposed to improve the performance of AM models by kernel methods. Specifically, we first applied kernel transforms to each possible training pair of faces sample and then mapped the high-dimensional feature space back to input space. Our scheme using modular autoassociative memory for face recognition is inspired by the same motivation as using autoencoders for optical character recognition (OCR), for which the advantages has been proven. By associative memory, all the prototypical faces of one particular person are used to reconstruct themselves and the reconstruction error for a probe face image is used to decide if the probe face is from the corresponding person. We carried out extensive experiments on three standard face recognition datasets, the FERET data, the XM2VTS data, and the ORL data. Detailed comparisons with earlier published results are provided and our proposed scheme offers better recognition accuracy on all of the face datasets.
Mobile-based text recognition from water quality devices
NASA Astrophysics Data System (ADS)
Dhakal, Shanti; Rahnemoonfar, Maryam
2015-03-01
Measuring water quality of bays, estuaries, and gulfs is a complicated and time-consuming process. YSI Sonde is an instrument used to measure water quality parameters such as pH, temperature, salinity, and dissolved oxygen. This instrument is taken to water bodies in a boat trip and researchers note down different parameters displayed by the instrument's display monitor. In this project, a mobile application is developed for Android platform that allows a user to take a picture of the YSI Sonde monitor, extract text from the image and store it in a file on the phone. The image captured by the application is first processed to remove perspective distortion. Probabilistic Hough line transform is used to identify lines in the image and the corner of the image is then obtained by determining the intersection of the detected horizontal and vertical lines. The image is warped using the perspective transformation matrix, obtained from the corner points of the source image and the destination image, hence, removing the perspective distortion. Mathematical morphology operation, black-hat is used to correct the shading of the image. The image is binarized using Otsu's binarization technique and is then passed to the Optical Character Recognition (OCR) software for character recognition. The extracted information is stored in a file on the phone and can be retrieved later for analysis. The algorithm was tested on 60 different images of YSI Sonde with different perspective features and shading. Experimental results, in comparison to ground-truth results, demonstrate the effectiveness of the proposed method.
Parallel processing considerations for image recognition tasks
NASA Astrophysics Data System (ADS)
Simske, Steven J.
2011-01-01
Many image recognition tasks are well-suited to parallel processing. The most obvious example is that many imaging tasks require the analysis of multiple images. From this standpoint, then, parallel processing need be no more complicated than assigning individual images to individual processors. However, there are three less trivial categories of parallel processing that will be considered in this paper: parallel processing (1) by task; (2) by image region; and (3) by meta-algorithm. Parallel processing by task allows the assignment of multiple workflows-as diverse as optical character recognition [OCR], document classification and barcode reading-to parallel pipelines. This can substantially decrease time to completion for the document tasks. For this approach, each parallel pipeline is generally performing a different task. Parallel processing by image region allows a larger imaging task to be sub-divided into a set of parallel pipelines, each performing the same task but on a different data set. This type of image analysis is readily addressed by a map-reduce approach. Examples include document skew detection and multiple face detection and tracking. Finally, parallel processing by meta-algorithm allows different algorithms to be deployed on the same image simultaneously. This approach may result in improved accuracy.
Automated sea floor extraction from underwater video
NASA Astrophysics Data System (ADS)
Kelly, Lauren; Rahmes, Mark; Stiver, James; McCluskey, Mike
2016-05-01
Ocean floor mapping using video is a method to simply and cost-effectively record large areas of the seafloor. Obtaining visual and elevation models has noteworthy applications in search and recovery missions. Hazards to navigation are abundant and pose a significant threat to the safety, effectiveness, and speed of naval operations and commercial vessels. This project's objective was to develop a workflow to automatically extract metadata from marine video and create image optical and elevation surface mosaics. Three developments made this possible. First, optical character recognition (OCR) by means of two-dimensional correlation, using a known character set, allowed for the capture of metadata from image files. Second, exploiting the image metadata (i.e., latitude, longitude, heading, camera angle, and depth readings) allowed for the determination of location and orientation of the image frame in mosaic. Image registration improved the accuracy of mosaicking. Finally, overlapping data allowed us to determine height information. A disparity map was created using the parallax from overlapping viewpoints of a given area and the relative height data was utilized to create a three-dimensional, textured elevation map.
Arabic handwritten: pre-processing and segmentation
NASA Astrophysics Data System (ADS)
Maliki, Makki; Jassim, Sabah; Al-Jawad, Naseer; Sellahewa, Harin
2012-06-01
This paper is concerned with pre-processing and segmentation tasks that influence the performance of Optical Character Recognition (OCR) systems and handwritten/printed text recognition. In Arabic, these tasks are adversely effected by the fact that many words are made up of sub-words, with many sub-words there associated one or more diacritics that are not connected to the sub-word's body; there could be multiple instances of sub-words overlap. To overcome these problems we investigate and develop segmentation techniques that first segment a document into sub-words, link the diacritics with their sub-words, and removes possible overlapping between words and sub-words. We shall also investigate two approaches for pre-processing tasks to estimate sub-words baseline, and to determine parameters that yield appropriate slope correction, slant removal. We shall investigate the use of linear regression on sub-words pixels to determine their central x and y coordinates, as well as their high density part. We also develop a new incremental rotation procedure to be performed on sub-words that determines the best rotation angle needed to realign baselines. We shall demonstrate the benefits of these proposals by conducting extensive experiments on publicly available databases and in-house created databases. These algorithms help improve character segmentation accuracy by transforming handwritten Arabic text into a form that could benefit from analysis of printed text.
Rectification of curved document images based on single view three-dimensional reconstruction.
Kang, Lai; Wei, Yingmei; Jiang, Jie; Bai, Liang; Lao, Songyang
2016-10-01
Since distortions in camera-captured document images significantly affect the accuracy of optical character recognition (OCR), distortion removal plays a critical role for document digitalization systems using a camera for image capturing. This paper proposes a novel framework that performs three-dimensional (3D) reconstruction and rectification of camera-captured document images. While most existing methods rely on additional calibrated hardware or multiple images to recover the 3D shape of a document page, or make a simple but not always valid assumption on the corresponding 3D shape, our framework is more flexible and practical since it only requires a single input image and is able to handle a general locally smooth document surface. The main contributions of this paper include a new iterative refinement scheme for baseline fitting from connected components of text line, an efficient discrete vertical text direction estimation algorithm based on convex hull projection profile analysis, and a 2D distortion grid construction method based on text direction function estimation using 3D regularization. In order to examine the performance of our proposed method, both qualitative and quantitative evaluation and comparison with several recent methods are conducted in our experiments. The experimental results demonstrate that the proposed method outperforms relevant approaches for camera-captured document image rectification, in terms of improvements on both visual distortion removal and OCR accuracy.
Image-based mobile service: automatic text extraction and translation
NASA Astrophysics Data System (ADS)
Berclaz, Jérôme; Bhatti, Nina; Simske, Steven J.; Schettino, John C.
2010-01-01
We present a new mobile service for the translation of text from images taken by consumer-grade cell-phone cameras. Such capability represents a new paradigm for users where a simple image provides the basis for a service. The ubiquity and ease of use of cell-phone cameras enables acquisition and transmission of images anywhere and at any time a user wishes, delivering rapid and accurate translation over the phone's MMS and SMS facilities. Target text is extracted completely automatically, requiring no bounding box delineation or related user intervention. The service uses localization, binarization, text deskewing, and optical character recognition (OCR) in its analysis. Once the text is translated, an SMS message is sent to the user with the result. Further novelties include that no software installation is required on the handset, any service provider or camera phone can be used, and the entire service is implemented on the server side.
An integrated information retrieval and document management system
NASA Technical Reports Server (NTRS)
Coles, L. Stephen; Alvarez, J. Fernando; Chen, James; Chen, William; Cheung, Lai-Mei; Clancy, Susan; Wong, Alexis
1993-01-01
This paper describes the requirements and prototype development for an intelligent document management and information retrieval system that will be capable of handling millions of pages of text or other data. Technologies for scanning, Optical Character Recognition (OCR), magneto-optical storage, and multiplatform retrieval using a Standard Query Language (SQL) will be discussed. The semantic ambiguity inherent in the English language is somewhat compensated-for through the use of coefficients or weighting factors for partial synonyms. Such coefficients are used both for defining structured query trees for routine queries and for establishing long-term interest profiles that can be used on a regular basis to alert individual users to the presence of relevant documents that may have just arrived from an external source, such as a news wire service. Although this attempt at evidential reasoning is limited in comparison with the latest developments in AI Expert Systems technology, it has the advantage of being commercially available.
Page segmentation using script identification vectors: A first look
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hochberg, J.; Cannon, M.; Kelly, P.
1997-07-01
Document images in which different scripts, such as Chinese and Roman, appear on a single page pose a problem for optical character recognition (OCR) systems. This paper explores the use of script identification vectors in the analysis of multilingual document images. A script identification vector is calculated for each connected component in a document. The vector expresses the closest distance between the component and templates developed for each of thirteen scripts, including Arabic, Chinese, Cyrillic, and Roman. The authors calculate the first three principal components within the resulting thirteen-dimensional space for each image. By mapping these components to red, green,more » and blue, they can visualize the information contained in the script identification vectors. The visualization of several multilingual images suggests that the script identification vectors can be used to segment images into script-specific regions as large as several paragraphs or as small as a few characters. The visualized vectors also reveal distinctions within scripts, such as font in Roman documents, and kanji vs. kana in Japanese. Results are best for documents containing highly dissimilar scripts such as Roman and Japanese. Documents containing similar scripts, such as Roman and Cyrillic will require further investigation.« less
Multioriented and curved text lines extraction from Indian documents.
Pal, U; Roy, Partha Pratim
2004-08-01
There are printed artistic documents where text lines of a single page may not be parallel to each other. These text lines may have different orientations or the text lines may be curved shapes. For the optical character recognition (OCR) of these documents, we need to extract such lines properly. In this paper, we propose a novel scheme, mainly based on the concept of water reservoir analogy, to extract individual text lines from printed Indian documents containing multioriented and/or curve text lines. A reservoir is a metaphor to illustrate the cavity region of a character where water can be stored. In the proposed scheme, at first, connected components are labeled and identified either as isolated or touching. Next, each touching component is classified either straight type (S-type) or curve type (C-type), depending on the reservoir base-area and envelope points of the component. Based on the type (S-type or C-type) of a component two candidate points are computed from each touching component. Finally, candidate regions (neighborhoods of the candidate points) of the candidate points of each component are detected and after analyzing these candidate regions, components are grouped to get individual text lines.
Composition of a dewarped and enhanced document image from two view images.
Koo, Hyung Il; Kim, Jinho; Cho, Nam Ik
2009-07-01
In this paper, we propose an algorithm to compose a geometrically dewarped and visually enhanced image from two document images taken by a digital camera at different angles. Unlike the conventional works that require special equipment or assumptions on the contents of books or complicated image acquisition steps, we estimate the unfolded book or document surface from the corresponding points between two images. For this purpose, the surface and camera matrices are estimated using structure reconstruction, 3-D projection analysis, and random sample consensus-based curve fitting with the cylindrical surface model. Because we do not need any assumption on the contents of books, the proposed method can be applied not only to optical character recognition (OCR), but also to the high-quality digitization of pictures in documents. In addition to the dewarping for a structurally better image, image mosaic is also performed for further improving the visual quality. By finding better parts of images (with less out of focus blur and/or without specular reflections) from either of views, we compose a better image by stitching and blending them. These processes are formulated as energy minimization problems that can be solved using a graph cut method. Experiments on many kinds of book or document images show that the proposed algorithm robustly works and yields visually pleasing results. Also, the OCR rate of the resulting image is comparable to that of document images from a flatbed scanner.
Comparison of the scanned pages of the contractual documents
NASA Astrophysics Data System (ADS)
Andreeva, Elena; Arlazarov, Vladimir V.; Manzhikov, Temudzhin; Slavin, Oleg
2018-04-01
In this paper the problem statement is given to compare the digitized pages of the official papers. Such problem appears during the comparison of two customer copies signed at different times between two parties with a view to find the possible modifications introduced on the one hand. This problem is a practically significant in the banking sector during the conclusion of contracts in a paper format. The method of comparison based on the recognition, which consists in the comparison of two bag-of-words, which are the recognition result of the master and test pages, is suggested. The described experiments were conducted using the OCR Tesseract and the siamese neural network. The advantages of the suggested method are the steady operation of the comparison algorithm and the high exacting precision, and one of the disadvantages is the dependence on the chosen OCR.
Historical files from Federal government mineral exploration-assistance programs, 1950 to 1974
Frank, David G.
2010-01-01
Congress enacted the Defense Production Act in 1950 to provide funding and support for the exploration and development of critical mineral resources. From 1950 to 1974, three Department of the Interior agencies carried out this mission. Contracts with mine owners provided financial assistance for mineral exploration on a joint-participation basis. These contracts are documented in more than 5,000 'dockets' now archived online by the U.S. Geological Survey. This archive provides access to unique and difficult to recreate information, such as drill logs, assay results, and underground geologic maps, that is invaluable to land and resource management organizations and the minerals industry. An effort to preserve the data began in 2009, and the entire collection of dockets was electronically scanned. The scanning process used optical character recognition (OCR) when possible, and files were converted into Portable Document Format (.pdf) files, which require Adobe Reader or similar software for viewing. In 2010, the scans were placed online (http://minerals.usgs.gov/dockets/) and are available to download free of charge.
Correcting geometric and photometric distortion of document images on a smartphone
NASA Astrophysics Data System (ADS)
Simon, Christian; Williem; Park, In Kyu
2015-01-01
A set of document image processing algorithms for improving the optical character recognition (OCR) capability of smartphone applications is presented. The scope of the problem covers the geometric and photometric distortion correction of document images. The proposed framework was developed to satisfy industrial requirements. It is implemented on an off-the-shelf smartphone with limited resources in terms of speed and memory. Geometric distortions, i.e., skew and perspective distortion, are corrected by sending horizontal and vertical vanishing points toward infinity in a downsampled image. Photometric distortion includes image degradation from moiré pattern noise and specular highlights. Moiré pattern noise is removed using low-pass filters with different sizes independently applied to the background and text region. The contrast of the text in a specular highlighted area is enhanced by locally enlarging the intensity difference between the background and text while the noise is suppressed. Intensive experiments indicate that the proposed methods show a consistent and robust performance on a smartphone with a runtime of less than 1 s.
Electronic patient registration and tracking at mass vaccination clinics: a clinical study.
Billittier, Anthony J; Lupiani, Patrick; Masterson, Gary; Masterson, Tim; Zak, Christopher
2003-01-01
To protect the citizens of the United States from the use of dangerous biological agents, the Center for Disease Control and Prevention (CDC) has been actively preparing to deal with the consequences of such an attack. Their plans include the deployment of mass immunization clinics to handle postevent vaccinations. As part of the planning efforts by the Western New York Public Health Alliance, a Web-based electronic patient registration and tracking system was developed and tested at a recent trial smallpox vaccination clinic. Initial goals were to determine what the pitfalls and benefits of using such a system might be in comparison to other methods of data collection. This exercise proved that use of an electronic system capable of scanning two-dimensional bar codes was superior to both paper-based and optical character recognition (OCR) methods of data collection and management. Major improvements in speed and/or accuracy were evident in all areas of the clinic, especially in patient registration, vaccine tracking and postclinic data analysis.
A New Pivoting and Iterative Text Detection Algorithm for Biomedical Images
DOE Office of Scientific and Technical Information (OSTI.GOV)
Xu, Songhua; Krauthammer, Prof. Michael
2010-01-01
There is interest to expand the reach of literature mining to include the analysis of biomedical images, which often contain a paper's key findings. Examples include recent studies that use Optical Character Recognition (OCR) to extract image text, which is used to boost biomedical image retrieval and classification. Such studies rely on the robust identification of text elements in biomedical images, which is a non-trivial task. In this work, we introduce a new text detection algorithm for biomedical images based on iterative projection histograms. We study the effectiveness of our algorithm by evaluating the performance on a set of manuallymore » labeled random biomedical images, and compare the performance against other state-of-the-art text detection algorithms. We demonstrate that our projection histogram-based text detection approach is well suited for text detection in biomedical images, and that the iterative application of the algorithm boosts performance to an F score of .60. We provide a C++ implementation of our algorithm freely available for academic use.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gastelum, Zoe N.; Henry, Michael J.
2013-11-13
In FY2013, the PIE International Safeguards team demonstrated our development progress to U.S. Department of Energy (DOE) staff from the Office of Nonproliferation and International Security (NA-24, our client) and the Office of Defense Nuclear Nonproliferation Research and Development (NA-22). Following the demonstration, the team was asked by our client to complete additional development prior to a planned demonstration at the International Atomic Energy Agency (IAEA), scheduled tentatively for January or spring of 2014. The team discussed four potential areas for development (in priority order), and will develop them as time and funding permit prior to an IAEA demonstration. Themore » four capability areas are: 1. Addition of equipment manuals to PIE-accessible files 2. Optical character recognition (OCR) of photographed text 3. Barcode reader with information look-up from a database 4. Add Facilities to Data Model 5. Geospatial capabilities with information integration Each area will be described below in a use case.« less
Lateral Entorhinal Cortex is Critical for Novel Object-Context Recognition
Wilson, David IG; Langston, Rosamund F; Schlesiger, Magdalene I; Wagner, Monica; Watanabe, Sakurako; Ainge, James A
2013-01-01
Episodic memory incorporates information about specific events or occasions including spatial locations and the contextual features of the environment in which the event took place. It has been modeled in rats using spontaneous exploration of novel configurations of objects, their locations, and the contexts in which they are presented. While we have a detailed understanding of how spatial location is processed in the brain relatively little is known about where the nonspatial contextual components of episodic memory are processed. Initial experiments measured c-fos expression during an object-context recognition (OCR) task to examine which networks within the brain process contextual features of an event. Increased c-fos expression was found in the lateral entorhinal cortex (LEC; a major hippocampal afferent) during OCR relative to control conditions. In a subsequent experiment it was demonstrated that rats with lesions of LEC were unable to recognize object-context associations yet showed normal object recognition and normal context recognition. These data suggest that contextual features of the environment are integrated with object identity in LEC and demonstrate that recognition of such object-context associations requires the LEC. This is consistent with the suggestion that contextual features of an event are processed in LEC and that this information is combined with spatial information from medial entorhinal cortex to form episodic memory in the hippocampus. © 2013 Wiley Periodicals, Inc. PMID:23389958
A unified approach for development of Urdu Corpus for OCR and demographic purpose
NASA Astrophysics Data System (ADS)
Choudhary, Prakash; Nain, Neeta; Ahmed, Mushtaq
2015-02-01
This paper presents a methodology for the development of an Urdu handwritten text image Corpus and application of Corpus linguistics in the field of OCR and information retrieval from handwritten document. Compared to other language scripts, Urdu script is little bit complicated for data entry. To enter a single character it requires a combination of multiple keys entry. Here, a mixed approach is proposed and demonstrated for building Urdu Corpus for OCR and Demographic data collection. Demographic part of database could be used to train a system to fetch the data automatically, which will be helpful to simplify existing manual data-processing task involved in the field of data collection such as input forms like Passport, Ration Card, Voting Card, AADHAR, Driving licence, Indian Railway Reservation, Census data etc. This would increase the participation of Urdu language community in understanding and taking benefit of the Government schemes. To make availability and applicability of database in a vast area of corpus linguistics, we propose a methodology for data collection, mark-up, digital transcription, and XML metadata information for benchmarking.
Development of a written music-recognition system using Java and open source technologies
NASA Astrophysics Data System (ADS)
Loibner, Gernot; Schwarzl, Andreas; Kovač, Matthias; Paulus, Dietmar; Pölzleitner, Wolfgang
2005-10-01
We report on the development of a software system to recognize and interpret printed music. The overall goal is to scan printed music sheets, analyze and recognize the notes, timing, and written text, and derive the all necessary information to use the computers MIDI sound system to play the music. This function is primarily useful for musicians who want to digitize printed music for editing purposes. There exist a number of commercial systems that offer such a functionality. However, on testing these systems, we were astonished on how weak they behave in their pattern recognition parts. Although we submitted very clear and rather flawless scanning input, none of these systems was able to e.g. recognize all notes, staff lines, and systems. They all require a high degree of interaction, post-processing, and editing to get a decent digital version of the hard copy material. In this paper we focus on the pattern recognition area. In a first approach we tested more or less standard methods of adaptive thresholding, blob detection, line detection, and corner detection to find the notes, staff lines, and candidate objects subject to OCR. Many of the objects on this type of material can be learned in a training phase. None of the commercial systems we saw offers the option to train special characters or unusual signatures. A second goal in this project is to use a modern software engineering platform. We were interested in how well Java and open source technologies are suitable for pattern recognition and machine vision. The scanning of music served as a case-study.
Postprocessing for character recognition using pattern features and linguistic information
NASA Astrophysics Data System (ADS)
Yoshikawa, Takatoshi; Okamoto, Masayosi; Horii, Hiroshi
1993-04-01
We propose a new method of post-processing for character recognition using pattern features and linguistic information. This method corrects errors in the recognition of handwritten Japanese sentences containing Kanji characters. This post-process method is characterized by having two types of character recognition. Improving the accuracy of the character recognition rate of Japanese characters is made difficult by the large number of characters, and the existence of characters with similar patterns. Therefore, it is not practical for a character recognition system to recognize all characters in detail. First, this post-processing method generates a candidate character table by recognizing the simplest features of characters. Then, it selects words corresponding to the character from the candidate character table by referring to a word and grammar dictionary before selecting suitable words. If the correct character is included in the candidate character table, this process can correct an error, however, if the character is not included, it cannot correct an error. Therefore, if this method can presume a character does not exist in a candidate character table by using linguistic information (word and grammar dictionary). It then can verify a presumed character by character recognition using complex features. When this method is applied to an online character recognition system, the accuracy of character recognition improves 93.5% to 94.7%. This proved to be the case when it was used for the editorials of a Japanese newspaper (Asahi Shinbun).
Post processing for offline Chinese handwritten character string recognition
NASA Astrophysics Data System (ADS)
Wang, YanWei; Ding, XiaoQing; Liu, ChangSong
2012-01-01
Offline Chinese handwritten character string recognition is one of the most important research fields in pattern recognition. Due to the free writing style, large variability in character shapes and different geometric characteristics, Chinese handwritten character string recognition is a challenging problem to deal with. However, among the current methods over-segmentation and merging method which integrates geometric information, character recognition information and contextual information, shows a promising result. It is found experimentally that a large part of errors are segmentation error and mainly occur around non-Chinese characters. In a Chinese character string, there are not only wide characters namely Chinese characters, but also narrow characters like digits and letters of the alphabet. The segmentation error is mainly caused by uniform geometric model imposed on all segmented candidate characters. To solve this problem, post processing is employed to improve recognition accuracy of narrow characters. On one hand, multi-geometric models are established for wide characters and narrow characters respectively. Under multi-geometric models narrow characters are not prone to be merged. On the other hand, top rank recognition results of candidate paths are integrated to boost final recognition of narrow characters. The post processing method is investigated on two datasets, in total 1405 handwritten address strings. The wide character recognition accuracy has been improved lightly and narrow character recognition accuracy has been increased up by 10.41% and 10.03% respectively. It indicates that the post processing method is effective to improve recognition accuracy of narrow characters.
Degraded Chinese rubbing images thresholding based on local first-order statistics
NASA Astrophysics Data System (ADS)
Wang, Fang; Hou, Ling-Ying; Huang, Han
2017-06-01
It is a necessary step for Chinese character segmentation from degraded document images in Optical Character Recognizer (OCR); however, it is challenging due to various kinds of noising in such an image. In this paper, we present three local first-order statistics method that had been adaptive thresholding for segmenting text and non-text of Chinese rubbing image. Both visual inspection and numerically investigate for the segmentation results of rubbing image had been obtained. In experiments, it obtained better results than classical techniques in the binarization of real Chinese rubbing image and PHIBD 2012 datasets.
NASA Astrophysics Data System (ADS)
Bezmaternykh, P. V.; Nikolaev, D. P.; Arlazarov, V. L.
2018-04-01
Textual blocks rectification or slant correction is an important stage of document image processing in OCR systems. This paper considers existing methods and introduces an approach for the construction of such algorithms based on Fast Hough Transform analysis. A quality measurement technique is proposed and obtained results are shown for both printed and handwritten textual blocks processing as a part of an industrial system of identity documents recognition on mobile devices.
Clustering of Farsi sub-word images for whole-book recognition
NASA Astrophysics Data System (ADS)
Soheili, Mohammad Reza; Kabir, Ehsanollah; Stricker, Didier
2015-01-01
Redundancy of word and sub-word occurrences in large documents can be effectively utilized in an OCR system to improve recognition results. Most OCR systems employ language modeling techniques as a post-processing step; however these techniques do not use important pictorial information that exist in the text image. In case of large-scale recognition of degraded documents, this information is even more valuable. In our previous work, we proposed a subword image clustering method for the applications dealing with large printed documents. In our clustering method, the ideal case is when all equivalent sub-word images lie in one cluster. To overcome the issues of low print quality, the clustering method uses an image matching algorithm for measuring the distance between two sub-word images. The measured distance with a set of simple shape features were used to cluster all sub-word images. In this paper, we analyze the effects of adding more shape features on processing time, purity of clustering, and the final recognition rate. Previously published experiments have shown the efficiency of our method on a book. Here we present extended experimental results and evaluate our method on another book with totally different font face. Also we show that the number of the new created clusters in a page can be used as a criteria for assessing the quality of print and evaluating preprocessing phases.
Recognition of Time Stamps on Full-Disk Hα Images Using Machine Learning Methods
NASA Astrophysics Data System (ADS)
Xu, Y.; Huang, N.; Jing, J.; Liu, C.; Wang, H.; Fu, G.
2016-12-01
Observation and understanding of the physics of the 11-year solar activity cycle and 22-year magnetic cycle are among the most important research topics in solar physics. The solar cycle is responsible for magnetic field and particle fluctuation in the near-earth environment that have been found increasingly important in affecting the living of human beings in the modern era. A systematic study of large-scale solar activities, as made possible by our rich data archive, will further help us to understand the global-scale magnetic fields that are closely related to solar cycles. The long-time-span data archive includes both full-disk and high-resolution Hα images. Prior to the widely use of CCD cameras in 1990s, 35-mm films were the major media to store images. The research group at NJIT recently finished the digitization of film data obtained by the National Solar Observatory (NSO) and Big Bear Solar Observatory (BBSO) covering the period of 1953 to 2000. The total volume of data exceeds 60 TB. To make this huge database scientific valuable, some processing and calibration are required. One of the most important steps is to read the time stamps on all of the 14 million images, which is almost impossible to be done manually. We implemented three different methods to recognize the time stamps automatically, including Optical Character Recognition (OCR), Classification Tree and TensorFlow. The latter two are known as machine learning algorithms which are very popular now a day in pattern recognition area. We will present some sample images and the results of clock recognition from all three methods.
A new pivoting and iterative text detection algorithm for biomedical images.
Xu, Songhua; Krauthammer, Michael
2010-12-01
There is interest to expand the reach of literature mining to include the analysis of biomedical images, which often contain a paper's key findings. Examples include recent studies that use Optical Character Recognition (OCR) to extract image text, which is used to boost biomedical image retrieval and classification. Such studies rely on the robust identification of text elements in biomedical images, which is a non-trivial task. In this work, we introduce a new text detection algorithm for biomedical images based on iterative projection histograms. We study the effectiveness of our algorithm by evaluating the performance on a set of manually labeled random biomedical images, and compare the performance against other state-of-the-art text detection algorithms. We demonstrate that our projection histogram-based text detection approach is well suited for text detection in biomedical images, and that the iterative application of the algorithm boosts performance to an F score of .60. We provide a C++ implementation of our algorithm freely available for academic use. Copyright © 2010 Elsevier Inc. All rights reserved.
The Solar Rotation in the 1930s from the Sunspot and Flocculi Catalogs of the Ebro Observatory
NASA Astrophysics Data System (ADS)
de Paula, V.; Curto, J. J.; Casas, R.
2016-10-01
The tables of sunspot and flocculi heliographic positions included in the catalogs published by the Ebro Observatory in the 1930s have recently been recovered and converted into digital format by using optical character recognition (OCR) technology. We here analyzed these data by computing the angular velocity of several sunspot and flocculi groups. A difference was found in the rotational velocity for sunspots and flocculi groups at high latitudes, and we also detected an asymmetry between the northern and southern hemispheres, which is especially marked for the flocculi groups. The results were then fitted with a differential-rotation law [ω=a+b sin2 B] to compare the data obtained with the results published by other authors. A dependence on the latitude that is consistent with former studies was found. Finally, we studied the possible relationship between the sunspot/flocculi group areas and their corresponding angular velocity. There are strong indications that the rotational velocity of a sunspot/flocculi group is reduced (in relation to the differential rotation law) when its maximum area is larger.
Restoring 2D content from distorted documents.
Brown, Michael S; Sun, Mingxuan; Yang, Ruigang; Yun, Lin; Seales, W Brent
2007-11-01
This paper presents a framework to restore the 2D content printed on documents in the presence of geometric distortion and non-uniform illumination. Compared with textbased document imaging approaches that correct distortion to a level necessary to obtain sufficiently readable text or to facilitate optical character recognition (OCR), our work targets nontextual documents where the original printed content is desired. To achieve this goal, our framework acquires a 3D scan of the document's surface together with a high-resolution image. Conformal mapping is used to rectify geometric distortion by mapping the 3D surface back to a plane while minimizing angular distortion. This conformal "deskewing" assumes no parametric model of the document's surface and is suitable for arbitrary distortions. Illumination correction is performed by using the 3D shape to distinguish content gradient edges from illumination gradient edges in the high-resolution image. Integration is performed using only the content edges to obtain a reflectance image with significantly less illumination artifacts. This approach makes no assumptions about light sources and their positions. The results from the geometric and photometric correction are combined to produce the final output.
Good initialization model with constrained body structure for scene text recognition
NASA Astrophysics Data System (ADS)
Zhu, Anna; Wang, Guoyou; Dong, Yangbo
2016-09-01
Scene text recognition has gained significant attention in the computer vision community. Character detection and recognition are the promise of text recognition and affect the overall performance to a large extent. We proposed a good initialization model for scene character recognition from cropped text regions. We use constrained character's body structures with deformable part-based models to detect and recognize characters in various backgrounds. The character's body structures are achieved by an unsupervised discriminative clustering approach followed by a statistical model and a self-build minimum spanning tree model. Our method utilizes part appearance and location information, and combines character detection and recognition in cropped text region together. The evaluation results on the benchmark datasets demonstrate that our proposed scheme outperforms the state-of-the-art methods both on scene character recognition and word recognition aspects.
Hidden Markov models for character recognition.
Vlontzos, J A; Kung, S Y
1992-01-01
A hierarchical system for character recognition with hidden Markov model knowledge sources which solve both the context sensitivity problem and the character instantiation problem is presented. The system achieves 97-99% accuracy using a two-level architecture and has been implemented using a systolic array, thus permitting real-time (1 ms per character) multifont and multisize printed character recognition as well as handwriting recognition.
Author name recognition in degraded journal images
NASA Astrophysics Data System (ADS)
de Bodard de la Jacopière, Aliette; Likforman-Sulem, Laurence
2006-01-01
A method for extracting names in degraded documents is presented in this article. The documents targeted are images of photocopied scientific journals from various scientific domains. Due to the degradation, there is poor OCR recognition, and pieces of other articles appear on the sides of the image. The proposed approach relies on the combination of a low-level textual analysis and an image-based analysis. The textual analysis extracts robust typographic features, while the image analysis selects image regions of interest through anchor components. We report results on the University of Washington benchmark database.
Kannada character recognition system using neural network
NASA Astrophysics Data System (ADS)
Kumar, Suresh D. S.; Kamalapuram, Srinivasa K.; Kumar, Ajay B. R.
2013-03-01
Handwriting recognition has been one of the active and challenging research areas in the field of pattern recognition. It has numerous applications which include, reading aid for blind, bank cheques and conversion of any hand written document into structural text form. As there is no sufficient number of works on Indian language character recognition especially Kannada script among 15 major scripts in India. In this paper an attempt is made to recognize handwritten Kannada characters using Feed Forward neural networks. A handwritten Kannada character is resized into 20x30 Pixel. The resized character is used for training the neural network. Once the training process is completed the same character is given as input to the neural network with different set of neurons in hidden layer and their recognition accuracy rate for different Kannada characters has been calculated and compared. The results show that the proposed system yields good recognition accuracy rates comparable to that of other handwritten character recognition systems.
Vehicle license plate recognition based on geometry restraints and multi-feature decision
NASA Astrophysics Data System (ADS)
Wu, Jianwei; Wang, Zongyue
2005-10-01
Vehicle license plate (VLP) recognition is of great importance to many traffic applications. Though researchers have paid much attention to VLP recognition there has not been a fully operational VLP recognition system yet for many reasons. This paper discusses a valid and practical method for vehicle license plate recognition based on geometry restraints and multi-feature decision including statistical and structural features. In general, the VLP recognition includes the following steps: the location of VLP, character segmentation, and character recognition. This paper discusses the three steps in detail. The characters of VLP are always declining caused by many factors, which makes it more difficult to recognize the characters of VLP, therefore geometry restraints such as the general ratio of length and width, the adjacent edges being perpendicular are used for incline correction. Image Moment has been proved to be invariant to translation, rotation and scaling therefore image moment is used as one feature for character recognition. Stroke is the basic element for writing and hence taking it as a feature is helpful to character recognition. Finally we take the image moment, the strokes and the numbers of each stroke for each character image and some other structural features and statistical features as the multi-feature to match each character image with sample character images so that each character image can be recognized by BP neural net. The proposed method combines statistical and structural features for VLP recognition, and the result shows its validity and efficiency.
Progress in recognizing typeset mathematics
NASA Astrophysics Data System (ADS)
Fateman, Richard J.; Tokuyasu, Taku A.
1996-03-01
Printed mathematics has a number of features which distinguish it from conventional text. These include structure in two dimensions (fractions, exponents, limits), frequent font changes, symbols with variable shape (quotient bars), and substantially differing notational conventions from source to source. When compounded with more generic problems such as noise and merged or broken characters, printed mathematics offers a challenging arena for recognition. Our project was initially driven by the goal of scanning and parsing some 5,000 pages of elaborate mathematics (tables of definite integrals). While our prototype system demonstrates success on translating noise-free typeset equations into Lisp expressions appropriate for further processing, a more semantic top-down approach appears necessary for higher levels of performance. Such an approach may benefit the incorporation of these programs into a more general document processing viewpoint. We intend to release to the public our somewhat refined prototypes as utility programs in the hope that they will be of general use in the construction of custom OCR packages. These utilities are quite fast even as originally prototyped in Lisp, where they may be of particular interest to those working on 'intelligent' optical processing. Some routines have been re-written in C++ as well. Additional programs providing formula recognition and parsing also form a part of this system. It is important however to realize that distinct conflicting grammars are needed to cover variations in contemporary and historical typesetting, and thus a single simple solution is not possible.
Chinese character recognition based on Gabor feature extraction and CNN
NASA Astrophysics Data System (ADS)
Xiong, Yudian; Lu, Tongwei; Jiang, Yongyuan
2018-03-01
As an important application in the field of text line recognition and office automation, Chinese character recognition has become an important subject of pattern recognition. However, due to the large number of Chinese characters and the complexity of its structure, there is a great difficulty in the Chinese character recognition. In order to solve this problem, this paper proposes a method of printed Chinese character recognition based on Gabor feature extraction and Convolution Neural Network(CNN). The main steps are preprocessing, feature extraction, training classification. First, the gray-scale Chinese character image is binarized and normalized to reduce the redundancy of the image data. Second, each image is convoluted with Gabor filter with different orientations, and the feature map of the eight orientations of Chinese characters is extracted. Third, the feature map through Gabor filters and the original image are convoluted with learning kernels, and the results of the convolution is the input of pooling layer. Finally, the feature vector is used to classify and recognition. In addition, the generalization capacity of the network is improved by Dropout technology. The experimental results show that this method can effectively extract the characteristics of Chinese characters and recognize Chinese characters.
Binarization algorithm for document image with complex background
NASA Astrophysics Data System (ADS)
Miao, Shaojun; Lu, Tongwei; Min, Feng
2015-12-01
The most important step in image preprocessing for Optical Character Recognition (OCR) is binarization. Due to the complex background or varying light in the text image, binarization is a very difficult problem. This paper presents the improved binarization algorithm. The algorithm can be divided into several steps. First, the background approximation can be obtained by the polynomial fitting, and the text is sharpened by using bilateral filter. Second, the image contrast compensation is done to reduce the impact of light and improve contrast of the original image. Third, the first derivative of the pixels in the compensated image are calculated to get the average value of the threshold, then the edge detection is obtained. Fourth, the stroke width of the text is estimated through a measuring of distance between edge pixels. The final stroke width is determined by choosing the most frequent distance in the histogram. Fifth, according to the value of the final stroke width, the window size is calculated, then a local threshold estimation approach can begin to binaries the image. Finally, the small noise is removed based on the morphological operators. The experimental result shows that the proposed method can effectively remove the noise caused by complex background and varying light.
A New Pivoting and Iterative Text Detection Algorithm for Biomedical Images
Xu, Songhua; Krauthammer, Michael
2010-01-01
There is interest to expand the reach of literature mining to include the analysis of biomedical images, which often contain a paper’s key findings. Examples include recent studies that use Optical Character Recognition (OCR) to extract image text, which is used to boost biomedical image retrieval and classification. Such studies rely on the robust identification of text elements in biomedical images, which is a non-trivial task. In this work, we introduce a new text detection algorithm for biomedical images based on iterative projection histograms. We study the effectiveness of our algorithm by evaluating the performance on a set of manually labeled random biomedical images, and compare the performance against other state-of-the-art text detection algorithms. In this paper, we demonstrate that a projection histogram-based text detection approach is well suited for text detection in biomedical images, with a performance of F score of .60. The approach performs better than comparable approaches for text detection. Further, we show that the iterative application of the algorithm is boosting overall detection performance. A C++ implementation of our algorithm is freely available through email request for academic use. PMID:20887803
Indexed Captioned Searchable Videos: A Learning Companion for STEM Coursework
NASA Astrophysics Data System (ADS)
Tuna, Tayfun; Subhlok, Jaspal; Barker, Lecia; Shah, Shishir; Johnson, Olin; Hovey, Christopher
2017-02-01
Videos of classroom lectures have proven to be a popular and versatile learning resource. A key shortcoming of the lecture video format is accessing the content of interest hidden in a video. This work meets this challenge with an advanced video framework featuring topical indexing, search, and captioning (ICS videos). Standard optical character recognition (OCR) technology was enhanced with image transformations for extraction of text from video frames to support indexing and search. The images and text on video frames is analyzed to divide lecture videos into topical segments. The ICS video player integrates indexing, search, and captioning in video playback providing instant access to the content of interest. This video framework has been used by more than 70 courses in a variety of STEM disciplines and assessed by more than 4000 students. Results presented from the surveys demonstrate the value of the videos as a learning resource and the role played by videos in a students learning process. Survey results also establish the value of indexing and search features in a video platform for education. This paper reports on the development and evaluation of ICS videos framework and over 5 years of usage experience in several STEM courses.
Mathematical morphology-based shape feature analysis for Chinese character recognition systems
NASA Astrophysics Data System (ADS)
Pai, Tun-Wen; Shyu, Keh-Hwa; Chen, Ling-Fan; Tai, Gwo-Chin
1995-04-01
This paper proposes an efficient technique of shape feature extraction based on the application of mathematical morphology theory. A new shape complexity index for preclassification of machine printed Chinese Character Recognition (CCR) is also proposed. For characters represented in different fonts/sizes or in a low resolution environment, a more stable local feature such as shape structure is preferred for character recognition. Morphological valley extraction filters are applied to extract the protrusive strokes from four sides of an input Chinese character. The number of extracted local strokes reflects the shape complexity of each side. These shape features of characters are encoded as corresponding shape complexity indices. Based on the shape complexity index, data base is able to be classified into 16 groups prior to recognition procedures. The performance of associating with shape feature analysis reclaims several characters from misrecognized character sets and results in an average of 3.3% improvement of recognition rate from an existing recognition system. In addition to enhance the recognition performance, the extracted stroke information can be further analyzed and classified its own stroke type. Therefore, the combination of extracted strokes from each side provides a means for data base clustering based on radical or subword components. It is one of the best solutions for recognizing high complexity characters such as Chinese characters which are divided into more than 200 different categories and consist more than 13,000 characters.
Fast approach for toner saving
NASA Astrophysics Data System (ADS)
Safonov, Ilia V.; Kurilin, Ilya V.; Rychagov, Michael N.; Lee, Hokeun; Kim, Sangho; Choi, Donchul
2011-01-01
Reducing toner consumption is an important task in modern printing devices and has a significant positive ecological impact. Existing toner saving approaches have two main drawbacks: appearance of hardcopy in toner saving mode is worse in comparison with normal mode; processing of whole rendered page bitmap requires significant computational costs. We propose to add small holes of various shapes and sizes to random places inside a character bitmap stored in font cache. Such random perforation scheme is based on processing pipeline in RIP of standard printer languages Postscript and PCL. Processing of text characters only, and moreover, processing of each character for given font and size alone, is an extremely fast procedure. The approach does not deteriorate halftoned bitmap and business graphics and provide toner saving for typical office documents up to 15-20%. Rate of toner saving is adjustable. Alteration of resulted characters' appearance is almost indistinguishable in comparison with solid black text due to random placement of small holes inside the character regions. The suggested method automatically skips small fonts to preserve its quality. Readability of text processed by proposed method is fine. OCR programs process that scanned hardcopy successfully too.
Portable Language-Independent Adaptive Translation from OCR. Phase 1
2009-04-01
including brute-force k-Nearest Neighbors ( kNN ), fast approximate kNN using hashed k-d trees, classification and regression trees, and locality...achieved by refinements in ground-truthing protocols. Recent algorithmic improvements to our approximate kNN classifier using hashed k-D trees allows...recent years discriminative training has been shown to outperform phonetic HMMs estimated using ML for speech recognition. Standard ML estimation
Rapid Naming Speed and Chinese Character Recognition
ERIC Educational Resources Information Center
Liao, Chen-Huei; Georgiou, George K.; Parrila, Rauno
2008-01-01
We examined the relationship between rapid naming speed (RAN) and Chinese character recognition accuracy and fluency. Sixty-three grade 2 and 54 grade 4 Taiwanese children were administered four RAN tasks (colors, digits, Zhu-Yin-Fu-Hao, characters), and two character recognition tasks. RAN tasks accounted for more reading variance in grade 4 than…
Real-Time Pattern Recognition - An Industrial Example
NASA Astrophysics Data System (ADS)
Fitton, Gary M.
1981-11-01
Rapid advancements in cost effective sensors and micro computers are now making practical the on-line implementation of pattern recognition based systems for a variety of industrial applications requiring high processing speeds. One major application area for real time pattern recognition is in the sorting of packaged/cartoned goods at high speed for automated warehousing and return goods cataloging. While there are many OCR and bar code readers available to perform these functions, it is often impractical to use such codes (package too small, adverse esthetics, poor print quality) and an approach which recognizes an item by its graphic content alone is desirable. This paper describes a specific application within the tobacco industry, that of sorting returned cigarette goods by brand and size.
Prediction of Marginal Mass Required for Successful Islet Transplantation
Papas, Klearchos K.; Colton, Clark K.; Qipo, Andi; Wu, Haiyan; Nelson, Rebecca A.; Hering, Bernhard J.; Weir, Gordon C.; Koulmanda, Maria
2013-01-01
Islet quality assessment methods for predicting diabetes reversal (DR) following transplantation are needed. We investigated two islet parameters, oxygen consumption rate (OCR) and OCR per DNA content, to predict transplantation outcome and explored the impact of islet quality on marginal islet mass for DR. Outcomes in immunosuppressed diabetic mice were evaluated by transplanting mixtures of healthy and purposely damaged rat islets for systematic variation of OCR/DNA over a wide range. The probability of DR increased with increasing transplanted OCR and OCR/DNA. On coordinates of OCR versus OCR/DNA, data fell into regions in which DR occurred in all, some, or none of the animals with a sharp threshold of around 150-nmol/min mg DNA. A model incorporating both parameters predicted transplantation outcome with sensitivity and specificity of 93% and 94%, respectively. Marginal mass was not constant, depended on OCR/DNA, and increased from 2,800 to over 100,000 islet equivalents/kg body weight as OCR/DNA decreased. We conclude that measurements of OCR and OCR/DNA are useful for predicting transplantation outcome in this model system, and OCR/DNA can be used to estimate the marginal mass required for reversing diabetes. Because human clinical islet preparations in a previous study had OCR/DNA values in the range of 100–150-nmol/min mg DNA, our findings suggest that substantial improvement in transplantation outcome may accompany increasedOCR/DNAin clinical islet preparations. PMID:20233002
Sub-word image clustering in Farsi printed books
NASA Astrophysics Data System (ADS)
Soheili, Mohammad Reza; Kabir, Ehsanollah; Stricker, Didier
2015-02-01
Most OCR systems are designed for the recognition of a single page. In case of unfamiliar font faces, low quality papers and degraded prints, the performance of these products drops sharply. However, an OCR system can use redundancy of word occurrences in large documents to improve recognition results. In this paper, we propose a sub-word image clustering method for the applications dealing with large printed documents. We assume that the whole document is printed by a unique unknown font with low quality print. Our proposed method finds clusters of equivalent sub-word images with an incremental algorithm. Due to the low print quality, we propose an image matching algorithm for measuring the distance between two sub-word images, based on Hamming distance and the ratio of the area to the perimeter of the connected components. We built a ground-truth dataset of more than 111000 sub-word images to evaluate our method. All of these images were extracted from an old Farsi book. We cluster all of these sub-words, including isolated letters and even punctuation marks. Then all centers of created clusters are labeled manually. We show that all sub-words of the book can be recognized with more than 99.7% accuracy by assigning the label of each cluster center to all of its members.
Word-level recognition of multifont Arabic text using a feature vector matching approach
NASA Astrophysics Data System (ADS)
Erlandson, Erik J.; Trenkle, John M.; Vogt, Robert C., III
1996-03-01
Many text recognition systems recognize text imagery at the character level and assemble words from the recognized characters. An alternative approach is to recognize text imagery at the word level, without analyzing individual characters. This approach avoids the problem of individual character segmentation, and can overcome local errors in character recognition. A word-level recognition system for machine-printed Arabic text has been implemented. Arabic is a script language, and is therefore difficult to segment at the character level. Character segmentation has been avoided by recognizing text imagery of complete words. The Arabic recognition system computes a vector of image-morphological features on a query word image. This vector is matched against a precomputed database of vectors from a lexicon of Arabic words. Vectors from the database with the highest match score are returned as hypotheses for the unknown image. Several feature vectors may be stored for each word in the database. Database feature vectors generated using multiple fonts and noise models allow the system to be tuned to its input stream. Used in conjunction with database pruning techniques, this Arabic recognition system has obtained promising word recognition rates on low-quality multifont text imagery.
NASA Astrophysics Data System (ADS)
Kaur, Jaswinder; Jagdev, Gagandeep, Dr.
2018-01-01
Optical character recognition is concerned with the recognition of optically processed characters. The recognition is done offline after the writing or printing has been completed, unlike online recognition where the computer has to recognize the characters instantly as they are drawn. The performance of character recognition depends upon the quality of scanned documents. The preprocessing steps are used for removing low-frequency background noise and normalizing the intensity of individual scanned documents. Several filters are used for reducing certain image details and enabling an easier or faster evaluation. The primary aim of the research work is to recognize handwritten and machine written characters and differentiate them. The language opted for the research work is Punjabi Gurmukhi and tool utilized is Matlab.
Character context: a shape descriptor for Arabic handwriting recognition
NASA Astrophysics Data System (ADS)
Mudhsh, Mohammed; Almodfer, Rolla; Duan, Pengfei; Xiong, Shengwu
2017-11-01
In the handwriting recognition field, designing good descriptors are substantial to obtain rich information of the data. However, the handwriting recognition research of a good descriptor is still an open issue due to unlimited variation in human handwriting. We introduce a "character context descriptor" that efficiently dealt with the structural characteristics of Arabic handwritten characters. First, the character image is smoothed and normalized, then the character context descriptor of 32 feature bins is built based on the proposed "distance function." Finally, a multilayer perceptron with regularization is used as a classifier. On experimentation with a handwritten Arabic characters database, the proposed method achieved a state-of-the-art performance with recognition rate equal to 98.93% and 99.06% for the 66 and 24 classes, respectively.
Handwritten recognition of Tamil vowels using deep learning
NASA Astrophysics Data System (ADS)
Ram Prashanth, N.; Siddarth, B.; Ganesh, Anirudh; Naveen Kumar, Vaegae
2017-11-01
We come across a large volume of handwritten texts in our daily lives and handwritten character recognition has long been an important area of research in pattern recognition. The complexity of the task varies among different languages and it so happens largely due to the similarity between characters, distinct shapes and number of characters which are all language-specific properties. There have been numerous works on character recognition of English alphabets and with laudable success, but regional languages have not been dealt with very frequently and with similar accuracies. In this paper, we explored the performance of Deep Belief Networks in the classification of Handwritten Tamil vowels, and conclusively compared the results obtained. The proposed method has shown satisfactory recognition accuracy in light of difficulties faced with regional languages such as similarity between characters and minute nuances that differentiate them. We can further extend this to all the Tamil characters.
Optical character recognition based on nonredundant correlation measurements.
Braunecker, B; Hauck, R; Lohmann, A W
1979-08-15
The essence of character recognition is a comparison between the unknown character and a set of reference patterns. Usually, these reference patterns are all possible characters themselves, the whole alphabet in the case of letter characters. Obviously, N analog measurements are highly redundant, since only K = log(2)N binary decisions are enough to identify one out of N characters. Therefore, we devised K reference patterns accordingly. These patterns, called principal components, are found by digital image processing, but used in an optical analog computer. We will explain the concept of principal components, and we will describe experiments with several optical character recognition systems, based on this concept.
Children's Recognition of Cartoon Voices.
ERIC Educational Resources Information Center
Spence, Melanie J.; Rollins, Pamela R.; Jerger, Susan
2002-01-01
A study examined developmental changes in talker recognition skills by assessing 72 children's (ages 3-5) recognition of 20 cartoon characters' voices. Four- and 5-year-old children recognized more of the voices than did 3-year-olds. All children were more accurate at recognizing more familiar characters than less familiar characters. (Contains…
Online recognition of Chinese characters: the state-of-the-art.
Liu, Cheng-Lin; Jaeger, Stefan; Nakagawa, Masaki
2004-02-01
Online handwriting recognition is gaining renewed interest owing to the increase of pen computing applications and new pen input devices. The recognition of Chinese characters is different from western handwriting recognition and poses a special challenge. To provide an overview of the technical status and inspire future research, this paper reviews the advances in online Chinese character recognition (OLCCR), with emphasis on the research works from the 1990s. Compared to the research in the 1980s, the research efforts in the 1990s aimed to further relax the constraints of handwriting, namely, the adherence to standard stroke orders and stroke numbers and the restriction of recognition to isolated characters only. The target of recognition has shifted from regular script to fluent script in order to better meet the requirements of practical applications. The research works are reviewed in terms of pattern representation, character classification, learning/adaptation, and contextual processing. We compare important results and discuss possible directions of future research.
Recognition of Telugu characters using neural networks.
Sukhaswami, M B; Seetharamulu, P; Pujari, A K
1995-09-01
The aim of the present work is to recognize printed and handwritten Telugu characters using artificial neural networks (ANNs). Earlier work on recognition of Telugu characters has been done using conventional pattern recognition techniques. We make an initial attempt here of using neural networks for recognition with the aim of improving upon earlier methods which do not perform effectively in the presence of noise and distortion in the characters. The Hopfield model of neural network working as an associative memory is chosen for recognition purposes initially. Due to limitation in the capacity of the Hopfield neural network, we propose a new scheme named here as the Multiple Neural Network Associative Memory (MNNAM). The limitation in storage capacity has been overcome by combining multiple neural networks which work in parallel. It is also demonstrated that the Hopfield network is suitable for recognizing noisy printed characters as well as handwritten characters written by different "hands" in a variety of styles. Detailed experiments have been carried out using several learning strategies and results are reported. It is shown here that satisfactory recognition is possible using the proposed strategy. A detailed preprocessing scheme of the Telugu characters from digitized documents is also described.
ERIC Educational Resources Information Center
Hsiao, Janet H.; Cheung, Kit
2016-01-01
In Chinese orthography, the most common character structure consists of a semantic radical on the left and a phonetic radical on the right (SP characters); the minority, opposite arrangement also exists (PS characters). Recent studies showed that SP character processing is more left hemisphere (LH) lateralized than PS character processing.…
1988-04-01
much less for characters than for other objects, and 200 0.98 thus OCR appears to represent a worst-case guideline. 400 (10 To quantify the effect of...mm, yma = 5 mm, , = 0.6328 mn, and J = 400 mm. the effect of the positional weighting factor given b From Eq. (6), we then find a > 23 line pairs/mm...is the square radius r2 = (x2 + yN2) of each input point i required. We used n = 400 fringes in Eq. (4) for a = 40 Po. The effect of the r2 = (X2 + y2
Increasing the efficiency of digitization workflows for herbarium specimens.
Tulig, Melissa; Tarnowsky, Nicole; Bevans, Michael; Anthony Kirchgessner; Thiers, Barbara M
2012-01-01
The New York Botanical Garden Herbarium has been databasing and imaging its estimated 7.3 million plant specimens for the past 17 years. Due to the size of the collection, we have been selectively digitizing fundable subsets of specimens, making successive passes through the herbarium with each new grant. With this strategy, the average rate for databasing complete records has been 10 specimens per hour. With 1.3 million specimens databased, this effort has taken about 130,000 hours of staff time. At this rate, to complete the herbarium and digitize the remaining 6 million specimens, another 600,000 hours would be needed. Given the current biodiversity and economic crises, there is neither the time nor money to complete the collection at this rate.Through a combination of grants over the last few years, The New York Botanical Garden has been testing new protocols and tactics for increasing the rate of digitization through combinations of data collaboration, field book digitization, partial data entry and imaging, and optical character recognition (OCR) of specimen images. With the launch of the National Science Foundation's new Advancing Digitization of Biological Collections program, we hope to move forward with larger, more efficient digitization projects, capturing data from larger portions of the herbarium at a fraction of the cost and time.
Haston, Elspeth; Cubey, Robert; Pullan, Martin; Atkins, Hannah; Harris, David J
2012-01-01
Digitisation programmes in many institutes frequently involve disparate and irregular funding, diverse selection criteria and scope, with different members of staff managing and operating the processes. These factors have influenced the decision at the Royal Botanic Garden Edinburgh to develop an integrated workflow for the digitisation of herbarium specimens which is modular and scalable to enable a single overall workflow to be used for all digitisation projects. This integrated workflow is comprised of three principal elements: a specimen workflow, a data workflow and an image workflow.The specimen workflow is strongly linked to curatorial processes which will impact on the prioritisation, selection and preparation of the specimens. The importance of including a conservation element within the digitisation workflow is highlighted. The data workflow includes the concept of three main categories of collection data: label data, curatorial data and supplementary data. It is shown that each category of data has its own properties which influence the timing of data capture within the workflow. Development of software has been carried out for the rapid capture of curatorial data, and optical character recognition (OCR) software is being used to increase the efficiency of capturing label data and supplementary data. The large number and size of the images has necessitated the inclusion of automated systems within the image workflow.
Selecting a restoration technique to minimize OCR error.
Cannon, M; Fugate, M; Hush, D R; Scovel, C
2003-01-01
This paper introduces a learning problem related to the task of converting printed documents to ASCII text files. The goal of the learning procedure is to produce a function that maps documents to restoration techniques in such a way that on average the restored documents have minimum optical character recognition error. We derive a general form for the optimal function and use it to motivate the development of a nonparametric method based on nearest neighbors. We also develop a direct method of solution based on empirical error minimization for which we prove a finite sample bound on estimation error that is independent of distribution. We show that this empirical error minimization problem is an extension of the empirical optimization problem for traditional M-class classification with general loss function and prove computational hardness for this problem. We then derive a simple iterative algorithm called generalized multiclass ratchet (GMR) and prove that it produces an optimal function asymptotically (with probability 1). To obtain the GMR algorithm we introduce a new data map that extends Kesler's construction for the multiclass problem and then apply an algorithm called Ratchet to this mapped data, where Ratchet is a modification of the Pocket algorithm . Finally, we apply these methods to a collection of documents and report on the experimental results.
Increasing the efficiency of digitization workflows for herbarium specimens
Tulig, Melissa; Tarnowsky, Nicole; Bevans, Michael; Anthony Kirchgessner; Thiers, Barbara M.
2012-01-01
Abstract The New York Botanical Garden Herbarium has been databasing and imaging its estimated 7.3 million plant specimens for the past 17 years. Due to the size of the collection, we have been selectively digitizing fundable subsets of specimens, making successive passes through the herbarium with each new grant. With this strategy, the average rate for databasing complete records has been 10 specimens per hour. With 1.3 million specimens databased, this effort has taken about 130,000 hours of staff time. At this rate, to complete the herbarium and digitize the remaining 6 million specimens, another 600,000 hours would be needed. Given the current biodiversity and economic crises, there is neither the time nor money to complete the collection at this rate. Through a combination of grants over the last few years, The New York Botanical Garden has been testing new protocols and tactics for increasing the rate of digitization through combinations of data collaboration, field book digitization, partial data entry and imaging, and optical character recognition (OCR) of specimen images. With the launch of the National Science Foundation’s new Advancing Digitization of Biological Collections program, we hope to move forward with larger, more efficient digitization projects, capturing data from larger portions of the herbarium at a fraction of the cost and time. PMID:22859882
Looking at 3,000,000 References Without Growing Grey Hair
NASA Astrophysics Data System (ADS)
Demleitner, M.; Accomazzi, A.; Eichhorn, G.; Grant, C. S.; Kurtz, M. J.; Murray, S. S.
1999-12-01
The article service of the Astrophysics Data System (ADS, http://adswww.harvard.edu) currently holds about 500,000 pages scanned from astronomical journals and conference proceedings. This data set not only facilitates an easy and convenient access to the majority of the astronomical literature from anywhere on the Internet but also allows highly automatized extraction of the information contained in the articles. As first steps towards processing and indexing the full texts of the articles, the ADS has been extracting abstracts and references from the bitmap images of the articles since May 1999. In this poster we describe the procedures and strategies to (a) automatically identify the regions within a paper containing the abstract or the references, (b) spot and correct errors in the data base or the identification of the regions, (c) resolve references obtained by optical character recognition (OCR) with its inherent uncertainties to parsed references (i.e., bibcodes) and (d) incorporate the data collected in this way into the ADS abstract service. We also give an overview of the extent of additional bibliographical material from this source. We estimate that by January 2000, these procedures will have yielded about 14,000 abstracts and 1,000,000 citation pairs (out of a total of 3,000,000 references) not previously present in the ADS.
CCD imaging technology and the war on crime
NASA Astrophysics Data System (ADS)
McNeill, Glenn E.
1992-08-01
Linear array based CCD technology has been successfully used in the development of an Automatic Currency Reader/Comparator (ACR/C) system. The ACR/C system is designed to provide a method for tracking US currency in the organized crime and drug trafficking environments where large amounts of cash are involved in illegal transactions and money laundering activities. United States currency notes can be uniquely identified by the combination of the denomination serial number and series year. The ACR/C system processes notes at five notes per second using a custom transport a stationary linear array and optical character recognition (OCR) techniques to make such identifications. In this way large sums of money can be " marked" (using the system to read and store their identifiers) and then circulated within various crime networks. The system can later be used to read and compare confiscated notes to the known sets of identifiers from the " marked" set to document a trail of criminal activities. With the ACR/C law enforcement agencies can efficiently identify currency without actually marking it. This provides an undetectable means for making each note individually traceable and facilitates record keeping for providing evidence in a court of law. In addition when multiple systems are used in conjunction with a central data base the system can be used to track currency geographically. 1.
Line Segmentation in Handwritten Assamese and Meetei Mayek Script Using Seam Carving Based Algorithm
NASA Astrophysics Data System (ADS)
Kumar, Chandan Jyoti; Kalita, Sanjib Kr.
Line segmentation is a key stage in an Optical Character Recognition system. This paper primarily concerns the problem of text line extraction on color and grayscale manuscript pages of two major North-east Indian regional Scripts, Assamese and Meetei Mayek. Line segmentation of handwritten text in Assamese and Meetei Mayek scripts is an uphill task primarily because of the structural features of both the scripts and varied writing styles. Line segmentation of a document image is been achieved by using the Seam carving technique, in this paper. Researchers from various regions used this approach for content aware resizing of an image. However currently many researchers are implementing Seam Carving for line segmentation phase of OCR. Although it is a language independent technique, mostly experiments are done over Arabic, Greek, German and Chinese scripts. Two types of seams are generated, medial seams approximate the orientation of each text line, and separating seams separated one line of text from another. Experiments are performed extensively over various types of documents and detailed analysis of the evaluations reflects that the algorithm performs well for even documents with multiple scripts. In this paper, we present a comparative study of accuracy of this method over different types of data.
Common constraints limit Korean and English character recognition in peripheral vision.
He, Yingchen; Kwon, MiYoung; Legge, Gordon E
2018-01-01
The visual span refers to the number of adjacent characters that can be recognized in a single glance. It is viewed as a sensory bottleneck in reading for both normal and clinical populations. In peripheral vision, the visual span for English characters can be enlarged after training with a letter-recognition task. Here, we examined the transfer of training from Korean to English characters for a group of bilingual Korean native speakers. In the pre- and posttests, we measured visual spans for Korean characters and English letters. Training (1.5 hours × 4 days) consisted of repetitive visual-span measurements for Korean trigrams (strings of three characters). Our training enlarged the visual spans for Korean single characters and trigrams, and the benefit transferred to untrained English symbols. The improvement was largely due to a reduction of within-character and between-character crowding in Korean recognition, as well as between-letter crowding in English recognition. We also found a negative correlation between the size of the visual span and the average pattern complexity of the symbol set. Together, our results showed that the visual span is limited by common sensory (crowding) and physical (pattern complexity) factors regardless of the language script, providing evidence that the visual span reflects a universal bottleneck for text recognition.
Common constraints limit Korean and English character recognition in peripheral vision
He, Yingchen; Kwon, MiYoung; Legge, Gordon E.
2018-01-01
The visual span refers to the number of adjacent characters that can be recognized in a single glance. It is viewed as a sensory bottleneck in reading for both normal and clinical populations. In peripheral vision, the visual span for English characters can be enlarged after training with a letter-recognition task. Here, we examined the transfer of training from Korean to English characters for a group of bilingual Korean native speakers. In the pre- and posttests, we measured visual spans for Korean characters and English letters. Training (1.5 hours × 4 days) consisted of repetitive visual-span measurements for Korean trigrams (strings of three characters). Our training enlarged the visual spans for Korean single characters and trigrams, and the benefit transferred to untrained English symbols. The improvement was largely due to a reduction of within-character and between-character crowding in Korean recognition, as well as between-letter crowding in English recognition. We also found a negative correlation between the size of the visual span and the average pattern complexity of the symbol set. Together, our results showed that the visual span is limited by common sensory (crowding) and physical (pattern complexity) factors regardless of the language script, providing evidence that the visual span reflects a universal bottleneck for text recognition. PMID:29327041
Character Recognition Using Genetically Trained Neural Networks
DOE Office of Scientific and Technical Information (OSTI.GOV)
Diniz, C.; Stantz, K.M.; Trahan, M.W.
1998-10-01
Computationally intelligent recognition of characters and symbols addresses a wide range of applications including foreign language translation and chemical formula identification. The combination of intelligent learning and optimization algorithms with layered neural structures offers powerful techniques for character recognition. These techniques were originally developed by Sandia National Laboratories for pattern and spectral analysis; however, their ability to optimize vast amounts of data make them ideal for character recognition. An adaptation of the Neural Network Designer soflsvare allows the user to create a neural network (NN_) trained by a genetic algorithm (GA) that correctly identifies multiple distinct characters. The initial successfidmore » recognition of standard capital letters can be expanded to include chemical and mathematical symbols and alphabets of foreign languages, especially Arabic and Chinese. The FIN model constructed for this project uses a three layer feed-forward architecture. To facilitate the input of characters and symbols, a graphic user interface (GUI) has been developed to convert the traditional representation of each character or symbol to a bitmap. The 8 x 8 bitmap representations used for these tests are mapped onto the input nodes of the feed-forward neural network (FFNN) in a one-to-one correspondence. The input nodes feed forward into a hidden layer, and the hidden layer feeds into five output nodes correlated to possible character outcomes. During the training period the GA optimizes the weights of the NN until it can successfully recognize distinct characters. Systematic deviations from the base design test the network's range of applicability. Increasing capacity, the number of letters to be recognized, requires a nonlinear increase in the number of hidden layer neurodes. Optimal character recognition performance necessitates a minimum threshold for the number of cases when genetically training the net. And, the amount of noise significantly degrades character recognition efficiency, some of which can be overcome by adding noise during training and optimizing the form of the network's activation fimction.« less
Building Hierarchical Representations for Oracle Character and Sketch Recognition.
Jun Guo; Changhu Wang; Roman-Rangel, Edgar; Hongyang Chao; Yong Rui
2016-01-01
In this paper, we study oracle character recognition and general sketch recognition. First, a data set of oracle characters, which are the oldest hieroglyphs in China yet remain a part of modern Chinese characters, is collected for analysis. Second, typical visual representations in shape- and sketch-related works are evaluated. We analyze the problems suffered when addressing these representations and determine several representation design criteria. Based on the analysis, we propose a novel hierarchical representation that combines a Gabor-related low-level representation and a sparse-encoder-related mid-level representation. Extensive experiments show the effectiveness of the proposed representation in both oracle character recognition and general sketch recognition. The proposed representation is also complementary to convolutional neural network (CNN)-based models. We introduce a solution to combine the proposed representation with CNN-based models, and achieve better performances over both approaches. This solution has beaten humans at recognizing general sketches.
Shinozaki, Takahiro
2018-01-01
Human-computer interface systems whose input is based on eye movements can serve as a means of communication for patients with locked-in syndrome. Eye-writing is one such system; users can input characters by moving their eyes to follow the lines of the strokes corresponding to characters. Although this input method makes it easy for patients to get started because of their familiarity with handwriting, existing eye-writing systems suffer from slow input rates because they require a pause between input characters to simplify the automatic recognition process. In this paper, we propose a continuous eye-writing recognition system that achieves a rapid input rate because it accepts characters eye-written continuously, with no pauses. For recognition purposes, the proposed system first detects eye movements using electrooculography (EOG), and then a hidden Markov model (HMM) is applied to model the EOG signals and recognize the eye-written characters. Additionally, this paper investigates an EOG adaptation that uses a deep neural network (DNN)-based HMM. Experiments with six participants showed an average input speed of 27.9 character/min using Japanese Katakana as the input target characters. A Katakana character-recognition error rate of only 5.0% was achieved using 13.8 minutes of adaptation data. PMID:29425248
Character recognition using a neural network model with fuzzy representation
NASA Technical Reports Server (NTRS)
Tavakoli, Nassrin; Seniw, David
1992-01-01
The degree to which digital images are recognized correctly by computerized algorithms is highly dependent upon the representation and the classification processes. Fuzzy techniques play an important role in both processes. In this paper, the role of fuzzy representation and classification on the recognition of digital characters is investigated. An experimental Neural Network model with application to character recognition was developed. Through a set of experiments, the effect of fuzzy representation on the recognition accuracy of this model is presented.
Papas, Klearchos K; Bellin, Melena D; Sutherland, David E R; Suszynski, Thomas M; Kitzmann, Jennifer P; Avgoustiniatos, Efstathios S; Gruessner, Angelika C; Mueller, Kathryn R; Beilman, Gregory J; Balamurugan, Appakalai N; Loganathan, Gopalakrishnan; Colton, Clark K; Koulmanda, Maria; Weir, Gordon C; Wilhelm, Josh J; Qian, Dajun; Niland, Joyce C; Hering, Bernhard J
2015-01-01
Reliable in vitro islet quality assessment assays that can be performed routinely, prospectively, and are able to predict clinical transplant outcomes are needed. In this paper we present data on the utility of an assay based on cellular oxygen consumption rate (OCR) in predicting clinical islet autotransplant (IAT) insulin independence (II). IAT is an attractive model for evaluating characterization assays regarding their utility in predicting II due to an absence of confounding factors such as immune rejection and immunosuppressant toxicity. Membrane integrity staining (FDA/PI), OCR normalized to DNA (OCR/DNA), islet equivalent (IE) and OCR (viable IE) normalized to recipient body weight (IE dose and OCR dose), and OCR/DNA normalized to islet size index (ISI) were used to characterize autoislet preparations (n = 35). Correlation between pre-IAT islet product characteristics and II was determined using receiver operating characteristic analysis. Preparations that resulted in II had significantly higher OCR dose and IE dose (p<0.001). These islet characterization methods were highly correlated with II at 6-12 months post-IAT (area-under-the-curve (AUC) = 0.94 for IE dose and 0.96 for OCR dose). FDA/PI (AUC = 0.49) and OCR/DNA (AUC = 0.58) did not correlate with II. OCR/DNA/ISI may have some utility in predicting outcome (AUC = 0.72). Commonly used assays to determine whether a clinical islet preparation is of high quality prior to transplantation are greatly lacking in sensitivity and specificity. While IE dose is highly predictive, it does not take into account islet cell quality. OCR dose, which takes into consideration both islet cell quality and quantity, may enable a more accurate and prospective evaluation of clinical islet preparations.
Principal Component 2-D Long Short-Term Memory for Font Recognition on Single Chinese Characters.
Tao, Dapeng; Lin, Xu; Jin, Lianwen; Li, Xuelong
2016-03-01
Chinese character font recognition (CCFR) has received increasing attention as the intelligent applications based on optical character recognition becomes popular. However, traditional CCFR systems do not handle noisy data effectively. By analyzing in detail the basic strokes of Chinese characters, we propose that font recognition on a single Chinese character is a sequence classification problem, which can be effectively solved by recurrent neural networks. For robust CCFR, we integrate a principal component convolution layer with the 2-D long short-term memory (2DLSTM) and develop principal component 2DLSTM (PC-2DLSTM) algorithm. PC-2DLSTM considers two aspects: 1) the principal component layer convolution operation helps remove the noise and get a rational and complete font information and 2) simultaneously, 2DLSTM deals with the long-range contextual processing along scan directions that can contribute to capture the contrast between character trajectory and background. Experiments using the frequently used CCFR dataset suggest the effectiveness of PC-2DLSTM compared with other state-of-the-art font recognition methods.
Sunspot drawings handwritten character recognition method based on deep learning
NASA Astrophysics Data System (ADS)
Zheng, Sheng; Zeng, Xiangyun; Lin, Ganghua; Zhao, Cui; Feng, Yongli; Tao, Jinping; Zhu, Daoyuan; Xiong, Li
2016-05-01
High accuracy scanned sunspot drawings handwritten characters recognition is an issue of critical importance to analyze sunspots movement and store them in the database. This paper presents a robust deep learning method for scanned sunspot drawings handwritten characters recognition. The convolution neural network (CNN) is one algorithm of deep learning which is truly successful in training of multi-layer network structure. CNN is used to train recognition model of handwritten character images which are extracted from the original sunspot drawings. We demonstrate the advantages of the proposed method on sunspot drawings provided by Chinese Academy Yunnan Observatory and obtain the daily full-disc sunspot numbers and sunspot areas from the sunspot drawings. The experimental results show that the proposed method achieves a high recognition accurate rate.
NASA Astrophysics Data System (ADS)
Hagita, Norihiro; Sawaki, Minako
1995-03-01
Most conventional methods in character recognition extract geometrical features such as stroke direction, connectivity of strokes, etc., and compare them with reference patterns in a stored dictionary. Unfortunately, geometrical features are easily degraded by blurs, stains and the graphical background designs used in Japanese newspaper headlines. This noise must be removed before recognition commences, but no preprocessing method is completely accurate. This paper proposes a method for recognizing degraded characters and characters printed on graphical background designs. This method is based on the binary image feature method and uses binary images as features. A new similarity measure, called the complementary similarity measure, is used as a discriminant function. It compares the similarity and dissimilarity of binary patterns with reference dictionary patterns. Experiments are conducted using the standard character database ETL-2 which consists of machine-printed Kanji, Hiragana, Katakana, alphanumeric, an special characters. The results show that this method is much more robust against noise than the conventional geometrical feature method. It also achieves high recognition rates of over 92% for characters with textured foregrounds, over 98% for characters with textured backgrounds, over 98% for outline fonts, and over 99% for reverse contrast characters.
Ocular Counter Rolling in Astronauts After Short- and Long-Duration Spaceflight.
Reschke, Millard F; Wood, Scott J; Clément, Gilles
2018-05-17
Ocular counter-rolling (OCR) is a reflex generated by the activation of the gravity sensors in the inner ear that stabilizes gaze and posture during head tilt. We compared the OCR measures that were obtained in 6 astronauts before, during, and after a spaceflight lasting 4-6 days with the OCR measures obtained from 6 astronauts before and after a spaceflight lasting 4-9 months. OCR in the short-duration fliers was measured using the afterimage method during head tilt at 15°, 30°, and 45°. OCR in the long-duration fliers was measured using video-oculography during whole body tilt at 25°. A control group of 7 subjects was used to compare OCR measures during head tilt and whole body tilt. No OCR occurred during head tilt in microgravity, and the response returned to normal within 2 hours of return from short-duration spaceflight. However, the amplitude of OCR was reduced for several days after return from long-duration spaceflight. This decrease in amplitude was not accompanied by changes in the asymmetry of OCR between right and left head tilt. These results indicate that the adaptation of otolith-driven reflexes to microgravity is a long-duration process.
Efficient automatic OCR word validation using word partial format derivation and language model
NASA Astrophysics Data System (ADS)
Chen, Siyuan; Misra, Dharitri; Thoma, George R.
2010-01-01
In this paper we present an OCR validation module, implemented for the System for Preservation of Electronic Resources (SPER) developed at the U.S. National Library of Medicine.1 The module detects and corrects suspicious words in the OCR output of scanned textual documents through a procedure of deriving partial formats for each suspicious word, retrieving candidate words by partial-match search from lexicons, and comparing the joint probabilities of N-gram and OCR edit transformation corresponding to the candidates. The partial format derivation, based on OCR error analysis, efficiently and accurately generates candidate words from lexicons represented by ternary search trees. In our test case comprising a historic medico-legal document collection, this OCR validation module yielded the correct words with 87% accuracy and reduced the overall OCR word errors by around 60%.
Pc-based car license plate reading
NASA Astrophysics Data System (ADS)
Tanabe, Katsuyoshi; Marubayashi, Eisaku; Kawashima, Harumi; Nakanishi, Tadashi; Shio, Akio
1994-03-01
A PC-based car license plate recognition system has been developed. The system recognizes Chinese characters and Japanese phonetic hiragana characters as well as six digits on Japanese license plates. The system consists of a CCD camera, vehicle sensors, a strobe unit, a monitoring center, and an i486-based PC. The PC includes in its extension slots: a vehicle detector board, a strobe emitter board, and an image grabber board. When a passing vehicle is detected by the vehicle sensors, the strobe emits a pulse of light. The light pulse is synchronized with the time the vehicle image is frozen on an image grabber board. The recognition process is composed of three steps: image thresholding, character region extraction, and matching-based character recognition. The recognition software can handle obscured characters. Experimental results for hundreds of outdoor images showed high recognition performance within relatively short performance times. The results confirmed that the system is applicable to a wide variety of applications such as automatic vehicle identification and travel time measurement.
A comparison of 1D and 2D LSTM architectures for the recognition of handwritten Arabic
NASA Astrophysics Data System (ADS)
Yousefi, Mohammad Reza; Soheili, Mohammad Reza; Breuel, Thomas M.; Stricker, Didier
2015-01-01
In this paper, we present an Arabic handwriting recognition method based on recurrent neural network. We use the Long Short Term Memory (LSTM) architecture, that have proven successful in different printed and handwritten OCR tasks. Applications of LSTM for handwriting recognition employ the two-dimensional architecture to deal with the variations in both vertical and horizontal axis. However, we show that using a simple pre-processing step that normalizes the position and baseline of letters, we can make use of 1D LSTM, which is faster in learning and convergence, and yet achieve superior performance. In a series of experiments on IFN/ENIT database for Arabic handwriting recognition, we demonstrate that our proposed pipeline can outperform 2D LSTM networks. Furthermore, we provide comparisons with 1D LSTM networks trained with manually crafted features to show that the automatically learned features in a globally trained 1D LSTM network with our normalization step can even outperform such systems.
NASA Astrophysics Data System (ADS)
Megherbi, Dalila B.; Lodhi, S. M.; Boulenouar, A. J.
2001-03-01
This work is in the field of automated document processing. This work addresses the problem of representation and recognition of Urdu characters using Fourier representation and a Neural Network architecture. In particular, we show that a two-stage Neural Network scheme is used here to make classification of 36 Urdu characters into seven sub-classes namely subclasses characterized by seven proposed and defined fuzzy features specifically related to Urdu characters. We show that here Fourier Descriptors and Neural Network provide a remarkably simple way to draw definite conclusions from vague, ambiguous, noisy or imprecise information. In particular, we illustrate the concept of interest regions and describe a framing method that provides a way to make the proposed technique for Urdu characters recognition robust and invariant to scaling and translation. We also show that a given character rotation is dealt with by using the Hotelling transform. This transform is based upon the eigenvalue decomposition of the covariance matrix of an image, providing a method of determining the orientation of the major axis of an object within an image. Finally experimental results are presented to show the power and robustness of the proposed two-stage Neural Network based technique for Urdu character recognition, its fault tolerance, and high recognition accuracy.
Automated recognition and extraction of tabular fields for the indexing of census records
NASA Astrophysics Data System (ADS)
Clawson, Robert; Bauer, Kevin; Chidester, Glen; Pohontsch, Milan; Kennard, Douglas; Ryu, Jongha; Barrett, William
2013-01-01
We describe a system for indexing of census records in tabular documents with the goal of recognizing the content of each cell, including both headers and handwritten entries. Each document is automatically rectified, registered and scaled to a known template following which lines and fields are detected and delimited as cells in a tabular form. Whole-word or whole-phrase recognition of noisy machine-printed text is performed using a glyph library, providing greatly increased efficiency and accuracy (approaching 100%), while avoiding the problems inherent with traditional OCR approaches. Constrained handwriting recognition results for a single author reach as high as 98% and 94.5% for the Gender field and Birthplace respectively. Multi-author accuracy (currently 82%) can be improved through an increased training set. Active integration of user feedback in the system will accelerate the indexing of records while providing a tightly coupled learning mechanism for system improvement.
Recognition of Similar Shaped Handwritten Marathi Characters Using Artificial Neural Network
NASA Astrophysics Data System (ADS)
Jane, Archana P.; Pund, Mukesh A.
2012-03-01
The growing need have handwritten Marathi character recognition in Indian offices such as passport, railways etc has made it vital area of a research. Similar shape characters are more prone to misclassification. In this paper a novel method is provided to recognize handwritten Marathi characters based on their features extraction and adaptive smoothing technique. Feature selections methods avoid unnecessary patterns in an image whereas adaptive smoothing technique form smooth shape of charecters.Combination of both these approaches leads to the better results. Previous study shows that, no one technique achieves 100% accuracy in handwritten character recognition area. This approach of combining both adaptive smoothing & feature extraction gives better results (approximately 75-100) and expected outcomes.
NASA Astrophysics Data System (ADS)
Kitagawa, Etsuji; Tanaka, Shigenori; Abiko, Satoshi; Wakabayashi, Katsuma; Jiang, Wenyuan
Recently, an electronic delivery for various documents is carried out by Ministry of Land, Infrastructure, Transport and Tourism in construction fields. One of them is image data of construction photography that must be delivered with information of photography management items such as construction name or type of works, etc. However, there is a problem that a lot of cost is needed to treat contents of these items from characters printed and handwritten on blackboard into these image data. In this research, we develop the system which can treat contents of these items by extracting contents of these items from the image data of construction photography taken in various scenes with preprocessing the image, recognizing characters with OCR and correcting error with natural language process. And we confirm the effectiveness of the system, by experimenting in each function of system and in entire system.
Early Reading Development in Chinese-Speaking Children with Hearing Loss
ERIC Educational Resources Information Center
Chan, Yi-Chih; Yang, You-Jhen
2018-01-01
This study aims to explore early reading comprehension in Chinese-speaking children with hearing loss (HL) by examining character recognition and linguistic comprehension. Twenty-five children with HL received three measures relevant to character reading: phonological awareness (PA), morphological awareness (MA), and character recognition; two…
Automated document analysis system
NASA Astrophysics Data System (ADS)
Black, Jeffrey D.; Dietzel, Robert; Hartnett, David
2002-08-01
A software application has been developed to aid law enforcement and government intelligence gathering organizations in the translation and analysis of foreign language documents with potential intelligence content. The Automated Document Analysis System (ADAS) provides the capability to search (data or text mine) documents in English and the most commonly encountered foreign languages, including Arabic. Hardcopy documents are scanned by a high-speed scanner and are optical character recognized (OCR). Documents obtained in an electronic format bypass the OCR and are copied directly to a working directory. For translation and analysis, the script and the language of the documents are first determined. If the document is not in English, the document is machine translated to English. The documents are searched for keywords and key features in either the native language or translated English. The user can quickly review the document to determine if it has any intelligence content and whether detailed, verbatim human translation is required. The documents and document content are cataloged for potential future analysis. The system allows non-linguists to evaluate foreign language documents and allows for the quick analysis of a large quantity of documents. All document processing can be performed manually or automatically on a single document or a batch of documents.
Phung, V T; Khatri, M; Liland, K H; Slinde, E; Sørheim, O; Almøy, T; Saarem, K; Egelandsdal, B
2013-01-01
Animal and muscle characteristics were recorded for 41 cattle. The oxygen consumption rate (OCR) of M. semimembranosus was measured between 3.0-6.4h post mortem (PM3-6) and after 3 weeks in a vacuum pack at 4°C. Colour change measurements were performed following the 3 weeks using reflectance spectra (400-1,100 nm) and the colour coordinates L, a and b, with the samples being packaged in oxygen permeable film and stored at 4°C for 167 h. Significant individual animal differences in OCR at PM3-6 were found for mitochondrial complexes I and II. OCR of complex I declined with increased temperature and time PM, while residual oxygen-consuming side-reactions (ROX) did not. OCR of stored muscles was dominated by complex II respiration. A three-way regression between samples, colour variables collected upon air exposure and OCR of 3 weeks old fibres revealed a positive relationship between OCR and complex II activity and also between OCR and OCR(ROX). The presence of complex I and β-oxidation activities increased metmyoglobin formation. Copyright © 2012 Elsevier Ltd. All rights reserved.
Longcamp, Marieke; Boucard, Céline; Gilhodes, Jean-Claude; Anton, Jean-Luc; Roth, Muriel; Nazarian, Bruno; Velay, Jean-Luc
2008-05-01
Fast and accurate visual recognition of single characters is crucial for efficient reading. We explored the possible contribution of writing memory to character recognition processes. We evaluated the ability of adults to discriminate new characters from their mirror images after being taught how to produce the characters either by traditional pen-and-paper writing or with a computer keyboard. After training, we found stronger and longer lasting (several weeks) facilitation in recognizing the orientation of characters that had been written by hand compared to those typed. Functional magnetic resonance imaging recordings indicated that the response mode during learning is associated with distinct pathways during recognition of graphic shapes. Greater activity related to handwriting learning and normal letter identification was observed in several brain regions known to be involved in the execution, imagery, and observation of actions, in particular, the left Broca's area and bilateral inferior parietal lobules. Taken together, these results provide strong arguments in favor of the view that the specific movements memorized when learning how to write participate in the visual recognition of graphic shapes and letters.
MSL: Facilitating automatic and physical analysis of published scientific literature in PDF format.
Ahmed, Zeeshan; Dandekar, Thomas
2015-01-01
Published scientific literature contains millions of figures, including information about the results obtained from different scientific experiments e.g. PCR-ELISA data, microarray analysis, gel electrophoresis, mass spectrometry data, DNA/RNA sequencing, diagnostic imaging (CT/MRI and ultrasound scans), and medicinal imaging like electroencephalography (EEG), magnetoencephalography (MEG), echocardiography (ECG), positron-emission tomography (PET) images. The importance of biomedical figures has been widely recognized in scientific and medicine communities, as they play a vital role in providing major original data, experimental and computational results in concise form. One major challenge for implementing a system for scientific literature analysis is extracting and analyzing text and figures from published PDF files by physical and logical document analysis. Here we present a product line architecture based bioinformatics tool 'Mining Scientific Literature (MSL)', which supports the extraction of text and images by interpreting all kinds of published PDF files using advanced data mining and image processing techniques. It provides modules for the marginalization of extracted text based on different coordinates and keywords, visualization of extracted figures and extraction of embedded text from all kinds of biological and biomedical figures using applied Optimal Character Recognition (OCR). Moreover, for further analysis and usage, it generates the system's output in different formats including text, PDF, XML and images files. Hence, MSL is an easy to install and use analysis tool to interpret published scientific literature in PDF format.
Haston, Elspeth; Cubey, Robert; Pullan, Martin; Atkins, Hannah; Harris, David J
2012-01-01
Abstract Digitisation programmes in many institutes frequently involve disparate and irregular funding, diverse selection criteria and scope, with different members of staff managing and operating the processes. These factors have influenced the decision at the Royal Botanic Garden Edinburgh to develop an integrated workflow for the digitisation of herbarium specimens which is modular and scalable to enable a single overall workflow to be used for all digitisation projects. This integrated workflow is comprised of three principal elements: a specimen workflow, a data workflow and an image workflow. The specimen workflow is strongly linked to curatorial processes which will impact on the prioritisation, selection and preparation of the specimens. The importance of including a conservation element within the digitisation workflow is highlighted. The data workflow includes the concept of three main categories of collection data: label data, curatorial data and supplementary data. It is shown that each category of data has its own properties which influence the timing of data capture within the workflow. Development of software has been carried out for the rapid capture of curatorial data, and optical character recognition (OCR) software is being used to increase the efficiency of capturing label data and supplementary data. The large number and size of the images has necessitated the inclusion of automated systems within the image workflow. PMID:22859881
Comparison of the three optical platforms for measurement of cellular respiration.
Kondrashina, Alina V; Ogurtsov, Vladimir I; Papkovsky, Dmitri B
2015-01-01
We compared three optical platforms for measurement of cellular respiration: absolute oxygen consumption rates (OCRs) in hermetically sealed microcuvettes, relative OCRs measured in a 96-well plate with oil seal, and steady-state oxygenation of cells in an open 96-well plate. Using mouse embryonic fibroblasts cell line, the phosphorescent intracellular O2 probe MitoXpress-Intra, and time-resolved fluorescence reader, we determined algorithms for conversion of relative OCRs and cell oxygenation into absolute OCRs, thereby allowing simple high-throughput measurement of absolute OCR values. Copyright © 2014 Elsevier Inc. All rights reserved.
Quantum-Limited Image Recognition
1989-12-01
J. S. Bomba ,’Alpha-numeric character recognition using local operations,’ Fall Joint Comput. Conf., 218-224 (1959). 53. D. Barnea and H. Silverman...for Chapter 6 1. J. S. Bomba ,’Alpha-numeric character recognition using local operations,’ Fall Joint Comput. Conf., 218-224 (1959). 2. D. Bamea and H
ERIC Educational Resources Information Center
Liu, Tianyin; Chuk, Tin Yim; Yeh, Su-Ling; Hsiao, Janet H.
2016-01-01
Expertise in Chinese character recognition is marked by reduced holistic processing (HP), which depends mainly on writing rather than reading experience. Here we show that, while simplified and traditional Chinese readers demonstrated a similar level of HP when processing characters shared between the simplified and traditional scripts, simplified…
NASA Astrophysics Data System (ADS)
Xiong, Yan; Reichenbach, Stephen E.
1999-01-01
Understanding of hand-written Chinese characters is at such a primitive stage that models include some assumptions about hand-written Chinese characters that are simply false. So Maximum Likelihood Estimation (MLE) may not be an optimal method for hand-written Chinese characters recognition. This concern motivates the research effort to consider alternative criteria. Maximum Mutual Information Estimation (MMIE) is an alternative method for parameter estimation that does not derive its rationale from presumed model correctness, but instead examines the pattern-modeling problem in automatic recognition system from an information- theoretic point of view. The objective of MMIE is to find a set of parameters in such that the resultant model allows the system to derive from the observed data as much information as possible about the class. We consider MMIE for recognition of hand-written Chinese characters using on a simplified hidden Markov Random Field. MMIE provides improved performance improvement over MLE in this application.
Anderson, Christopher N; Grether, Gregory F
2010-02-22
In zones of sympatry between closely related species, species recognition errors in a competitive context can cause character displacement in agonistic signals and competitor recognition functions, just as species recognition errors in a mating context can cause character displacement in mating signals and mate recognition. These two processes are difficult to distinguish because the same traits can serve as both agonistic and mating signals. One solution is to test for sympatric shifts in recognition functions. We studied competitor recognition in Hetaerina damselflies by challenging territory holders with live tethered conspecific and heterospecific intruders. Heterospecific intruders elicited less aggression than conspecific intruders in species pairs with dissimilar wing coloration (H. occisa/H. titia, H. americana/H. titia) but not in species pairs with similar wing coloration (H. occisa/H. cruentata, H. americana/H. cruentata). Natural variation in the area of black wing pigmentation on H. titia intruders correlated negatively with heterospecific aggression. To directly examine the role of wing coloration, we blackened the wings of H. occisa or H. americana intruders and measured responses of conspecific territory holders. This treatment reduced territorial aggression at multiple sites where H. titia is present, but not at allopatric sites. These results provide strong evidence for agonistic character displacement.
Real-Time Detection and Reading of LED/LCD Displays for Visually Impaired Persons
Tekin, Ender; Coughlan, James M.; Shen, Huiying
2011-01-01
Modern household appliances, such as microwave ovens and DVD players, increasingly require users to read an LED or LCD display to operate them, posing a severe obstacle for persons with blindness or visual impairment. While OCR-enabled devices are emerging to address the related problem of reading text in printed documents, they are not designed to tackle the challenge of finding and reading characters in appliance displays. Any system for reading these characters must address the challenge of first locating the characters among substantial amounts of background clutter; moreover, poor contrast and the abundance of specular highlights on the display surface – which degrade the image in an unpredictable way as the camera is moved – motivate the need for a system that processes images at a few frames per second, rather than forcing the user to take several photos, each of which can take seconds to acquire and process, until one is readable. We describe a novel system that acquires video, detects and reads LED/LCD characters in real time, reading them aloud to the user with synthesized speech. The system has been implemented on both a desktop and a cell phone. Experimental results are reported on videos of display images, demonstrating the feasibility of the system. PMID:21804957
Word recognition using a lexicon constrained by first/last character decisions
NASA Astrophysics Data System (ADS)
Zhao, Sheila X.; Srihari, Sargur N.
1995-03-01
In lexicon based recognition of machine-printed word images, the size of the lexicon can be quite extensive. The recognition performance is closely related to the size of the lexicon. Recognition performance drops quickly when lexicon size increases. Here, we present an algorithm to improve the word recognition performance by reducing the size of the given lexicon. The algorithm utilizes the information provided by the first and last characters of a word to reduce the size of the given lexicon. Given a word image and a lexicon that contains the word in the image, the first and last characters are segmented and then recognized by a character classifier. The possible candidates based on the results given by the classifier are selected, which give us the sub-lexicon. Then a word shape analysis algorithm is applied to produce the final ranking of the given lexicon. The algorithm was tested on a set of machine- printed gray-scale word images which includes a wide range of print types and qualities.
Characterization of Adipose Tissue Product Quality Using Measurements of Oxygen Consumption Rate.
Suszynski, Thomas M; Sieber, David A; Mueller, Kathryn; Van Beek, Allen L; Cunningham, Bruce L; Kenkel, Jeffrey M
2018-03-14
Fat grafting is a common procedure in plastic surgery but associated with unpredictable graft retention. Adipose tissue (AT) "product" quality is affected by the methods used for harvest, processing and transfer, which vary widely amongst surgeons. Currently, there is no method available to accurately assess the quality of AT. In this study, we present a novel method for the assessment of AT product quality through direct measurements of oxygen consumption rate (OCR). OCR has exhibited potential in predicting outcomes following pancreatic islet transplant. Our study aim was to reapportion existing technology for its use with AT preparations and to confirm that these measurements are feasible. OCR was successfully measured for en bloc and postprocessed AT using a stirred microchamber system. OCR was then normalized to DNA content (OCR/DNA), which represents the AT product quality. Mean (±SE) OCR/DNA values for fresh en bloc and post-processed AT were 149.8 (± 9.1) and 61.1 (± 6.1) nmol/min/mg DNA, respectively. These preliminary data suggest that: (1) OCR and OCR/DNA measurements of AT harvested using conventional protocol are feasible; and (2) standard AT processing results in a decrease in overall AT product quality. OCR measurements of AT using existing technology can be done and enables accurate, real-time, quantitative assessment of the quality of AT product prior to transfer. The availability and further validation of this type of assay could enable optimization of fat grafting protocol by providing a tool for the more detailed study of procedural variables that affect AT product quality.
Comparison of crisp and fuzzy character networks in handwritten word recognition
NASA Technical Reports Server (NTRS)
Gader, Paul; Mohamed, Magdi; Chiang, Jung-Hsien
1992-01-01
Experiments involving handwritten word recognition on words taken from images of handwritten address blocks from the United States Postal Service mailstream are described. The word recognition algorithm relies on the use of neural networks at the character level. The neural networks are trained using crisp and fuzzy desired outputs. The fuzzy outputs were defined using a fuzzy k-nearest neighbor algorithm. The crisp networks slightly outperformed the fuzzy networks at the character level but the fuzzy networks outperformed the crisp networks at the word level.
Spatial-frequency spectra of printed characters and human visual perception.
Põder, Endel
2003-06-01
It is well known that certain spatial frequency (SF) bands are more important than others for character recognition. Solomon and Pelli [Nature 369 (1994) 395-397] have concluded that human pattern recognition mechanism is able to use only a narrow band from available SF spectrum of letters. However, the SF spectra of letters themselves have not been studied carefully. Here I report the results of an analysis of SF spectra of printed characters and discuss their relationship to the observed band-pass nature of letter recognition.
Effectiveness of feature and classifier algorithms in character recognition systems
NASA Astrophysics Data System (ADS)
Wilson, Charles L.
1993-04-01
At the first Census Optical Character Recognition Systems Conference, NIST generated accuracy data for more than character recognition systems. Most systems were tested on the recognition of isolated digits and upper and lower case alphabetic characters. The recognition experiments were performed on sample sizes of 58,000 digits, and 12,000 upper and lower case alphabetic characters. The algorithms used by the 26 conference participants included rule-based methods, image-based methods, statistical methods, and neural networks. The neural network methods included Multi-Layer Perceptron's, Learned Vector Quantitization, Neocognitrons, and cascaded neural networks. In this paper 11 different systems are compared using correlations between the answers of different systems, comparing the decrease in error rate as a function of confidence of recognition, and comparing the writer dependence of recognition. This comparison shows that methods that used different algorithms for feature extraction and recognition performed with very high levels of correlation. This is true for neural network systems, hybrid systems, and statistically based systems, and leads to the conclusion that neural networks have not yet demonstrated a clear superiority to more conventional statistical methods. Comparison of these results with the models of Vapnick (for estimation problems), MacKay (for Bayesian statistical models), Moody (for effective parameterization), and Boltzmann models (for information content) demonstrate that as the limits of training data variance are approached, all classifier systems have similar statistical properties. The limiting condition can only be approached for sufficiently rich feature sets because the accuracy limit is controlled by the available information content of the training set, which must pass through the feature extraction process prior to classification.
Application of the ANNA neural network chip to high-speed character recognition.
Sackinger, E; Boser, B E; Bromley, J; Lecun, Y; Jackel, L D
1992-01-01
A neural network with 136000 connections for recognition of handwritten digits has been implemented using a mixed analog/digital neural network chip. The neural network chip is capable of processing 1000 characters/s. The recognition system has essentially the same rate (5%) as a simulation of the network with 32-b floating-point precision.
Character Recognition Method by Time-Frequency Analyses Using Writing Pressure
NASA Astrophysics Data System (ADS)
Watanabe, Tatsuhito; Katsura, Seiichiro
With the development of information and communication technology, personal verification becomes more and more important. In the future ubiquitous society, the development of terminals handling personal information requires the personal verification technology. The signature is one of the personal verification methods; however, the number of characters is limited in the case of the signature and therefore false signature is used easily. Thus, personal identification is difficult from handwriting. This paper proposes a “haptic pen” that extracts the writing pressure, and shows a character recognition method by time-frequency analyses. Although the figures of characters written by different amanuenses are similar, the differences appear in the time-frequency domain. As a result, it is possible to use the proposed character recognition for personal identification more exactly. The experimental results showed the viability of the proposed method.
Combination of dynamic Bayesian network classifiers for the recognition of degraded characters
NASA Astrophysics Data System (ADS)
Likforman-Sulem, Laurence; Sigelle, Marc
2009-01-01
We investigate in this paper the combination of DBN (Dynamic Bayesian Network) classifiers, either independent or coupled, for the recognition of degraded characters. The independent classifiers are a vertical HMM and a horizontal HMM whose observable outputs are the image columns and the image rows respectively. The coupled classifiers, presented in a previous study, associate the vertical and horizontal observation streams into single DBNs. The scores of the independent and coupled classifiers are then combined linearly at the decision level. We compare the different classifiers -independent, coupled or linearly combined- on two tasks: the recognition of artificially degraded handwritten digits and the recognition of real degraded old printed characters. Our results show that coupled DBNs perform better on degraded characters than the linear combination of independent HMM scores. Our results also show that the best classifier is obtained by linearly combining the scores of the best coupled DBN and the best independent HMM.
Kagawa, Yuki; Miyahara, Hirotaka; Ota, Yuri; Tsuneda, Satoshi
2016-01-01
Estimating the oxygen consumption rates (OCRs) of mammalian cells in hypoxic environments is essential for designing and developing a three-dimensional (3-D) cell culture system. However, OCR measurements under hypoxic conditions are infrequently reported in the literature. Here, we developed a system for measuring OCRs at low oxygen levels. The system injects nitrogen gas into the environment and measures the oxygen concentration by an optical oxygen microsensor that consumes no oxygen. The developed system was applied to HepG2 cells in static culture. Specifically, we measured the spatial profiles of the local dissolved oxygen concentration in the medium, then estimated the OCRs of the cells. The OCRs, and also the pericellular oxygen concentrations, decreased nonlinearly as the oxygen partial pressure in the environment decreased from 19% to 1%. The OCRs also depended on the culture period and the matrix used for coating the dish surface. Using this system, we can precisely estimate the OCRs of various cell types under environments that mimic 3-D culture conditions, contributing crucial data for an efficient 3-D culture system design. © 2015 American Institute of Chemical Engineers.
Recognition of handprinted characters for automated cartography A progress report
NASA Technical Reports Server (NTRS)
Lybanon, M.; Brown, R. M.; Gronmeyer, L. K.
1980-01-01
A research program for developing handwritten character recognition techniques is reported. The generation of cartographic/hydrographic manuscripts is overviewed. The performance of hardware/software systems is discussed, along with future research problem areas and planned approaches.
Scene Text Recognition using Similarity and a Lexicon with Sparse Belief Propagation
Weinman, Jerod J.; Learned-Miller, Erik; Hanson, Allen R.
2010-01-01
Scene text recognition (STR) is the recognition of text anywhere in the environment, such as signs and store fronts. Relative to document recognition, it is challenging because of font variability, minimal language context, and uncontrolled conditions. Much information available to solve this problem is frequently ignored or used sequentially. Similarity between character images is often overlooked as useful information. Because of language priors, a recognizer may assign different labels to identical characters. Directly comparing characters to each other, rather than only a model, helps ensure that similar instances receive the same label. Lexicons improve recognition accuracy but are used post hoc. We introduce a probabilistic model for STR that integrates similarity, language properties, and lexical decision. Inference is accelerated with sparse belief propagation, a bottom-up method for shortening messages by reducing the dependency between weakly supported hypotheses. By fusing information sources in one model, we eliminate unrecoverable errors that result from sequential processing, improving accuracy. In experimental results recognizing text from images of signs in outdoor scenes, incorporating similarity reduces character recognition error by 19%, the lexicon reduces word recognition error by 35%, and sparse belief propagation reduces the lexicon words considered by 99.9% with a 12X speedup and no loss in accuracy. PMID:19696446
The video ocular counter-roll (vOCR): a clinical test to detect loss of otolith-ocular function
Otero-Millan, Jorge; Treviño, Carolina; Winnick, Ariel; Zee, David S.; Carey, John P.; Kheradmand, Amir
2017-01-01
Conclusion vOCR can detect loss of otolith-ocular function without specifying the side of vestibular loss. Since vOCR is measured with a simple head tilt maneuver, it can be potentially used as a bedside clinical test in combination with video head impulse test. Objective Video-oculography (VOG) goggles are being integrated into the bedside assessment of patients with vestibular disorders. Lacking, however, is a method to evaluate otolith function. This study validated a VOG test for loss of otolith function. Methods VOG was used to measure ocular counter-roll (vOCR) in 12 healthy controls, 14 patients with unilateral vestibular loss (UVL), and six patients with bilateral vestibular loss (BVL) with a static lateral head tilt of 30°. The results were compared with vestibular evoked myogenic potentials (VEMP), a widely-used laboratory test of otolith function. Results The average vOCR for healthy controls (4.6°) was significantly different from UVL (2.7°) and BVL (1.6°) patients (p < 0.0001). The vOCR and VEMP measurements were correlated across subjects, especially the click and tap oVEMPs (click oVEMP R = 0.45, tap oVEMP R = 0.51; p < 0.0003). The receiver operator characteristic (ROC) analysis showed that vOCR and VEMPs detected loss of otolith function equally well. The best threshold for vOCR to detect vestibular loss was at 3°. The vOCR values from the side of vestibular loss and the healthy side were not different in UVL patients (2.53° vs 2.8°; p = 0.59). PMID:28084887
Ocular Counter-Rolling During Centrifugation and Static Tilt
NASA Technical Reports Server (NTRS)
Cohen, Bernard; Clement, Gilles; Moore, Steven; Curthoys, Ian; Dai, Mingjia; Koizuka, Izumi; Kubo, Takeshi; Raphan, Theodore
2003-01-01
Activation of the gravity sensors in the inner ear-the otoliths-generates reflexes that act to maintain posture and gaze. Ocular counter-rolling (OCR) is an example of such a reflex. When the head is tilted to the side, the eyes rotate around the line of sight in the opposite direction (i.e., counter-rolling). While turning comers, undergoing centrifugation, or making side-to-side tilting head movements, the OCR reflex orients the eyes towards the sum of the accelerations from body movements and gravity. Deconditioning of otolith-mediated reflexes following adaptation to microgravity has been proposed as the basis of many of the postural, locomotor, and gaze control problems experienced by returning astronauts. Evidence suggests that OCR is reduced postflight in about 75% of astronauts tested; but the data are sparse, primarily due to difficulties in recording rotational eye movements. During the Neurolab mission, a short-arm human centrifuge was flown that generated sustained sideways accelerations of 0.5-G and one-G to the head and upper body. This produces OCR; and so for the first time, the responses to sustained centrifugation could be studied without the influence of Earth's gravity on the results. This allowed us to determine the relative importance of sideways and vertical acceleration in the generation of OCR. This also provided the first test of the effects of exposure to artificial gravity in space on postflight otolith-ocular reflexes. There was little difference between the responses to centrifugation in microgravity and on Earth. In both conditions, the induced OCR was roughly proportional to the applied acceleration, with the OCR magnitude during 0.5-G centrifugation approximately 60% of that generated during one-G centrifugation. The overall mean OCR from the four payload crewmembers in response to one-G of sideways acceleration was 5.7 plus or minus 1.1 degree (mean and SD) on Earth. Inflight one-G centrifugation generated 5.7 plus or minus 1.1 degree of OCR, which was a small but significant decrease in OCR magnitude. The postflight OCR was 5.9 plus or minus 1.4 degree, which was not significantly different from preflight values. During both 0.5-G and one-G centrifugation in microgravity, where the head vertical gravitational component was absent, the OCR magnitude was not significantly different from that produced by an equivalent acceleration during static tilt on Earth. This suggests that the larger OCR magnitude observed during centrifugation on Earth was due to the larger body vertical linear acceleration component, which may have activated either the otoliths or the body tilt receptors. In contrast to previous studies, there was no decrease in OCR gain postflight. Our findings raise the possibility that inflight exposure to artificial gravity, in the form of intermittent one-G and 0.5-G centripetal acceleration, may have been a countermeasure to deconditioning of otolith-based orientation reflexes.
Contribution of finger tracing to the recognition of Chinese characters.
Yim-Ng, Y Y; Varley, R; Andrade, J
2000-01-01
Finger tracing is a simulation of the act of writing without the use of pen and paper. It is claimed to help in the processing of Chinese characters, possibly by providing additional motor coding. In this study, blindfolded subjects were equally good at identifying Chinese characters and novel visual stimuli through passive movements made with the index finger of the preferred hand and those made with the last finger of that hand. This suggests that finger tracing provides a relatively high level of coding specific to individual characters, but non-specific to motor effectors. Beginning each stroke from the same location, i.e. removing spatial information, impaired recognition of the familiar characters and the novel nonsense figures. Passively tracing the strokes in a random sequence also impaired recognition of the characters. These results therefore suggest that the beneficial effect of finger tracing on writing or recall of Chinese characters is mediated by sequence and spatial information embedded in the motor movements, and that proprioceptive channel may play a part in mediating visuo-spatial information. Finger tracing may be a useful strategy for remediation of Chinese language impairments.
Imaging Systems: What, When, How.
ERIC Educational Resources Information Center
Lunin, Lois F.; And Others
1992-01-01
The three articles in this special section on document image files discuss intelligent character recognition, including comparison with optical character recognition; selection of displays for document image processing, focusing on paperlike displays; and imaging hardware, software, and vendors, including guidelines for system selection. (MES)
Neural system applied on an invariant industrial character recognition
NASA Astrophysics Data System (ADS)
Lecoeuche, Stephane; Deguillemont, Denis; Dubus, Jean-Paul
1997-04-01
Besides the variety of fonts, character recognition systems for the industrial world are confronted with specific problems like: the variety of support (metal, wood, paper, ceramics . . .) as well as the variety of marking (printing, engraving, . . .) and conditions of lighting. We present a system that is able to solve a part of this problem. It implements a collaboration between two neural networks. The first network specialized in vision allows the system to extract the character from an image. Besides this capability, we have equipped our system with characteristics allowing it to obtain an invariant model from the presented character. Thus, whatever the position, the size and the orientation of the character during the capture are, the model presented to the input of the second network will be identical. The second network, thanks to a learning phase, permits us to obtain a character recognition system independent of the type of fonts used. Furthermore, its capabilities of generalization permit us to recognize degraded and/or distorted characters. A feedback loop between the two networks permits the first one to modify the quality of vision.The cooperation between these two networks allows us to recognize characters whatever the support and the marking.
Stiefel, C; Schwack, W
2014-12-01
Organic UV filters are used as active ingredients in most sunscreens and also in a variety of daily care products. Their good (photo) stability is of special interest to guarantee protective function and to prevent interactions with the human skin. Due to the mostly electrophilic character of the UV filters, reactions with nucleophilic protein moieties like lysine side chains are conceivable. Prior studies showed that the UV filters octocrylene (OCR), butyl methoxydibenzoylmethane (BM-DBM), ethylhexyl salicylate (EHS), ethylhexyl methoxycinnamate (EHMC), benzophenone-3 (BP-3), ethylhexyl triazone (EHT) and dibenzoylmethane (DBM) were able to covalently bind to an HPTLC amino phase and the amino acid models ethanolamine and butylamine after slightly heating and/or radiation. Boc-protected lysine, the tetrapeptide Boc-Gly-Phe-Gly-Lys-OH, bovine serum albumin (BSA) and porcine gelatin were used as more complex models to determine the reactivity of the mentioned UV filters towards skin proteins under thermal or UV irradiation conditions. After gentle heating at 37°C, benzophenone imines were identified as reaction products of BP-3 and OCR with Boc-lysine and the tetrapeptide, whereas DBM and BM-DBM yielded enamines. For EHMC, a Michael-type reaction occurred, which resulted in addition of Boc-lysine or the tetrapeptide to the conjugated double bond. Ester aminolysis of EHS and EHT mainly afforded the corresponding amides. Reactions of the UV filters with BSA changed the UV spectrum of BSA, generally associated with an increase of the absorption strength in the UVA or UVB range. For all protein models, the UV filters showed an increasing reactivity in the order EHT < EHMC < EHS < BP-3 < OCR < DBM < BM-DBM. Especially the UV absorbers BM-DBM, OCR and BP-3, which are seen as common allergens or photoallergens, showed a high reactivity towards the different skin protein models. As the formation of protein adducts is recognized as important key element in the induction of skin sensitization, the results of this study can contribute to a better understanding of the underlying chemical mechanisms of such reactions. © 2014 Society of Cosmetic Scientists and the Société Française de Cosmétologie.
Hsiao, Janet Hui-Wen
2011-11-01
In Chinese orthography, a dominant character structure exists in which a semantic radical appears on the left and a phonetic radical on the right (SP characters); a minority opposite arrangement also exists (PS characters). As the number of phonetic radical types is much greater than semantic radical types, in SP characters the information is skewed to the right, whereas in PS characters it is skewed to the left. Through training a computational model for SP and PS character recognition that takes into account of the locations in which the characters appear in the visual field during learning, but does not assume any fundamental hemispheric processing difference, we show that visual field differences can emerge as a consequence of the fundamental structural differences in information between SP and PS characters, as opposed to the fundamental processing differences between the two hemispheres. This modeling result is also consistent with behavioral naming performance. This work provides strong evidence that perceptual learning, i.e., the information structure of word stimuli to which the readers have long been exposed, is one of the factors that accounts for hemispheric asymmetry effects in visual word recognition. Copyright © 2011 Elsevier Inc. All rights reserved.
Traleika Glacier X-Stack Extension Final Report
DOE Office of Scientific and Technical Information (OSTI.GOV)
Fryman, Joshua
The XStack Extension Project continued along the direction of the XStack program in exploring the software tools and frameworks to support a task-based community runtime towards the goal of Exascale programming. The momentum built as part of the XStack project, with the development of the task-based Open Community Runtime (OCR) and related tools, was carried through during the XStack Extension with the focus areas of easing application development, improving performance and supporting more features. The infrastructure set up for a community-driven open-source development continued to be used towards these areas, with continued co-development of runtime and applications. A variety ofmore » OCR programming environments were studied, as described in Sections Revolutionary Programming Environments & Applications – to assist with application development on OCR, and we develop OCR Translator, a ROSE-based source-to-source compiler that parses high-level annotations in an MPI program to generate equivalent OCR code. Figure 2 compares the number of OCR objects needed to generate the 2D stencil workload using the translator, against manual approaches based on SPMD library or native coding. The rate of increase with the translator, with an increase in number of ranks, is consistent with other approaches. This is explored further in Section OCR Translator.« less
Triplet-triplet energy transfer from a UV-A absorber butylmethoxydibenzoylmethane to UV-B absorbers.
Kikuchi, Azusa; Oguchi-Fujiyama, Nozomi; Miyazawa, Kazuyuki; Yagi, Mikio
2014-01-01
The phosphorescence decay of a UV-A absorber, 4-tert-butyl-4'-methoxydibenzolymethane (BMDBM) has been observed following a 355 nm laser excitation in the absence and presence of UV-B absorbers, 2-ethylhexyl 4-methoxycinnamate (octyl methoxycinnamate, OMC) and octocrylene (OCR) in ethanol at 77 K. The lifetime of the lowest excited triplet (T1) state of BMDBM is significantly reduced in the presence of OMC and OCR. The observed quenching of BMDBM triplet by OMC and OCR suggests that the intermolecular triplet-triplet energy transfer occurs from BMDBM to OMC and OCR. The T1 state of OCR is nonphosphorescent or very weakly phosphorescent. However, we have shown that the energy level of the T1 state of OCR is lower than that of the enol form of BMDBM. Our methodology of energy-donor phosphorescence decay measurements can be applied to the study of the triplet-triplet energy transfer between UV absorbers even if the energy acceptor is nonphosphorescent. In addition, the delayed fluorescence of BMDBM due to triplet-triplet annihilation was observed in the BMDBM-OMC and BMDBM-OCR mixtures in ethanol at 77 K. Delayed fluorescence is one of the deactivation processes of the excited states of BMDBM under our experimental conditions. © 2013 The American Society of Photobiology.
Effects of head tilt on visual field testing with a head-mounted perimeter imo
Matsumoto, Chota; Nomoto, Hiroki; Numata, Takuya; Eura, Mariko; Yamashita, Marika; Hashimoto, Shigeki; Okuyama, Sachiko; Kimura, Shinji; Yamanaka, Kenzo; Chiba, Yasutaka; Aihara, Makoto; Shimomura, Yoshikazu
2017-01-01
Purpose A newly developed head-mounted perimeter termed “imo” enables visual field (VF) testing without a fixed head position. Because the positional relationship between the subject’s head and the imo is fixed, the effects of head position changes on the test results are small compared with those obtained using a stationary perimeter. However, only ocular counter-roll (OCR) induced by head tilt might affect VF testing. To quantitatively reveal the effects of head tilt and OCR on the VF test results, we investigated the associations among the head-tilt angle, OCR amplitude and VF testing results. Subjects and methods For 20 healthy subjects, we binocularly recorded static OCR (s-OCR) while tilting the subject’s head at an arbitrary angle ranging from 0° to 60° rightward or leftward in 10° increments. By monitoring iris patterns, we evaluated the s-OCR amplitude. We also performed blind spot detection while tilting the subject’s head by an arbitrary angle ranging from 0° to 50° rightward or leftward in 10° increments to calculate the angle by which the blind spot rotates because of head tilt. Results The association between s-OCR amplitude and head-tilt angle showed a sinusoidal relationship. In blind spot detection, the blind spot rotated to the opposite direction of the head tilt, and the association between the rotation angle of the blind spot and the head-tilt angle also showed a sinusoidal relationship. The rotation angle of the blind spot was strongly correlated with the s-OCR amplitude (R2≥0.94, p<0.0001). A head tilt greater than 20° with imo causes interference between adjacent test areas. Conclusions Both the s-OCR amplitude and the rotation angle of the blind spot were correlated with the head-tilt angle by sinusoidal regression. The rotated VF was correlated with the s-OCR amplitude. During perimetry using imo, the change in the subject’s head tilt should be limited to 20°. PMID:28945777
ERIC Educational Resources Information Center
Li, Hong; Shu, Hua; McBride-Chang, Catherine; Liu, Hongyun; Peng, Hong
2012-01-01
Tasks tapping visual skills, orthographic knowledge, phonological awareness, speeded naming, morphological awareness and Chinese character recognition were administered to 184 kindergarteners and 273 primary school students from Beijing. Regression analyses indicated that only syllable deletion, morphological construction and speeded number naming…
MSL: Facilitating automatic and physical analysis of published scientific literature in PDF format
Ahmed, Zeeshan; Dandekar, Thomas
2018-01-01
Published scientific literature contains millions of figures, including information about the results obtained from different scientific experiments e.g. PCR-ELISA data, microarray analysis, gel electrophoresis, mass spectrometry data, DNA/RNA sequencing, diagnostic imaging (CT/MRI and ultrasound scans), and medicinal imaging like electroencephalography (EEG), magnetoencephalography (MEG), echocardiography (ECG), positron-emission tomography (PET) images. The importance of biomedical figures has been widely recognized in scientific and medicine communities, as they play a vital role in providing major original data, experimental and computational results in concise form. One major challenge for implementing a system for scientific literature analysis is extracting and analyzing text and figures from published PDF files by physical and logical document analysis. Here we present a product line architecture based bioinformatics tool ‘Mining Scientific Literature (MSL)’, which supports the extraction of text and images by interpreting all kinds of published PDF files using advanced data mining and image processing techniques. It provides modules for the marginalization of extracted text based on different coordinates and keywords, visualization of extracted figures and extraction of embedded text from all kinds of biological and biomedical figures using applied Optimal Character Recognition (OCR). Moreover, for further analysis and usage, it generates the system’s output in different formats including text, PDF, XML and images files. Hence, MSL is an easy to install and use analysis tool to interpret published scientific literature in PDF format. PMID:29721305
Physicists Get INSPIREd: INSPIRE Project and Grid Applications
NASA Astrophysics Data System (ADS)
Klem, Jukka; Iwaszkiewicz, Jan
2011-12-01
INSPIRE is the new high-energy physics scientific information system developed by CERN, DESY, Fermilab and SLAC. INSPIRE combines the curated and trusted contents of SPIRES database with Invenio digital library technology. INSPIRE contains the entire HEP literature with about one million records and in addition to becoming the reference HEP scientific information platform, it aims to provide new kinds of data mining services and metrics to assess the impact of articles and authors. Grid and cloud computing provide new opportunities to offer better services in areas that require large CPU and storage resources including document Optical Character Recognition (OCR) processing, full-text indexing of articles and improved metrics. D4Science-II is a European project that develops and operates an e-Infrastructure supporting Virtual Research Environments (VREs). It develops an enabling technology (gCube) which implements a mechanism for facilitating the interoperation of its e-Infrastructure with other autonomously running data e-Infrastructures. As a result, this creates the core of an e-Infrastructure ecosystem. INSPIRE is one of the e-Infrastructures participating in D4Science-II project. In the context of the D4Science-II project, the INSPIRE e-Infrastructure makes available some of its resources and services to other members of the resulting ecosystem. Moreover, it benefits from the ecosystem via a dedicated Virtual Organization giving access to an array of resources ranging from computing and storage resources of grid infrastructures to data and services.
Neuroanatomical term generation and comparison between two terminologies.
Srinivas, Prashanti R; Gusfield, Daniel; Mason, Oliver; Gertz, Michael; Hogarth, Michael; Stone, James; Jones, Edward G; Gorin, Fredric A
2003-01-01
An approach and software tools are described for identifying and extracting compound terms (CTs), acronyms and their associated contexts from textual material that is associated with neuroanatomical atlases. A set of simple syntactic rules were appended to the output of a commercially available part of speech (POS) tagger (Qtag v 3.01) that extracts CTs and their associated context from the texts of neuroanatomical atlases. This "hybrid" parser. appears to be highly sensitive and recognized 96% of the potentially germane neuroanatomical CTs and acronyms present in the cat and primate thalamic atlases. A comparison of neuroanatomical CTs and acronymsbetween the cat and primate atlas texts was initially performed using exact-term matching. The implementation of string-matching algorithms significantly improved the identification of relevant terms and acronyms between the two domains. The End Gap Free string matcher identified 98% of CTs and the Needleman Wunsch (NW) string matcher matched 36% of acronyms between the two atlases. Combining several simple grammatical and lexical rules with the POS tagger ("hybrid parser") (1) extracted complex neuroanatomical terms and acronyms from selected cat and primate thalamic atlases and (2) and facilitated the semi-automated generation of a highly granular thalamic terminology. The implementation of string-matching algorithms (1) reconciled terminological errors generated by optical character recognition (OCR) software used to generate the neuroanatomical text information and (2) increased the sensitivity of matching neuroanatomical terms and acronyms between the two neuroanatomical domains that were generated by the "hybrid" parser.
NASA Technical Reports Server (NTRS)
Clement, Gilles; Denise, Pierre; Reschke, Millard; Wood, Scott J.
2007-01-01
Ocular counter-rolling (OCR) induced by whole body tilt in roll has been explored after spaceflight as an indicator of the adaptation of the otolith function to microgravity. It has been claimed that the overall pattern of OCR responses during static body tilt after spaceflight is indicative of a decreased role of the otolith function, but the results of these studies have not been consistent, mostly due to large variations in the OCR within and across individuals. By contrast with static head tilt, off-vertical axis rotation (OVAR) presents the advantage of generating a sinusoidal modulation of OCR, allowing averaged measurements over several cycles, thus improving measurement accuracy. Accordingly, OCR and the sense of roll tilt were evaluated in seven astronauts before and after spaceflight during OVAR at 45 /s in darkness at two angles of tilt (10 and 20 ). There was no significant difference in OCR during OVAR immediately after landing compared to preflight. However, the amplitude of the perceived roll tilt during OVAR was significantly larger immediately postflight, and then returned to control values in the following days. Since the OCR response is predominantly attributed to the shearing force exerted on the utricular macula, the absence of change in OCR postflight suggests that the peripheral otolith organs function normally after short-term spaceflight. However, the increased sense of roll tilt indicates an adaptation in the central processing of gravitational input, presumably related to a re-weigthing of the internal representation of gravitational vertical as a result of adaptation to microgravity.
Optical character recognition with feature extraction and associative memory matrix
NASA Astrophysics Data System (ADS)
Sasaki, Osami; Shibahara, Akihito; Suzuki, Takamasa
1998-06-01
A method is proposed in which handwritten characters are recognized using feature extraction and an associative memory matrix. In feature extraction, simple processes such as shifting and superimposing patterns are executed. A memory matrix is generated with singular value decomposition and by modifying small singular values. The method is optically implemented with two liquid crystal displays. Experimental results for the recognition of 25 handwritten alphabet characters clearly shows the effectiveness of the method.
Yeari, Menahem; Isser, Michal; Schiff, Rachel
2017-07-01
A controversy has recently developed regarding the hypothesis that developmental dyslexia may be caused, in some cases, by a reduced visual attention span (VAS). To examine this hypothesis, independent of phonological abilities, researchers tested the ability of dyslexic participants to recognize arrays of unfamiliar visual characters. Employing this test, findings were rather equivocal: dyslexic participants exhibited poor performance in some studies but normal performance in others. The present study explored four methodological differences revealed between the two sets of studies that might underlie their conflicting results. Specifically, in two experiments we examined whether a VAS deficit is (a) specific to recognition of multi-character arrays as wholes rather than of individual characters within arrays, (b) specific to characters' position within arrays rather than to characters' identity, or revealed only under a higher attention load due to (c) low-discriminable characters, and/or (d) characters' short exposure. Furthermore, in this study we examined whether pure dyslexic participants who do not have attention disorder exhibit a reduced VAS. Although comorbidity of dyslexia and attention disorder is common and the ability to sustain attention for a long time plays a major rule in the visual recognition task, the presence of attention disorder was neither evaluated nor ruled out in previous studies. Findings did not reveal any differences between the performance of dyslexic and control participants on eight versions of the visual recognition task. These findings suggest that pure dyslexic individuals do not present a reduced visual attention span.
A study of payload specialist station monitor size constraints. [space shuttle orbiters
NASA Technical Reports Server (NTRS)
Kirkpatrick, M., III; Shields, N. L., Jr.; Malone, T. B.
1975-01-01
Constraints on the CRT display size for the shuttle orbiter cabin are studied. The viewing requirements placed on these monitors were assumed to involve display of imaged scenes providing visual feedback during payload operations and display of alphanumeric characters. Data on target recognition/resolution, target recognition, and range rate detection by human observers were utilized to determine viewing requirements for imaged scenes. Field-of-view and acuity requirements for a variety of payload operations were obtained along with the necessary detection capability in terms of range-to-target size ratios. The monitor size necessary to meet the acuity requirements was established. An empirical test was conducted to determine required recognition sizes for displayed alphanumeric characters. The results of the test were used to determine the number of characters which could be simultaneously displayed based on the recognition size requirements using the proposed monitor size. A CRT display of 20 x 20 cm is recommended. A portion of the display area is used for displaying imaged scenes and the remaining display area is used for alphanumeric characters pertaining to the displayed scene. The entire display is used for the character alone mode.
Aguer, Céline; Gambarotta, Daniela; Mailloux, Ryan J; Moffat, Cynthia; Dent, Robert; McPherson, Ruth; Harper, Mary-Ellen
2011-01-01
Human primary myotubes are highly glycolytic when cultured in high glucose medium rendering it difficult to study mitochondrial dysfunction. Galactose is known to enhance mitochondrial metabolism and could be an excellent model to study mitochondrial dysfunction in human primary myotubes. The aim of the present study was to 1) characterize the effect of differentiating healthy human myoblasts in galactose on oxidative metabolism and 2) determine whether galactose can pinpoint a mitochondrial malfunction in post-diabetic myotubes. Oxygen consumption rate (OCR), lactate levels, mitochondrial content, citrate synthase and cytochrome C oxidase activities, and AMPK phosphorylation were determined in healthy myotubes differentiated in different sources/concentrations of carbohydrates: 25 mM glucose (high glucose (HG)), 5 mM glucose (low glucose (LG)) or 10 mM galactose (GAL). Effect of carbohydrates on OCR was also determined in myotubes derived from post-diabetic patients and matched obese non-diabetic subjects. OCR was significantly increased whereas anaerobic glycolysis was significantly decreased in GAL myotubes compared to LG or HG myotubes. This increased OCR in GAL myotubes occurred in conjunction with increased cytochrome C oxidase activity and expression, as well as increased AMPK phosphorylation. OCR of post-diabetic myotubes was not different than that of obese non-diabetic myotubes when differentiated in LG or HG. However, whereas GAL increased OCR in obese non-diabetic myotubes, it did not affect OCR in post-diabetic myotubes, leading to a significant difference in OCR between groups. The lack of an increase in OCR in post-diabetic myotubes differentiated in GAL was in relation with unaltered cytochrome C oxidase activity levels or AMPK phosphorylation. Our results indicate that differentiating human primary myoblasts in GAL enhances aerobic metabolism. Because this cell culture model elicited an abnormal response in cells from post-diabetic patients, it may be useful in further studies of the molecular mechanisms of mitochondrial dysfunction.
The Inversion Effect for Chinese Characters is Modulated by Radical Organization.
Luo, Canhuang; Chen, Wei; Zhang, Ye
2017-06-01
In studies of visual object recognition, strong inversion effects accompany the acquisition of expertise and imply the involvement of configural processing. Chinese literacy results in sensitivity to the orthography of Chinese characters. While there is some evidence that this orthographic sensitivity results in an inversion effect, and thus involves configural processing, that processing might depend on exact orthographic properties. Chinese character recognition is believed to involve a hierarchical process, involving at least two lower levels of representation: strokes and radicals. Radicals are grouped into characters according to certain types of structure, i.e. left-right structure, top-bottom structure, or simple characters with only one radical by itself. These types of radical structures vary in both familiarity, and in hierarchical level (compound versus simple characters). In this study, we investigate whether the hierarchical-level or familiarity of radical-structure has an impact on the magnitude of the inversion effect. Participants were asked to do a matching task on pairs of either upright or inverted characters with all the types of structure. Inversion effects were measured based on both reaction time and response sensitivity. While an inversion effect was observed in all 3 conditions, the magnitude of the inversion effect varied with radical structure, being significantly larger for the most familiar type of structure: characters consisting of 2 radicals organized from left to right. These findings indicate that character recognition involves extraction of configural structure as well as radical processing which play different roles in the processing of compound characters and simple characters.
Recognition intent and visual word recognition.
Wang, Man-Ying; Ching, Chi-Le
2009-03-01
This study adopted a change detection task to investigate whether and how recognition intent affects the construction of orthographic representation in visual word recognition. Chinese readers (Experiment 1-1) and nonreaders (Experiment 1-2) detected color changes in radical components of Chinese characters. Explicit recognition demand was imposed in Experiment 2 by an additional recognition task. When the recognition was implicit, a bias favoring the radical location informative of character identity was found in Chinese readers (Experiment 1-1), but not nonreaders (Experiment 1-2). With explicit recognition demands, the effect of radical location interacted with radical function and word frequency (Experiment 2). An estimate of identification performance under implicit recognition was derived in Experiment 3. These findings reflect the joint influence of recognition intent and orthographic regularity in shaping readers' orthographic representation. The implication for the role of visual attention in word recognition was also discussed.
Rasmussen, Luke V; Peissig, Peggy L; McCarty, Catherine A; Starren, Justin
2012-06-01
Although the penetration of electronic health records is increasing rapidly, much of the historical medical record is only available in handwritten notes and forms, which require labor-intensive, human chart abstraction for some clinical research. The few previous studies on automated extraction of data from these handwritten notes have focused on monolithic, custom-developed recognition systems or third-party systems that require proprietary forms. We present an optical character recognition processing pipeline, which leverages the capabilities of existing third-party optical character recognition engines, and provides the flexibility offered by a modular custom-developed system. The system was configured and run on a selected set of form fields extracted from a corpus of handwritten ophthalmology forms. The processing pipeline allowed multiple configurations to be run, with the optimal configuration consisting of the Nuance and LEADTOOLS engines running in parallel with a positive predictive value of 94.6% and a sensitivity of 13.5%. While limitations exist, preliminary experience from this project yielded insights on the generalizability and applicability of integrating multiple, inexpensive general-purpose third-party optical character recognition engines in a modular pipeline.
Peissig, Peggy L; McCarty, Catherine A; Starren, Justin
2011-01-01
Background Although the penetration of electronic health records is increasing rapidly, much of the historical medical record is only available in handwritten notes and forms, which require labor-intensive, human chart abstraction for some clinical research. The few previous studies on automated extraction of data from these handwritten notes have focused on monolithic, custom-developed recognition systems or third-party systems that require proprietary forms. Methods We present an optical character recognition processing pipeline, which leverages the capabilities of existing third-party optical character recognition engines, and provides the flexibility offered by a modular custom-developed system. The system was configured and run on a selected set of form fields extracted from a corpus of handwritten ophthalmology forms. Observations The processing pipeline allowed multiple configurations to be run, with the optimal configuration consisting of the Nuance and LEADTOOLS engines running in parallel with a positive predictive value of 94.6% and a sensitivity of 13.5%. Discussion While limitations exist, preliminary experience from this project yielded insights on the generalizability and applicability of integrating multiple, inexpensive general-purpose third-party optical character recognition engines in a modular pipeline. PMID:21890871
Office for Civil Rights Annual Report to Congress, Fiscal Year 2006
ERIC Educational Resources Information Center
US Department of Education, 2007
2007-01-01
In this report, the Office for Civil Rights (OCR) provides a summary of its substantive achievements in FY 2006. OCR's mission is to ensure equal access to education and to promote educational excellence throughout the nation through vigorous enforcement of civil rights. This report details OCR's accomplishments in enforcing the civil rights laws…
Annual Report to Congress of the Office for Civil Rights. Fiscal Years 2007-08
ERIC Educational Resources Information Center
Office for Civil Rights, US Department of Education, 2009
2009-01-01
This paper is the Office for Civil Rights's (OCR's) "Annual Report to Congress" for fiscal years 2007 and 2008. This report details OCR's accomplishments in enforcing the civil rights laws under which OCR has been granted jurisdiction to address and remedy discrimination. These enforcement efforts include complaint investigation and resolution,…
Annual Report to Congress of the Office for Civil Rights. Fiscal Year 2006
ERIC Educational Resources Information Center
Office for Civil Rights, US Department of Education, 2007
2007-01-01
This paper is the Office for Civil Rights' (OCR's) "Annual Report to Congress" for fiscal year 2006. In this report, OCR provides a summary of its substantive achievements in FY 2006. OCR's mission is to ensure equal access to education and to promote educational excellence throughout the nation through vigorous enforcement of civil…
Enforcing Title IX. A Report of the U.S. Commission on Civil Rights.
ERIC Educational Resources Information Center
Commission on Civil Rights, Washington, DC.
This report reassesses for the Department of Education (ED) the enforcement effort of Title IX by the Office for Civil Rights (OCR) and offers recommendations. OCR is criticized for being very slow to issue important guidelines, process complaints, conduct compliance reviews, and enforce the law. Also OCR is charged with showing little commitment…
The Belousov-Zhabotinskii Reaction: Improving the Oregonator Model with the Arrhenius Equation
ERIC Educational Resources Information Center
Pellitero, Miguel Aller; Lamsfus, Carlos Alvarez; Borge, Javier
2013-01-01
Oscillating chemical reactions (OCRs) have been known since 1828, with the Belousov-Zhabotinskii (BZ) reaction the most studied example. Initially, OCRs were considered to be special cases due to the small number detected and because the oscillatory behavior did not seem to agree with the second law of thermodynamics. However, OCRs have become…
Perceptual expertise: can sensorimotor experience change holistic processing and left-side bias?
Tso, Ricky Van-yip; Au, Terry Kit-fong; Hsiao, Janet Hui-wen
2014-09-01
Holistic processing and left-side bias are both behavioral markers of expert face recognition. By contrast, expert recognition of characters in Chinese orthography involves left-side bias but reduced holistic processing, although faces and Chinese characters share many visual properties. Here, we examined whether this reduction in holistic processing of Chinese characters can be better explained by writing experience than by reading experience. Compared with Chinese nonreaders, Chinese readers who had limited writing experience showed increased holistic processing, whereas Chinese readers who could write characters fluently showed reduced holistic processing. This result suggests that writing and sensorimotor experience can modulate holistic-processing effects and that the reduced holistic processing observed in expert Chinese readers may depend mostly on writing experience. However, both expert writers and writers with limited experience showed similarly stronger left-side bias than novices did in processing mirror-symmetric Chinese characters; left-side bias may therefore be a robust expertise marker for object recognition that is uninfluenced by sensorimotor experience. © The Author(s) 2014.
Metabolic Profile of Pancreatic Acinar and Islet Tissue in Culture
Suszynski, Thomas M.; Mueller, Kathryn; Gruessner, Angelika C.; Papas, Klearchos K.
2016-01-01
The amount and condition of exocrine impurities may affect the quality of islet preparations especially during culture. In this study, the objective was to determine the oxygen demandand viability of islet and acinar tissue post-isolation and whether they change disproportionately while in culture. We compare the OCR normalized to DNA (OCR/DNA, a measure of fractional viability in units nmol/min/mg DNA), and percent change in OCR and DNA recoveries between adult porcine islet and acinar tissue from the same preparation (paired) over a 6-9 days of standard culture. Paired comparisons were done to quantify differences in OCR/DNA between islet and acinar tissue from the same preparation, at specified time points during culture; the mean (± standard error) OCR/DNA was 74.0 (±11.7) units higher for acinar (vs. islet) tissue on the day of isolation (n=16, p<0.0001), but 25.7 (±9.4) units lower after 1 day (n=8, p=0.03), 56.6 (±11.5) units lower after 2 days (n=12, p=0.0004), and 65.9 (±28.7) units lower after 8 days (n=4, p=0.2) in culture. DNA and OCR recoveries decreased at different rates for acinar versus islet tissue over 6-9 days in culture (n=6). DNA recovery decreased to 24±7% for acinar and 75±8% for islets (p=0.002). Similarly, OCR recovery decreased to 16±3% for acinar and remained virtually constant for islets (p=0.005). Differences in the metabolic profile of acinarand islet tissue should be considered when culturing impure islet preparations. OCR-based measurements may help optimize pre-IT culture protocols. PMID:25131082
Syntax-directed content analysis of videotext: application to a map detection recognition system
NASA Astrophysics Data System (ADS)
Aradhye, Hrishikesh; Herson, James A.; Myers, Gregory
2003-01-01
Video is an increasingly important and ever-growing source of information to the intelligence and homeland defense analyst. A capability to automatically identify the contents of video imagery would enable the analyst to index relevant foreign and domestic news videos in a convenient and meaningful way. To this end, the proposed system aims to help determine the geographic focus of a news story directly from video imagery by detecting and geographically localizing political maps from news broadcasts, using the results of videotext recognition in lieu of a computationally expensive, scale-independent shape recognizer. Our novel method for the geographic localization of a map is based on the premise that the relative placement of text superimposed on a map roughly corresponds to the geographic coordinates of the locations the text represents. Our scheme extracts and recognizes videotext, and iteratively identifies the geographic area, while allowing for OCR errors and artistic freedom. The fast and reliable recognition of such maps by our system may provide valuable context and supporting evidence for other sources, such as speech recognition transcripts. The concepts of syntax-directed content analysis of videotext presented here can be extended to other content analysis systems.
Using the web to validate document recognition results: experiments with business cards
NASA Astrophysics Data System (ADS)
Oertel, Clemens; O'Shea, Shauna; Bodnar, Adam; Blostein, Dorothea
2004-12-01
The World Wide Web is a vast information resource which can be useful for validating the results produced by document recognizers. Three computational steps are involved, all of them challenging: (1) use the recognition results in a Web search to retrieve Web pages that contain information similar to that in the document, (2) identify the relevant portions of the retrieved Web pages, and (3) analyze these relevant portions to determine what corrections (if any) should be made to the recognition result. We have conducted exploratory implementations of steps (1) and (2) in the business-card domain: we use fields of the business card to retrieve Web pages and identify the most relevant portions of those Web pages. In some cases, this information appears suitable for correcting OCR errors in the business card fields. In other cases, the approach fails due to stale information: when business cards are several years old and the business-card holder has changed jobs, then websites (such as the home page or company website) no longer contain information matching that on the business card. Our exploratory results indicate that in some domains it may be possible to develop effective means of querying the Web with recognition results, and to use this information to correct the recognition results and/or detect that the information is stale.
Using the web to validate document recognition results: experiments with business cards
NASA Astrophysics Data System (ADS)
Oertel, Clemens; O'Shea, Shauna; Bodnar, Adam; Blostein, Dorothea
2005-01-01
The World Wide Web is a vast information resource which can be useful for validating the results produced by document recognizers. Three computational steps are involved, all of them challenging: (1) use the recognition results in a Web search to retrieve Web pages that contain information similar to that in the document, (2) identify the relevant portions of the retrieved Web pages, and (3) analyze these relevant portions to determine what corrections (if any) should be made to the recognition result. We have conducted exploratory implementations of steps (1) and (2) in the business-card domain: we use fields of the business card to retrieve Web pages and identify the most relevant portions of those Web pages. In some cases, this information appears suitable for correcting OCR errors in the business card fields. In other cases, the approach fails due to stale information: when business cards are several years old and the business-card holder has changed jobs, then websites (such as the home page or company website) no longer contain information matching that on the business card. Our exploratory results indicate that in some domains it may be possible to develop effective means of querying the Web with recognition results, and to use this information to correct the recognition results and/or detect that the information is stale.
Hsiao, Janet H; Cheung, Kit
2016-03-01
In Chinese orthography, the most common character structure consists of a semantic radical on the left and a phonetic radical on the right (SP characters); the minority, opposite arrangement also exists (PS characters). Recent studies showed that SP character processing is more left hemisphere (LH) lateralized than PS character processing. Nevertheless, it remains unclear whether this is due to phonetic radical position or character type frequency. Through computational modeling with artificial lexicons, in which we implement a theory of hemispheric asymmetry in perception but do not assume phonological processing being LH lateralized, we show that the difference in character type frequency alone is sufficient to exhibit the effect that the dominant type has a stronger LH lateralization than the minority type. This effect is due to higher visual similarity among characters in the dominant type than the minority type, demonstrating the modulation of visual similarity of words on hemispheric lateralization. Copyright © 2015 Cognitive Science Society, Inc.
Keeping on Keeping on: OCR and Complaints of Racial Discrimination 50 Years after "Brown"
ERIC Educational Resources Information Center
Pollock, Mica
2005-01-01
This article, written by a former civil rights investigator in the U.S. Department of Education's Office for Civil Rights (OCR), contends that ordinary Americans advocating for equal educational opportunity for students of color might enlist OCR more actively and knowingly to help secure racial equality of opportunity 50 years after "Brown." Now a…
End-to-end system of license plate localization and recognition
NASA Astrophysics Data System (ADS)
Zhu, Siyu; Dianat, Sohail; Mestha, Lalit K.
2015-03-01
An end-to-end license plate recognition system is proposed. It is composed of preprocessing, detection, segmentation, and character recognition to find and recognize plates from camera-based still images. The system utilizes connected component (CC) properties to quickly extract the license plate region. A two-stage CC filtering is utilized to address both shape and spatial relationship information to produce high precision and to recall values for detection. Floating peak and valleys of projection profiles are used to cut the license plates into individual characters. A turning function-based method is proposed to quickly and accurately recognize each character. It is further accelerated using curvature histogram-based support vector machine. The INFTY dataset is used to train the recognition system, and MediaLab license plate dataset is used for testing. The proposed system achieved 89.45% F-measure for detection and 87.33% accuracy for overall recognition rate which is comparable to current state-of-the-art systems.
AN OPTICAL CHARACTER RECOGNITION RESEARCH AND DEMONSTRATION PROJECT.
ERIC Educational Resources Information Center
1968
RESEARCH AND DEVELOPMENT OF PROTOTYPE LIBRARY SYSTEMS WHICH UTILIZE OPTICAL CHARACTER RECOGNITION INPUT HAS CENTERED AROUND OPTICAL PAGE READERS AND DOCUMENT READERS. THE STATE-OF-THE-ART OF BOTH THESE OPTICAL SCANNERS IS SUCH THAT BOTH ARE ACCEPTABLE FOR LIBRARY INPUT PREPARATION. A DEMONSTRATION PROJECT UTILIZING THE TWO TYPES OF READERS, SINCE…
Lin, Nan; Yu, Xi; Zhao, Ying; Zhang, Mingxia
2016-01-01
This fMRI study aimed to identify the neural mechanisms underlying the recognition of Chinese multi-character words by partialling out the confounding effect of reaction time (RT). For this purpose, a special type of nonword-transposable nonword-was created by reversing the character orders of real words. These nonwords were included in a lexical decision task along with regular (non-transposable) nonwords and real words. Through conjunction analysis on the contrasts of transposable nonwords versus regular nonwords and words versus regular nonwords, the confounding effect of RT was eliminated, and the regions involved in word recognition were reliably identified. The word-frequency effect was also examined in emerged regions to further assess their functional roles in word processing. Results showed significant conjunctional effect and positive word-frequency effect in the bilateral inferior parietal lobules and posterior cingulate cortex, whereas only conjunctional effect was found in the anterior cingulate cortex. The roles of these brain regions in recognition of Chinese multi-character words were discussed.
Lin, Nan; Yu, Xi; Zhao, Ying; Zhang, Mingxia
2016-01-01
This fMRI study aimed to identify the neural mechanisms underlying the recognition of Chinese multi-character words by partialling out the confounding effect of reaction time (RT). For this purpose, a special type of nonword—transposable nonword—was created by reversing the character orders of real words. These nonwords were included in a lexical decision task along with regular (non-transposable) nonwords and real words. Through conjunction analysis on the contrasts of transposable nonwords versus regular nonwords and words versus regular nonwords, the confounding effect of RT was eliminated, and the regions involved in word recognition were reliably identified. The word-frequency effect was also examined in emerged regions to further assess their functional roles in word processing. Results showed significant conjunctional effect and positive word-frequency effect in the bilateral inferior parietal lobules and posterior cingulate cortex, whereas only conjunctional effect was found in the anterior cingulate cortex. The roles of these brain regions in recognition of Chinese multi-character words were discussed. PMID:26901644
Handprinted Forms and Characters
National Institute of Standards and Technology Data Gateway
NIST Handprinted Forms and Characters (Web, free access) NIST Special Database 19 contains NIST's entire corpus of training materials for handprinted document and character recognition. It supersedes NIST Special Databases 3 and 7.
The role of lexical variables in the visual recognition of Chinese characters: A megastudy analysis.
Sze, Wei Ping; Yap, Melvin J; Rickard Liow, Susan J
2015-01-01
Logographic Chinese orthography partially represents both phonology and semantics. By capturing the online processing of a large pool of Chinese characters, we were able to examine the relative salience of specific lexical variables when this nonalphabetic script is read. Using a sample of native mainland Chinese speakers (N = 35), lexical decision latencies for 1560 single characters were collated into a database, before the effects of a comprehensive range of variables were explored. Hierarchical regression analyses determined the unique item-level variance explained by orthographic (frequency, stroke count), semantic (age of learning, imageability, number of meanings), and phonological (consistency, phonological frequency) factors. Orthographic and semantic variables, respectively, accounted for more collective variance than the phonological variables. Significant main effects were further observed for the individual orthographic and semantic predictors. These results are consistent with the idea that skilled readers tend to rely on orthographic and semantic information when processing visually presented characters. This megastudy approach marks an important extension to existing work on Chinese character recognition, which hitherto has relied on factorial designs. Collectively, the findings reported here represent a useful set of empirical constraints for future computational models of character recognition.
Jersey number detection in sports video for athlete identification
NASA Astrophysics Data System (ADS)
Ye, Qixiang; Huang, Qingming; Jiang, Shuqiang; Liu, Yang; Gao, Wen
2005-07-01
Athlete identification is important for sport video content analysis since users often care about the video clips with their preferred athletes. In this paper, we propose a method for athlete identification by combing the segmentation, tracking and recognition procedures into a coarse-to-fine scheme for jersey number (digital characters on sport shirt) detection. Firstly, image segmentation is employed to separate the jersey number regions with its background. And size/pipe-like attributes of digital characters are used to filter out candidates. Then, a K-NN (K nearest neighbor) classifier is employed to classify a candidate into a digit in "0-9" or negative. In the recognition procedure, we use the Zernike moment features, which are invariant to rotation and scale for digital shape recognition. Synthetic training samples with different fonts are used to represent the pattern of digital characters with non-rigid deformation. Once a character candidate is detected, a SSD (smallest square distance)-based tracking procedure is started. The recognition procedure is performed every several frames in the tracking process. After tracking tens of frames, the overall recognition results are combined to determine if a candidate is a true jersey number or not by a voting procedure. Experiments on several types of sports video shows encouraging result.
NASA Astrophysics Data System (ADS)
Bai, Hao; Zhang, Xi-wen
2017-06-01
While Chinese is learned as a second language, its characters are taught step by step from their strokes to components, radicals to components, and their complex relations. Chinese Characters in digital ink from non-native language writers are deformed seriously, thus the global recognition approaches are poorer. So a progressive approach from bottom to top is presented based on hierarchical models. Hierarchical information includes strokes and hierarchical components. Each Chinese character is modeled as a hierarchical tree. Strokes in one Chinese characters in digital ink are classified with Hidden Markov Models and concatenated to the stroke symbol sequence. And then the structure of components in one ink character is extracted. According to the extraction result and the stroke symbol sequence, candidate characters are traversed and scored. Finally, the recognition candidate results are listed by descending. The method of this paper is validated by testing 19815 copies of the handwriting Chinese characters written by foreign students.
Holistic neural coding of Chinese character forms in bilateral ventral visual system.
Mo, Ce; Yu, Mengxia; Seger, Carol; Mo, Lei
2015-02-01
How are Chinese characters recognized and represented in the brain of skilled readers? Functional MRI fast adaptation technique was used to address this question. We found that neural adaptation effects were limited to identical characters in bilateral ventral visual system while no activation reduction was observed for partially overlapping characters regardless of the spatial location of the shared sub-character components, suggesting highly selective neuronal tuning to whole characters. The consistent neural profile across the entire ventral visual cortex indicates that Chinese characters are represented as mutually distinctive wholes rather than combinations of sub-character components, which presents a salient contrast to the left-lateralized, simple-to-complex neural representations of alphabetic words. Our findings thus revealed the cultural modulation effect on both local neuronal activity patterns and functional anatomical regions associated with written symbol recognition. Moreover, the cross-language discrepancy in written symbol recognition mechanism might stem from the language-specific early-stage learning experience. Copyright © 2014 The Authors. Published by Elsevier Inc. All rights reserved.
Deriving an Abstraction Network to Support Quality Assurance in OCRe
Ochs, Christopher; Agrawal, Ankur; Perl, Yehoshua; Halper, Michael; Tu, Samson W.; Carini, Simona; Sim, Ida; Noy, Natasha; Musen, Mark; Geller, James
2012-01-01
An abstraction network is an auxiliary network of nodes and links that provides a compact, high-level view of an ontology. Such a view lends support to ontology orientation, comprehension, and quality-assurance efforts. A methodology is presented for deriving a kind of abstraction network, called a partial-area taxonomy, for the Ontology of Clinical Research (OCRe). OCRe was selected as a representative of ontologies implemented using the Web Ontology Language (OWL) based on shared domains. The derivation of the partial-area taxonomy for the Entity hierarchy of OCRe is described. Utilizing the visualization of the content and structure of the hierarchy provided by the taxonomy, the Entity hierarchy is audited, and several errors and inconsistencies in OCRe’s modeling of its domain are exposed. After appropriate corrections are made to OCRe, a new partial-area taxonomy is derived. The generalizability of the paradigm of the derivation methodology to various families of biomedical ontologies is discussed. PMID:23304341
ERIC Educational Resources Information Center
Kwong, Elena; Burns, Matthew K.
2016-01-01
The current study examined the effectiveness of Incremental Rehearsal (IR) for teaching Chinese character recognition using a single-case experimental design. In addition, a morphological component was added to standard IR procedures (IRM) to take into account the role of morphological awareness in Chinese reading. Three kindergarten students in…
ERIC Educational Resources Information Center
Liu, Duo; McBride-Chang, Catherine
2014-01-01
In the present study, we explored the characteristics of morphological structure processing during word recognition among third grade Chinese children and its possible relationship with Chinese character reading. By using the modified priming lexical decision paradigm, a significant morphological structure priming effect was found in the subject…
Scene text recognition in mobile applications by character descriptor and structure configuration.
Yi, Chucai; Tian, Yingli
2014-07-01
Text characters and strings in natural scene can provide valuable information for many applications. Extracting text directly from natural scene images or videos is a challenging task because of diverse text patterns and variant background interferences. This paper proposes a method of scene text recognition from detected text regions. In text detection, our previously proposed algorithms are applied to obtain text regions from scene image. First, we design a discriminative character descriptor by combining several state-of-the-art feature detectors and descriptors. Second, we model character structure at each character class by designing stroke configuration maps. Our algorithm design is compatible with the application of scene text extraction in smart mobile devices. An Android-based demo system is developed to show the effectiveness of our proposed method on scene text information extraction from nearby objects. The demo system also provides us some insight into algorithm design and performance improvement of scene text extraction. The evaluation results on benchmark data sets demonstrate that our proposed scheme of text recognition is comparable with the best existing methods.
Structural model constructing for optical handwritten character recognition
NASA Astrophysics Data System (ADS)
Khaustov, P. A.; Spitsyn, V. G.; Maksimova, E. I.
2017-02-01
The article is devoted to the development of the algorithms for optical handwritten character recognition based on the structural models constructing. The main advantage of these algorithms is the low requirement regarding the number of reference images. The one-pass approach to a thinning of the binary character representation has been proposed. This approach is based on the joint use of Zhang-Suen and Wu-Tsai algorithms. The effectiveness of the proposed approach is confirmed by the results of the experiments. The article includes the detailed description of the structural model constructing algorithm’s steps. The proposed algorithm has been implemented in character processing application and has been approved on MNIST handwriting characters database. Algorithms that could be used in case of limited reference images number were used for the comparison.
Nonlinear filtering for character recognition in low quality document images
NASA Astrophysics Data System (ADS)
Diaz-Escobar, Julia; Kober, Vitaly
2014-09-01
Optical character recognition in scanned printed documents is a well-studied task, where the captured conditions like sheet position, illumination, contrast and resolution are controlled. Nowadays, it is more practical to use mobile devices for document capture than a scanner. So as a consequence, the quality of document images is often poor owing to presence of geometric distortions, nonhomogeneous illumination, low resolution, etc. In this work we propose to use multiple adaptive nonlinear composite filters for detection and classification of characters. Computer simulation results obtained with the proposed system are presented and discussed.
Radical Sensitivity Is the Key to Understanding Chinese Character Acquisition in Children
ERIC Educational Resources Information Center
Tong, Xiuhong; Tong, Xiuli; McBride, Catherine
2017-01-01
This study investigated Chinese children's development of sensitivity to positional (orthographic), phonological, and semantic cues of radicals in encoding novel Chinese characters. A newly designed picture-novel character mapping task, along with nonverbal reasoning ability, vocabulary, and Chinese character recognition were administered to 198…
ERIC Educational Resources Information Center
Pattillo, Suzan Trefry; Heller, Kathryn Wolf; Smith, Maureen
2004-01-01
The repeated-reading strategy and optical character recognition were paired to demonstrate a functional relationship between the combined strategies and two factors: the reading rates of students with visual impairments and the students' self-perceptions, or attitudes, toward reading. The results indicated that all five students increased their…
ERIC Educational Resources Information Center
Higgins, Eleanor L.; Raskind, Marshall H.
1997-01-01
Thirty-seven college students with learning disabilities were given a reading comprehension task under the following conditions: (1) using an optical character recognition/speech synthesis system; (2) having the text read aloud by a human reader; or (3) reading silently without assistance. Findings indicated that the greater the disability, the…
ERIC Educational Resources Information Center
Wu, Shiyu; Ma, Zheng
2017-01-01
Previous research has indicated that, in viewing a visual word, the activated phonological representation in turn activates its homophone, causing semantic interference. Using this mechanism of phonological mediation, this study investigated native-language phonological interference in visual recognition of Chinese two-character compounds by early…
Hybrid neuro-fuzzy approach for automatic vehicle license plate recognition
NASA Astrophysics Data System (ADS)
Lee, Hsi-Chieh; Jong, Chung-Shi
1998-03-01
Most currently available vehicle identification systems use techniques such as R.F., microwave, or infrared to help identifying the vehicle. Transponders are usually installed in the vehicle in order to transmit the corresponding information to the sensory system. It is considered expensive to install a transponder in each vehicle and the malfunction of the transponder will result in the failure of the vehicle identification system. In this study, novel hybrid approach is proposed for automatic vehicle license plate recognition. A system prototype is built which can be used independently or cooperating with current vehicle identification system in identifying a vehicle. The prototype consists of four major modules including the module for license plate region identification, the module for character extraction from the license plate, the module for character recognition, and the module for the SimNet neuro-fuzzy system. To test the performance of the proposed system, three hundred and eighty vehicle image samples are taken by a digital camera. The license plate recognition success rate of the prototype is approximately 91% while the character recognition success rate of the prototype is approximately 97%.
Anonymization of DICOM Electronic Medical Records for Radiation Therapy
Newhauser, Wayne; Jones, Timothy; Swerdloff, Stuart; Newhauser, Warren; Cilia, Mark; Carver, Robert; Halloran, Andy; Zhang, Rui
2014-01-01
Electronic medical records (EMR) and treatment plans are used in research on patient outcomes and radiation effects. In many situations researchers must remove protected health information (PHI) from EMRs. The literature contains several studies describing the anonymization of generic Digital Imaging and Communication in Medicine (DICOM) files and DICOM image sets but no publications were found that discuss the anonymization of DICOM radiation therapy plans, a key component of an EMR in a cancer clinic. In addition to this we were unable to find a commercial software tool that met the minimum requirements for anonymization and preservation of data integrity for radiation therapy research. The purpose of this study was to develop a prototype software code to meet the requirements for the anonymization of radiation therapy treatment plans and to develop a way to validate that code and demonstrate that it properly anonymized treatment plans and preserved data integrity. We extended an open-source code to process all relevant PHI and to allow for the automatic anonymization of multiple EMRs. The prototype code successfully anonymized multiple treatment plans in less than 1 minute per patient. We also tested commercial optical character recognition (OCR) algorithms for the detection of burned-in text on the images, but they were unable to reliably recognize text. In addition, we developed and tested an image filtering algorithm that allowed us to isolate and redact alpha-numeric text from a test radiograph. Validation tests verified that PHI was anonymized and data integrity, such as the relationship between DICOM unique identifiers (UID) was preserved. PMID:25147130
Anonymization of DICOM electronic medical records for radiation therapy.
Newhauser, Wayne; Jones, Timothy; Swerdloff, Stuart; Newhauser, Warren; Cilia, Mark; Carver, Robert; Halloran, Andy; Zhang, Rui
2014-10-01
Electronic medical records (EMR) and treatment plans are used in research on patient outcomes and radiation effects. In many situations researchers must remove protected health information (PHI) from EMRs. The literature contains several studies describing the anonymization of generic Digital Imaging and Communication in Medicine (DICOM) files and DICOM image sets but no publications were found that discuss the anonymization of DICOM radiation therapy plans, a key component of an EMR in a cancer clinic. In addition to this we were unable to find a commercial software tool that met the minimum requirements for anonymization and preservation of data integrity for radiation therapy research. The purpose of this study was to develop a prototype software code to meet the requirements for the anonymization of radiation therapy treatment plans and to develop a way to validate that code and demonstrate that it properly anonymized treatment plans and preserved data integrity. We extended an open-source code to process all relevant PHI and to allow for the automatic anonymization of multiple EMRs. The prototype code successfully anonymized multiple treatment plans in less than 1min/patient. We also tested commercial optical character recognition (OCR) algorithms for the detection of burned-in text on the images, but they were unable to reliably recognize text. In addition, we developed and tested an image filtering algorithm that allowed us to isolate and redact alpha-numeric text from a test radiograph. Validation tests verified that PHI was anonymized and data integrity, such as the relationship between DICOM unique identifiers (UID) was preserved. Copyright © 2014 Elsevier Ltd. All rights reserved.
Development of a Machine-Vision System for Recording of Force Calibration Data
NASA Astrophysics Data System (ADS)
Heamawatanachai, Sumet; Chaemthet, Kittipong; Changpan, Tawat
This paper presents the development of a new system for recording of force calibration data using machine vision technology. Real time camera and computer system were used to capture images of the reading from the instruments during calibration. Then, the measurement images were transformed and translated to numerical data using optical character recognition (OCR) technique. These numerical data along with raw images were automatically saved to memories as the calibration database files. With this new system, the human error of recording would be eliminated. The verification experiments were done by using this system for recording the measurement results from an amplifier (DMP 40) with load cell (HBM-Z30-10kN). The NIMT's 100-kN deadweight force standard machine (DWM-100kN) was used to generate test forces. The experiments setup were done in 3 categories; 1) dynamics condition (record during load changing), 2) statics condition (record during fix load), and 3) full calibration experiments in accordance with ISO 376:2011. The captured images from dynamics condition experiment gave >94% without overlapping of number. The results from statics condition experiment were >98% images without overlapping. All measurement images without overlapping were translated to number by the developed program with 100% accuracy. The full calibration experiments also gave 100% accurate results. Moreover, in case of incorrect translation of any result, it is also possible to trace back to the raw calibration image to check and correct it. Therefore, this machine-vision-based system and program should be appropriate for recording of force calibration data.
Document recognition serving people with disabilities
NASA Astrophysics Data System (ADS)
Fruchterman, James R.
2007-01-01
Document recognition advances have improved the lives of people with print disabilities, by providing accessible documents. This invited paper provides perspectives on the author's career progression from document recognition professional to social entrepreneur applying this technology to help people with disabilities. Starting with initial thoughts about optical character recognition in college, it continues with the creation of accurate omnifont character recognition that did not require training. It was difficult to make a reading machine for the blind in a commercial setting, which led to the creation of a nonprofit social enterprise to deliver these devices around the world. This network of people with disabilities scanning books drove the creation of Bookshare.org, an online library of scanned books. Looking forward, the needs for improved document recognition technology to further lower the barriers to reading are discussed. Document recognition professionals should be proud of the positive impact their work has had on some of society's most disadvantaged communities.
Character recognition from trajectory by recurrent spiking neural networks.
Jiangrong Shen; Kang Lin; Yueming Wang; Gang Pan
2017-07-01
Spiking neural networks are biologically plausible and power-efficient on neuromorphic hardware, while recurrent neural networks have been proven to be efficient on time series data. However, how to use the recurrent property to improve the performance of spiking neural networks is still a problem. This paper proposes a recurrent spiking neural network for character recognition using trajectories. In the network, a new encoding method is designed, in which varying time ranges of input streams are used in different recurrent layers. This is able to improve the generalization ability of our model compared with general encoding methods. The experiments are conducted on four groups of the character data set from University of Edinburgh. The results show that our method can achieve a higher average recognition accuracy than existing methods.
Loss of otolith function with age is associated with increased postural sway measures.
Serrador, Jorge M; Lipsitz, Lewis A; Gopalakrishnan, Gosala S; Black, F Owen; Wood, Scott J
2009-11-06
Loss of balance and increased fall risk is a common problem associated with aging. Changes in vestibular function occur with aging but the contribution of reduced vestibular otolith function to fall risk remains unknown. We examined a population of 151 healthy individuals (aged 21-93) for both balance (sway measures) and ocular counter-rolling (OCR) function. We assessed balance function with eyes open and closed on a firm surface, eyes open and closed on a foam surface and OCR during +/-20 degree roll tilt at 0.005 Hz. Subjects demonstrated a significant age-related reduction in OCR and increase in postural sway. The effect of age on OCR was greater in females than males. The reduction in OCR was strongly correlated with the mediolateral measures of sway with eyes closed. This correlation was also present in the elderly group alone, suggesting that aging alone does not account for this effect. OCR decreased linearly with age and at a greater rate in females than males. This loss of vestibular otolith-ocular function is associated with increased mediolateral measures of sway which have been shown to be related to increased risk of falls. These data suggest a role for loss of otolith function in contributing to fall risk in the elderly. Further prospective, longitudinal studies are necessary to confirm these findings.
Optical character recognition of camera-captured images based on phase features
NASA Astrophysics Data System (ADS)
Diaz-Escobar, Julia; Kober, Vitaly
2015-09-01
Nowadays most of digital information is obtained using mobile devices specially smartphones. In particular, it brings the opportunity for optical character recognition in camera-captured images. For this reason many recognition applications have been recently developed such as recognition of license plates, business cards, receipts and street signal; document classification, augmented reality, language translator and so on. Camera-captured images are usually affected by geometric distortions, nonuniform illumination, shadow, noise, which make difficult the recognition task with existing systems. It is well known that the Fourier phase contains a lot of important information regardless of the Fourier magnitude. So, in this work we propose a phase-based recognition system exploiting phase-congruency features for illumination/scale invariance. The performance of the proposed system is tested in terms of miss classifications and false alarms with the help of computer simulation.
Aging per se Increases the Susceptibility to Free Fatty Acid–Induced Insulin Resistance
Huffman, Derek M.; Fishman, Sigal; Jerschow, Elina; Heo, Hye J.; Atzmon, Gil; Schechter, Clyde; Muzumdar, Radhika H.
2010-01-01
Elevations in systemic free fatty acids (FFA) contribute to insulin resistance. To determine the effects of an acute elevation in FFA on insulin action with aging, we infused saline or intralipid (IL) during a hyperinsulinemic–euglycemic clamp in three groups of rats: young ad libitum–fed (YAL), old ad libitum–fed (OAL), and old on lifelong calorie restriction (OCR). The OCR group was included to distinguish between aging per se and age-related changes in body fat distribution. IL induced marked insulin resistance in both YAL and OCR, but the onset of insulin resistance was approximately two to three times more rapid in OCR as compared with YAL. In response to IL infusion, plasminogen-activating inhibitor-1 (PAI-1) expression was increased in subcutaneous fat from OAL animals. In visceral fat, a marked increase in PAI-1 and interleukin-6 expression was observed in OAL and OCR rats, but not YAL, in response to IL treatment. Thus, aging per se increases the inflammatory response to excess nutrients and vulnerability to FFA-induced insulin resistance with aging. PMID:20504893
ERIC Educational Resources Information Center
Hsiao, Janet Hui-wen
2011-01-01
In Chinese orthography, a dominant character structure exists in which a semantic radical appears on the left and a phonetic radical on the right (SP characters); a minority opposite arrangement also exists (PS characters). As the number of phonetic radical types is much greater than semantic radical types, in SP characters the information is…
A distinguishing method of printed and handwritten legal amount on Chinese bank check
NASA Astrophysics Data System (ADS)
Zhu, Ningbo; Lou, Zhen; Yang, Jingyu
2003-09-01
While carrying out Optical Chinese Character Recognition, distinguishing the font between printed and handwritten characters at the early phase is necessary, because there is so much difference between the methods on recognizing these two types of characters. In this paper, we proposed a good method on how to banish seals and its relative standards that can judge whether they should be banished. Meanwhile, an approach on clearing up scattered noise shivers after image segmentation is presented. Four sets of classifying features that show discrimination between printed and handwritten characters are well adopted. The proposed approach was applied to an automatic check processing system and tested on about 9031 checks. The recognition rate is more than 99.5%.
Image based automatic water meter reader
NASA Astrophysics Data System (ADS)
Jawas, N.; Indrianto
2018-01-01
Water meter is used as a tool to calculate water consumption. This tool works by utilizing water flow and shows the calculation result with mechanical digit counter. Practically, in everyday use, an operator will manually check the digit counter periodically. The Operator makes logs of the number shows by water meter to know the water consumption. This manual operation is time consuming and prone to human error. Therefore, in this paper we propose an automatic water meter digit reader from digital image. The digits sequence is detected by utilizing contour information of the water meter front panel.. Then an OCR method is used to get the each digit character. The digit sequence detection is an important part of overall process. It determines the success of overall system. The result shows promising results especially in sequence detection.
NASA Astrophysics Data System (ADS)
Jamróz, Dorota; Wójcik, Marek; Lindgren, Jan
2000-09-01
Infrared spectra of mixtures of water and deuteroacetonitrile containing the Cr 3+ cation have been studied as a function of concentration, time and temperature. The CN stretching vibration of CD 3CN molecules has been used as a probe of the structural environments. The CN band in the spectra of the solutions is a superposition of four subbands, which may be attributed to CD 3CN bound in the first, second, and third solvation shells of the cation and to non-bound CD 3CN. The character of changes of the integral intensities of the subbands with time for various H 2O:Cr 3+ molar ratios are explained by suggesting mechanisms of molecular replacement within the solvation shells of Cr 3+.
Practical automatic Arabic license plate recognition system
NASA Astrophysics Data System (ADS)
Mohammad, Khader; Agaian, Sos; Saleh, Hani
2011-02-01
Since 1970's, the need of an automatic license plate recognition system, sometimes referred as Automatic License Plate Recognition system, has been increasing. A license plate recognition system is an automatic system that is able to recognize a license plate number, extracted from image sensors. In specific, Automatic License Plate Recognition systems are being used in conjunction with various transportation systems in application areas such as law enforcement (e.g. speed limit enforcement) and commercial usages such as parking enforcement and automatic toll payment private and public entrances, border control, theft and vandalism control. Vehicle license plate recognition has been intensively studied in many countries. Due to the different types of license plates being used, the requirement of an automatic license plate recognition system is different for each country. [License plate detection using cluster run length smoothing algorithm ].Generally, an automatic license plate localization and recognition system is made up of three modules; license plate localization, character segmentation and optical character recognition modules. This paper presents an Arabic license plate recognition system that is insensitive to character size, font, shape and orientation with extremely high accuracy rate. The proposed system is based on a combination of enhancement, license plate localization, morphological processing, and feature vector extraction using the Haar transform. The performance of the system is fast due to classification of alphabet and numerals based on the license plate organization. Experimental results for license plates of two different Arab countries show an average of 99 % successful license plate localization and recognition in a total of more than 20 different images captured from a complex outdoor environment. The results run times takes less time compared to conventional and many states of art methods.
Zhang, Jie; Wu, Xiaohong; Yu, Yanmei; Luo, Daisheng
2013-01-01
In optical printed Chinese character recognition (OPCCR), many classifiers have been proposed for the recognition. Among the classifiers, support vector machine (SVM) might be the best classifier. However, SVM is a classifier for two classes. When it is used for multi-classes in OPCCR, its computation is time-consuming. Thus, we propose a neighbor classes based SVM (NC-SVM) to reduce the computation consumption of SVM. Experiments of NC-SVM classification for OPCCR have been done. The results of the experiments have shown that the NC-SVM we proposed can effectively reduce the computation time in OPCCR.
Heuristic algorithm for optical character recognition of Arabic script
NASA Astrophysics Data System (ADS)
Yarman-Vural, Fatos T.; Atici, A.
1996-02-01
In this paper, a heuristic method is developed for segmentation, feature extraction and recognition of the Arabic script. The study is part of a large project for the transcription of the documents in Ottoman Archives. A geometrical and topological feature analysis method is developed for segmentation and feature extraction stages. Chain code transformation is applied to main strokes of the characters which are then classified by the hidden Markov model (HMM) in the recognition stage. Experimental results indicate that the performance of the proposed method is impressive, provided that the thinning process does not yield spurious branches.
Handwritten character recognition using background analysis
NASA Astrophysics Data System (ADS)
Tascini, Guido; Puliti, Paolo; Zingaretti, Primo
1993-04-01
The paper describes a low-cost handwritten character recognizer. It is constituted by three modules: the `acquisition' module, the `binarization' module, and the `core' module. The core module can be logically partitioned into six steps: character dilation, character circumscription, region and `profile' analysis, `cut' analysis, decision tree descent, and result validation. Firstly, it reduces the resolution of the binarized regions and detects the minimum rectangle (MR) which encloses the character; the MR partitions the background into regions that surround the character or are enclosed by it, and allows it to define features as `profiles' and `cuts;' a `profile' is the set of vertical or horizontal minimum distances between a side of the MR and the character itself; a `cut' is a vertical or horizontal image segment delimited by the MR. Then, the core module classifies the character by descending along the decision tree on the basis of the analysis of regions around the character, in particular of the `profiles' and `cuts,' and without using context information. Finally, it recognizes the character or reactivates the core module by analyzing validation test results. The recognizer is largely insensible to character discontinuity and is able to detect Arabic numerals and English alphabet capital letters. The recognition rate of a 32 X 32 pixel character is of about 97% after the first iteration, and of over 98% after the second iteration.
Developing Multimedia Supplementary Materials to Support Learning Beginning Level Chinese Characters
ERIC Educational Resources Information Center
Xu, Lisha
2017-01-01
Studies investigating beginner Chinese learners' character learning strategies found that learners considered orthographic knowledge the most useful factor (Ke, 1998; Shen, 2005). Orthographic recognition correlates with character identification and production and can be used by advanced learners to solve word identification problems (Everson,…
NASA Technical Reports Server (NTRS)
Moore, Steven T.; Cohen, Bernard; Clement, Gilles; Raphan, Theodore
1999-01-01
On Earth, the low-frequency afferent signal from the otoliths encodes head tilt with respect to the gravitational vertical, and the higher frequency components reflect both tilt and linear acceleration of the head. In microgravity, static tilt of the head does not influence otolith output, and the relationship between sensory input from the vestibular organs, and the visual, proprioceptive and somatosensory systems, would be disrupted. Several researchers have proposed that in 0-g this conflict may induce a reinterpretation of all otolith signals by the brain to encode only linear translation (otolith tilt-translation reinterpretation or OTTR). Ocular counter-rolling (OCR) is a low-frequency otolith-mediated reflex, which generates compensatory torsional eye movements (rotation about the visual axis) towards the spatial vertical during static roll tilt with a gain of approximately 10%. Transient linear acceleration and off-axis centrifugation at a constant angular velocity can also generate OCR. According to the OTTR hypothesis, OCR should be reduced in microgravity, and immediately upon return from a 0-g environment. Results to date have been inconclusive. OCR was reduced following the 10 day Spacelab-1 mission in response to leftward roll tilts (28-56% in 3 subjects and unchanged in one subject), and sinusoidal linear oscillations at 0.4 and 0.8 Hz. OCR gain declined 70% in four monkeys following a 14 day COSMOS mission. Following a 30 day MIR mission OCR gain decreased in one astronaut, but increased in two others following a 180 day mission. We have studied the affect of microgravity on low-frequency otolith function as part of a larger study of the interaction of vision and the vestibular system. This experiment (E-047) involved off-axis centrifugation of payload crewmembers and flew aboard the recent Neurolab mission (STS 90). Presented below are preliminary results focusing on perception and the OCR response during both centrifugation and static tilt.
The effect of character contextual diversity on eye movements in Chinese sentence reading.
Chen, Qingrong; Zhao, Guoxia; Huang, Xin; Yang, Yiming; Tanenhaus, Michael K
2017-12-01
Chen, Huang, et al. (Psychonomic Bulletin & Review, 2017) found that when reading two-character Chinese words embedded in sentence contexts, contextual diversity (CD), a measure of the proportion of texts in which a word appears, affected fixation times to words. When CD is controlled, however, frequency did not affect reading times. Two experiments used the same experimental designs to examine whether there are frequency effects of the first character of two-character words when CD is controlled. In Experiment 1, yoked triples of characters from a control group, a group matched for character CD that is lower in frequency, and a group matched in frequency with the control group, but higher in character CD, were rotated through the same sentence frame. In Experiment 2 each character from a larger set was embedded in a separate sentence frame, allowing for a larger difference in log frequency compared to Experiment 1 (0.8 and 0.4, respectively). In both experiments, early and later eye movement measures were significantly shorter for characters with higher CD than for characters with lower CD, with no effects of character frequency. These results place constraints on models of visual word recognition and suggest ways in which Chinese can be used to tease apart the nature of context effects in word recognition and language processing in general.
Segmental Rescoring in Text Recognition
2014-02-04
description relates to rescoring text hypotheses in text recognition based on segmental features. Offline printed text and handwriting recognition (OHR) can... Handwriting , College Park, Md., 2006, which is incorporated by reference here. For the set of training images 202, a character modeler 208 receives
Character displacement of Cercopithecini primate visual signals
Allen, William L.; Stevens, Martin; Higham, James P.
2014-01-01
Animal visual signals have the potential to act as an isolating barrier to prevent interbreeding of populations through a role in species recognition. Within communities of competing species, species recognition signals are predicted to undergo character displacement, becoming more visually distinctive from each other, however this pattern has rarely been identified. Using computational face recognition algorithms to model primate face processing, we demonstrate that the face patterns of guenons (tribe: Cercopithecini) have evolved under selection to become more visually distinctive from those of other guenon species with whom they are sympatric. The relationship between the appearances of sympatric species suggests that distinguishing conspecifics from other guenon species has been a major driver of diversification in guenon face appearance. Visual signals that have undergone character displacement may have had an important role in the tribe’s radiation, keeping populations that became geographically separated reproductively isolated on secondary contact. PMID:24967517
Shape analysis modeling for character recognition
NASA Astrophysics Data System (ADS)
Khan, Nadeem A. M.; Hegt, Hans A.
1998-10-01
Optimal shape modeling of character-classes is crucial for achieving high performance on recognition of mixed-font, hand-written or (and) poor quality text. A novel scheme is presented in this regard focusing on constructing such structural models that can be hierarchically examined. These models utilize a certain `well-thought' set of shape primitives. They are simplified enough to ignore the inter- class variations in font-type or writing style yet retaining enough details for discrimination between the samples of the similar classes. Thus the number of models per class required can be kept minimal without sacrificing the recognition accuracy. In this connection a flexible multi- stage matching scheme exploiting the proposed modeling is also described. This leads to a system which is robust against various distortions and degradation including those related to cases of touching and broken characters. Finally, we present some examples and test results as a proof-of- concept demonstrating the validity and the robustness of the approach.
Cao, Fan; Vu, Marianne; Chan, Derek Ho Lung; Lawrence, Jason M; Harris, Lindsay N; Guan, Qun; Xu, Yi; Perfetti, Charles A
2013-07-01
We examined the hypothesis that learning to write Chinese characters influences the brain's reading network for characters. Students from a college Chinese class learned 30 characters in a character-writing condition and 30 characters in a pinyin-writing condition. After learning, functional magnetic resonance imaging collected during passive viewing showed different networks for reading Chinese characters and English words, suggesting accommodation to the demands of the new writing system through short-term learning. Beyond these expected differences, we found specific effects of character writing in greater activation (relative to pinyin writing) in bilateral superior parietal lobules and bilateral lingual gyri in both a lexical decision and an implicit writing task. These findings suggest that character writing establishes a higher quality representation of the visual-spatial structure of the character and its orthography. We found a greater involvement of bilateral sensori-motor cortex (SMC) for character-writing trained characters than pinyin-writing trained characters in the lexical decision task, suggesting that learning by doing invokes greater interaction with sensori-motor information during character recognition. Furthermore, we found a correlation of recognition accuracy with activation in right superior parietal lobule, right lingual gyrus, and left SMC, suggesting that these areas support the facilitative effect character writing has on reading. Finally, consistent with previous behavioral studies, we found character-writing training facilitates connections with semantics by producing greater activation in bilateral middle temporal gyri, whereas pinyin-writing training facilitates connections with phonology by producing greater activation in right inferior frontal gyrus. Copyright © 2012 Wiley Periodicals, Inc.
Real-time assessment of encapsulated neonatal porcine islets prior to clinical xenotransplantation.
Kitzmann, Jennifer P; Law, Lee; Shome, Avik; Muzina, Marija; Elliott, Robert B; Mueller, Kate R; Schuurman, Henk-Jan; Papas, Klearchos K
2012-01-01
Porcine islet transplantation is emerging as an attractive option for the treatment of patients with type 1 diabetes, with the possibility of providing islets of higher and more consistent quality and in larger volumes than available from human pancreata. The use of encapsulated neonatal porcine islets (ENPI) is appealing because it can address islet supply limitations while reducing the need for anti-rejection therapy. Pre-transplant characterization of ENPI viability and potency is an essential component of the production process. We applied the validated assay for oxygen consumption rate normalized for DNA content (OCR/DNA) to characterize ENPI viability. ENPI of low viscosity and high m alginate were prepared according to standard methods and characterized at various culture time points up to 5 weeks. The OCR/DNA (nmol/min·mgDNA ± SEM) of ENPI (235 ± 10, n = 9) was comparable to that of free NPI (255 ± 14, n = 13). After encapsulation, NPI OCR/DNA was sustained over a culture period of up to 5 weeks. The average OCR/DNA of ENPI cultured longer than 9 days was higher than that of freshly encapsulated NPI. This is the first characterization of ENPI by a validated and more sensitive method for product viability. The NPI encapsulation process does not compromise viability as measured by OCR/DNA, and ENPI can be cultured for up to 5 weeks with maintenance of viability. ENPI meet or exceed current adult porcine islet product release criteria (established at the University of Minnesota) for preclinical xenotransplantation in terms of OCR/DNA. © 2012 John Wiley & Sons A/S.
Oxygen consumption rate of early pre-antral follicles from vitrified human ovarian cortical tissue
ISHIKAWA, Takayuki; KYOYA, Toshihiko; NAKAMURA, Yusuke; SATO, Eimei; TOMIYAMA, Tatsuhiro; KYONO, Koichi
2014-01-01
The study of human ovarian tissue transplantation and cryopreservation has advanced significantly. Autotransplantation of human pre-antral follicles isolated from cryopreserved cortical tissue is a promising option for the preservation of fertility in young cancer patients. The purpose of the present study was to reveal the effect of vitrification after low-temperature transportation of human pre-antral follicles by using the oxygen consumption rate (OCR). Cortical tissues from 9 ovaries of female-to-male transsexuals were vitrified after transportation (6 or 18 h). The follicles were enzymatically isolated from nonvitrified tissue (group I, 18 h of transportation), vitrified-warmed tissue (group II, 6 and 18 h of transportation) and vitrified-warmed tissue that had been incubated for 24 h (group III, 6 and 18 h of transportation). OCR measurement and the LIVE/DEAD viability assay were performed. Despite the ischemic condition, the isolated pre-antral follicles in group I consumed oxygen, and the mean OCRs increased with developmental stage. Neither the transportation time nor patient age seemed to affect the OCR in this group. Meanwhile, the mean OCR was significantly lower (P < 0.05) in group II but was comparable to that of group I after 24 h of incubation. The integrity of vitrified-warmed primordial and primary follicles was clearly corroborated by the LIVE/DEAD viability assay. These results demonstrate that the OCR can be used to directly estimate the effect of vitrification on the viability of primordial and primary follicles and to select the viable primordial and primary follicles from vitrified-warmed follicles. PMID:25262776
Zhang, Jie; Wu, Xiaohong; Yu, Yanmei; Luo, Daisheng
2013-01-01
In optical printed Chinese character recognition (OPCCR), many classifiers have been proposed for the recognition. Among the classifiers, support vector machine (SVM) might be the best classifier. However, SVM is a classifier for two classes. When it is used for multi-classes in OPCCR, its computation is time-consuming. Thus, we propose a neighbor classes based SVM (NC-SVM) to reduce the computation consumption of SVM. Experiments of NC-SVM classification for OPCCR have been done. The results of the experiments have shown that the NC-SVM we proposed can effectively reduce the computation time in OPCCR. PMID:23536777
Hsu, Wei-Chih; Yu, Tsan-Ying; Chen, Kuan-Liang
2009-12-10
Wafer identifications (wafer ID) can be used to identify wafers from each other so that wafer processing can be traced easily. Wafer ID recognition is one of the problems of optical character recognition. The process to recognize wafer IDs is similar to that used in recognizing car license-plate characters. However, due to some unique characteristics, such as the irregular space between two characters and the unsuccessive strokes of wafer ID, it will not get a good result to recognize wafer ID by directly utilizing the approaches used in car license-plate character recognition. Wafer ID scratches are engraved by a laser scribe almost along the following four fixed directions: horizontal, vertical, plus 45 degrees , and minus 45 degrees orientations. The closer to the center line of a wafer ID scratch, the higher the gray level will be. These and other characteristics increase the difficulty to recognize the wafer ID. In this paper a wafer ID recognition scheme based on an asterisk-shape filter and a high-low score comparison method is proposed to cope with the serious influence of uneven luminance and make recognition more efficiently. Our proposed approach consists of some processing stages. Especially in the final recognition stage, a template-matching method combined with stroke analysis is used as a recognizing scheme. This is because wafer IDs are composed of Semiconductor Equipment and Materials International (SEMI) standard Arabic numbers and English alphabets, and thus the template ID images are easy to obtain. Furthermore, compared with the approach that requires prior training, such as a support vector machine, which often needs a large amount of training image samples, no prior training is required for our approach. The testing results show that our proposed scheme can efficiently and correctly segment out and recognize the wafer ID with high performance.
A Large Scale Code Resolution Service Network in the Internet of Things
Yu, Haining; Zhang, Hongli; Fang, Binxing; Yu, Xiangzhan
2012-01-01
In the Internet of Things a code resolution service provides a discovery mechanism for a requester to obtain the information resources associated with a particular product code immediately. In large scale application scenarios a code resolution service faces some serious issues involving heterogeneity, big data and data ownership. A code resolution service network is required to address these issues. Firstly, a list of requirements for the network architecture and code resolution services is proposed. Secondly, in order to eliminate code resolution conflicts and code resolution overloads, a code structure is presented to create a uniform namespace for code resolution records. Thirdly, we propose a loosely coupled distributed network consisting of heterogeneous, independent; collaborating code resolution services and a SkipNet based code resolution service named SkipNet-OCRS, which not only inherits DHT's advantages, but also supports administrative control and autonomy. For the external behaviors of SkipNet-OCRS, a novel external behavior mode named QRRA mode is proposed to enhance security and reduce requester complexity. For the internal behaviors of SkipNet-OCRS, an improved query algorithm is proposed to increase query efficiency. It is analyzed that integrating SkipNet-OCRS into our resolution service network can meet our proposed requirements. Finally, simulation experiments verify the excellent performance of SkipNet-OCRS. PMID:23202207
Sim, Ida; Tu, Samson W.; Carini, Simona; Lehmann, Harold P.; Pollock, Brad H.; Peleg, Mor; Wittkowski, Knut M.
2013-01-01
To date, the scientific process for generating, interpreting, and applying knowledge has received less informatics attention than operational processes for conducting clinical studies. The activities of these scientific processes — the science of clinical research — are centered on the study protocol, which is the abstract representation of the scientific design of a clinical study. The Ontology of Clinical Research (OCRe) is an OWL 2 model of the entities and relationships of study design protocols for the purpose of computationally supporting the design and analysis of human studies. OCRe’s modeling is independent of any specific study design or clinical domain. It includes a study design typology and a specialized module called ERGO Annotation for capturing the meaning of eligibility criteria. In this paper, we describe the key informatics use cases of each phase of a study’s scientific lifecycle, present OCRe and the principles behind its modeling, and describe applications of OCRe and associated technologies to a range of clinical research use cases. OCRe captures the central semantics that underlies the scientific processes of clinical research and can serve as an informatics foundation for supporting the entire range of knowledge activities that constitute the science of clinical research. PMID:24239612
Kitzmann, JP; O’Gorman, D; Kin, T; Gruessner, AC; Senior, P; Imes, S; Gruessner, RW; Shapiro, AMJ; Papas, KK
2014-01-01
Human islet allotransplant (ITx) for the treatment of type 1 diabetes is in phase III clinical registration trials in the US and standard of care in several other countries. Current islet product release criteria include viability based on cell membrane integrity stains, glucose stimulated insulin release (GSIR), and islet equivalent (IE) dose based on counts. However, only a fraction of patients transplanted with islets that meet or exceed these release criteria become insulin independent following one transplant. Measurements of islet oxygen consumption rate (OCR) have been reported as highly predictive of transplant outcome in many models. In this paper we report on the assessment of clinical islet allograft preparations using islet oxygen consumption rate (OCR) dose (or viable IE dose) and current product release assays in a series of 13 first transplant recipients. The predictive capability of each assay was examined and successful graft function was defined as 100% insulin independence within 45 days post-transplant. Results showed that OCR dose was most predictive of CTO. IE dose was also highly predictive, while GSIR and membrane integrity stains were not. In conclusion, OCR dose can predict CTO with high specificity and sensitivity and is a useful tool for evaluating islet preparations prior to clinical ITx. PMID:25131089
A large scale code resolution service network in the Internet of Things.
Yu, Haining; Zhang, Hongli; Fang, Binxing; Yu, Xiangzhan
2012-11-07
In the Internet of Things a code resolution service provides a discovery mechanism for a requester to obtain the information resources associated with a particular product code immediately. In large scale application scenarios a code resolution service faces some serious issues involving heterogeneity, big data and data ownership. A code resolution service network is required to address these issues. Firstly, a list of requirements for the network architecture and code resolution services is proposed. Secondly, in order to eliminate code resolution conflicts and code resolution overloads, a code structure is presented to create a uniform namespace for code resolution records. Thirdly, we propose a loosely coupled distributed network consisting of heterogeneous, independent; collaborating code resolution services and a SkipNet based code resolution service named SkipNet-OCRS, which not only inherits DHT’s advantages, but also supports administrative control and autonomy. For the external behaviors of SkipNet-OCRS, a novel external behavior mode named QRRA mode is proposed to enhance security and reduce requester complexity. For the internal behaviors of SkipNet-OCRS, an improved query algorithm is proposed to increase query efficiency. It is analyzed that integrating SkipNet-OCRS into our resolution service network can meet our proposed requirements. Finally, simulation experiments verify the excellent performance of SkipNet-OCRS.
Glycolysis-respiration relationships in a neuroblastoma cell line.
Swerdlow, Russell H; E, Lezi; Aires, Daniel; Lu, Jianghua
2013-04-01
Although some reciprocal glycolysis-respiration relationships are well recognized, the relationship between reduced glycolysis flux and mitochondrial respiration has not been critically characterized. We concomitantly measured the extracellular acidification rate (ECAR) and oxygen consumption rate (OCR) of SH-SY5Y neuroblastoma cells under free and restricted glycolysis flux conditions. Under conditions of fixed energy demand ECAR and OCR values showed a reciprocal relationship. In addition to observing an expected Crabtree effect in which increasing glucose availability raised the ECAR and reduced the OCR, a novel reciprocal relationship was documented in which reducing the ECAR via glucose deprivation or glycolysis inhibition increased the OCR. Substituting galactose for glucose, which reduces net glycolysis ATP yield without blocking glycolysis flux, similarly reduced the ECAR and increased the OCR. We further determined how reduced ECAR conditions affect proteins that associate with energy sensing and energy response pathways. ERK phosphorylation, SIRT1, and HIF1a decreased while AKT, p38, and AMPK phosphorylation increased. These data document a novel intracellular glycolysis-respiration effect in which restricting glycolysis flux increases mitochondrial respiration. Since this effect can be used to manipulate cell bioenergetic infrastructures, this particular glycolysis-respiration effect can practically inform the development of new mitochondrial medicine approaches. Copyright © 2012 Elsevier B.V. All rights reserved.
The activation of segmental and tonal information in visual word recognition.
Li, Chuchu; Lin, Candise Y; Wang, Min; Jiang, Nan
2013-08-01
Mandarin Chinese has a logographic script in which graphemes map onto syllables and morphemes. It is not clear whether Chinese readers activate phonological information during lexical access, although phonological information is not explicitly represented in Chinese orthography. In the present study, we examined the activation of phonological information, including segmental and tonal information in Chinese visual word recognition, using the Stroop paradigm. Native Mandarin speakers named the presentation color of Chinese characters in Mandarin. The visual stimuli were divided into five types: color characters (e.g., , hong2, "red"), homophones of the color characters (S+T+; e.g., , hong2, "flood"), different-tone homophones (S+T-; e.g., , hong1, "boom"), characters that shared the same tone but differed in segments with the color characters (S-T+; e.g., , ping2, "bottle"), and neutral characters (S-T-; e.g., , qian1, "leading through"). Classic Stroop facilitation was shown in all color-congruent trials, and interference was shown in the incongruent trials. Furthermore, the Stroop effect was stronger for S+T- than for S-T+ trials, and was similar between S+T+ and S+T- trials. These findings suggested that both tonal and segmental forms of information play roles in lexical constraints; however, segmental information has more weight than tonal information. We proposed a revised visual word recognition model in which the functions of both segmental and suprasegmental types of information and their relative weights are taken into account.
NASA Astrophysics Data System (ADS)
Rishi, Rahul; Choudhary, Amit; Singh, Ravinder; Dhaka, Vijaypal Singh; Ahlawat, Savita; Rao, Mukta
2010-02-01
In this paper we propose a system for classification problem of handwritten text. The system is composed of preprocessing module, supervised learning module and recognition module on a very broad level. The preprocessing module digitizes the documents and extracts features (tangent values) for each character. The radial basis function network is used in the learning and recognition modules. The objective is to analyze and improve the performance of Multi Layer Perceptron (MLP) using RBF transfer functions over Logarithmic Sigmoid Function. The results of 35 experiments indicate that the Feed Forward MLP performs accurately and exhaustively with RBF. With the change in weight update mechanism and feature-drawn preprocessing module, the proposed system is competent with good recognition show.
The Effects of Graphic Similarity on Japanese Recognition of Simplified Chinese Characters
ERIC Educational Resources Information Center
Teng, Xiaochun; Yamada, Jun
2017-01-01
The pedagogical and theoretical questions addressed in this study relate to the extent to which native Japanese readers with little or no knowledge of Chinese characters recognize Chinese characters that are viewed as abbreviations of the kanji they already know. Three graphic similarity functions (i.e., an orthographically acceptable similarity,…
Drury, J. P.; Grether, G. F.
2014-01-01
Traits that mediate intraspecific social interactions may overlap in closely related sympatric species, resulting in costly between-species interactions. Such interactions have principally interested investigators studying the evolution of reproductive isolation via reproductive character displacement (RCD) or reinforcement, yet in addition to reproductive interference, interspecific trait overlap can lead to costly between-species aggression. Previous research on rubyspot damselflies (Hetaerina spp.) demonstrated that sympatric shifts in male wing colour patterns and competitor recognition reduce interspecific aggression, supporting the hypothesis that agonistic character displacement (ACD) drove trait shifts. However, a recent theoretical model shows that RCD overshadows ACD if the same male trait is used for both female mate recognition and male competitor recognition. To determine whether female mate recognition is based on male wing coloration in Hetaerina, we conducted a phenotype manipulation experiment. Compared to control males, male H. americana with wings manipulated to resemble a sympatric congener (H. titia) suffered no reduction in mating success. Thus, female mate recognition is not based on species differences in male wing coloration. Experimental males did, however, experience higher interspecific fighting rates and reduced survival compared to controls. These results greatly strengthen the case for ACD and highlight the mechanistic distinction between ACD and RCD. PMID:25339724
NASA Astrophysics Data System (ADS)
Morris, Phillip A.
The prevalence of low-cost side scanning sonar systems mounted on small recreational vessels has created improved opportunities to identify and map submerged navigational hazards in freshwater impoundments. However, these economical sensors also present unique challenges for automated techniques. This research explores related literature in automated sonar imagery processing and mapping technology, proposes and implements a framework derived from these sources, and evaluates the approach with video collected from a recreational grade sonar system. Image analysis techniques including optical character recognition and an unsupervised computer automated detection (CAD) algorithm are employed to extract the transducer GPS coordinates and slant range distance of objects protruding from the lake bottom. The retrieved information is formatted for inclusion into a spatial mapping model. Specific attributes of the sonar sensors are modeled such that probability profiles may be projected onto a three dimensional gridded map. These profiles are computed from multiple points of view as sonar traces crisscross or come near each other. As lake levels fluctuate over time so do the elevation points of view. With each sonar record, the probability of a hazard existing at certain elevations at the respective grid points is updated with Bayesian mechanics. As reinforcing data is collected, the confidence of the map improves. Given a lake's current elevation and a vessel draft, a final generated map can identify areas of the lake that have a high probability of containing hazards that threaten navigation. The approach is implemented in C/C++ utilizing OpenCV, Tesseract OCR, and QGIS open source software and evaluated in a designated test area at Lake Lavon, Collin County, Texas.
Rescuing Paleomagnetic Data from Deep-Sea Cores Through the IEDA-CCNY Data Internship Program
NASA Astrophysics Data System (ADS)
Ismail, A.; Randel, C.; Palumbo, R. V.; Carter, M.; Cai, Y.; Kent, D. V.; Lehnert, K.; Block, K. A.
2016-12-01
Paleomagnetic data provides essential information for evaluating the chronostratigraphy of sedimentary cores. Lamont research vessels Vema and Robert Conrad collected over 10,000 deep-sea sediment cores around the world from 1953 to 1989. 10% of these cores have been sampled for paleomagnetic analyses at Lamont. Over the years, only 10% of these paleomagnetic records have been published. Moreover, data listings were only rarely made available in older publications because electronic appendices were not available and cyberinfrastructure was not in place for publishing and preserving these data. As a result, the majority of these datasets exist only as fading computer printouts in binders on the investigator's bookshelf. This summer, undergraduate students from the NSF-funded IEDA-CCNY Data Internship Program started digitizing this enormous dataset under the supervision of Dennis Kent, the current custodian of the data and one of the investigators who oversaw some of the data collection process, and an active leader in the field. Undergraduate students worked on digitizing paper records, proof-reading and organizing the data sheets for future integration into an appropriate repository. Through observing and plotting the data, the students learned about how sediment cores and paleomagnetic data are collected and used in research, and the best practices in data publishing and preservation from IEDA (Interdisciplinary Earth Data Alliance) team members. The students also compared different optical character recognition (OCR) softwares and established an efficient workflow to digitize these datasets. These datasets will eventually be incorporated in the Magnetics Information Consortium (MagIC), so that they can be easily compared with similar datasets and have the potential to generate new findings. Through this data rescue project, the students had the opportunity to learn about an important field of scientific research and interact with world-class scientists.
De-identification of unstructured paper-based health records for privacy-preserving secondary use.
Fenz, Stefan; Heurix, Johannes; Neubauer, Thomas; Rella, Antonio
2014-07-01
Abstract Whenever personal data is processed, privacy is a serious issue. Especially in the document-centric e-health area, the patients' privacy must be preserved in order to prevent any negative repercussions for the patient. Clinical research, for example, demands structured health records to carry out efficient clinical trials, whereas legislation (e.g. HIPAA) regulates that only de-identified health records may be used for research. However, unstructured and often paper-based data dominates information technology, especially in the healthcare sector. Existing approaches are geared towards data in English-language documents only and have not been designed to handle the recognition of erroneous personal data which is the result of the OCR-based digitization of paper-based health records.
Fast title extraction method for business documents
NASA Astrophysics Data System (ADS)
Katsuyama, Yutaka; Naoi, Satoshi
1997-04-01
Conventional electronic document filing systems are inconvenient because the user must specify the keywords in each document for later searches. To solve this problem, automatic keyword extraction methods using natural language processing and character recognition have been developed. However, these methods are slow, especially for japanese documents. To develop a practical electronic document filing system, we focused on the extraction of keyword areas from a document by image processing. Our fast title extraction method can automatically extract titles as keywords from business documents. All character strings are evaluated for similarity by rating points associated with title similarity. We classified these points as four items: character sitting size, position of character strings, relative position among character strings, and string attribution. Finally, the character string that has the highest rating is selected as the title area. The character recognition process is carried out on the selected area. It is fast because this process must recognize a small number of patterns in the restricted area only, and not throughout the entire document. The mean performance of this method is an accuracy of about 91 percent and a 1.8 sec. processing time for an examination of 100 Japanese business documents.
Imaging Prostate Cancer (Pca) Phenotype and Evolution
2014-10-01
Extracellular flux analysis experiments with the Seahorse system showed a marked decrease in OCR after inhibition of ATP synthase by oligomycin...measured in each well 34 h after seeding the cells, using the Seahorse extracellular flux analyzer, as also described in Methods section. OCR
Integrative Lifecourse and Genetic Analysis of Military Working Dogs
2012-10-01
Recognition), ICR (Intelligent Character Recognition) and HWR ( Handwriting Recognition). A number of various software packages were evaluated and we have...the third-party software is able to recognize check-boxes and columns and do a reasonable job with handwriting – which is does. This workflow will
Gremlin 1 Identifies a Skeletal Stem Cell with Bone, Cartilage, and Reticular Stromal Potential
Worthley, Daniel L.; Churchill, Michael; Compton, Jocelyn T.; Tailor, Yagnesh; Rao, Meenakshi; Si, Yiling; Levin, Daniel; Schwartz, Matthew G.; Uygur, Aysu; Hayakawa, Yoku; Gross, Stefanie; Renz, Bernhard W.; Setlik, Wanda; Martinez, Ashley N.; Chen, Xiaowei; Nizami, Saqib; Lee, Heon Goo; Kang, H. Paco; Caldwell, Jon-Michael; Asfaha, Samuel; Westphalen, C. Benedikt; Graham, Trevor; Jin, Guangchun; Nagar, Karan; Wang, Hongshan; Kheirbek, Mazen A.; Kolhe, Alka; Carpenter, Jared; Glaire, Mark; Nair, Abhinav; Renders, Simon; Manieri, Nicholas; Muthupalani, Sureshkumar; Fox, James G.; Reichert, Maximilian; Giraud, Andrew S.; Schwabe, Robert F.; Pradere, Jean-Phillipe; Walton, Katherine; Prakash, Ajay; Gumucio, Deborah; Rustgi, Anil K.; Stappenbeck, Thaddeus S.; Friedman, Richard A.; Gershon, Michael D.; Sims, Peter; Grikscheit, Tracy; Lee, Francis Y.; Karsenty, Gerard; Mukherjee, Siddhartha; Wang, Timothy C.
2014-01-01
The stem cells that maintain and repair the postnatal skeleton remain undefined. One model suggests that perisinusoidal mesenchymal stem cells (MSCs) give rise to osteoblasts, chondrocytes, marrow stromal cells, and adipocytes, although the existence of these cells has not been proven through fate-mapping experiments. We demonstrate here that expression of the bone morphogenetic protein (BMP) antagonist gremlin 1 defines a population of osteochondroreticular (OCR) stem cells in the bone marrow. OCR stem cells self-renew and generate osteoblasts, chondrocytes, and reticular marrow stromal cells, but not adipocytes. OCR stem cells are concentrated within the metaphysis of long bones not in the perisinusoidal space and are needed for bone development, bone remodeling, and fracture repair. Grem1 expression also identifies intestinal reticular stem cells (iRSCs) that are cells of origin for the periepithelial intestinal mesenchymal sheath. Grem1 expression identifies distinct connective tissue stem cells in both the bone (OCR stem cells) and the intestine (iRSCs). PMID:25594183
Enhancing vestibular function in the elderly with imperceptible electrical stimulation.
Serrador, Jorge M; Deegan, Brian M; Geraghty, Maria C; Wood, Scott J
2018-01-10
Age-related loss of vestibular function can result in decrements in gaze stabilization and increased fall risk in the elderly. This study was designed to see if low levels of electrical stochastic noise applied transcutaneously to the vestibular system can improve a gaze stabilization reflex in young and elderly subject groups. Ocular counter-rolling (OCR) using a video-based technique was obtained in 16 subjects during low frequency passive roll tilts. Consistent with previous studies, there was a significant reduction in OCR gains in the elderly compared to the young group. Imperceptible stochastic noise significantly increased OCR in the elderly (Mean 23%, CI: 17-35%). Increases in OCR gain were greatest for those with lowest baseline gain and were negligible in those with normal gain. Since stimulation was effective at low levels undetectable to subjects, stochastic noise may provide a new treatment alternative to enhance vestibular function, specifically otolith-ocular reflexes, in the elderly or patient populations with reduced otolith-ocular function.
Grabež, V; Kathri, M; Phung, V; Moe, K M; Slinde, E; Skaugen, M; Saarem, K; Egelandsdal, B
2015-04-01
Oxygen consumption rate (OCR) of muscle fibers from bovine semimembranosus muscle of 41 animals was investigated 3 to 4 h and 3 wk postmortem. Significant relations (P < 0.05) were found between OCR measurements and Warner-Bratzler shear force measurement. Muscles with high mitochondrial OCR after 3 to 4 h and low nonmitochondrial oxygen consumption gave more tender meat. Tender (22.92 ± 2.2 N/cm2) and tough (72.98 ± 7.2 N/cm2) meat samples (4 samples each), separated based on their OCR measurements, were selected for proteomic studies using mitochondria isolated approximately 2.5 h postmortem. Twenty-six differently expressed proteins (P < 0.05) were identified in tender meat and 19 in tough meat. In tender meat, the more prevalent antioxidant and chaperon enzymes may reduce reactive oxygen species and prolong oxygen removal by the electron transport system (ETS). Glycolytic, Krebs cycle, and ETS enzymes were also more abundant in tender meat
ERIC Educational Resources Information Center
Wang, Hsueh-Cheng; Schotter, Elizabeth R.; Angele, Bernhard; Yang, Jinmian; Simovici, Dan; Pomplun, Marc; Rayner, Keith
2013-01-01
Previous research indicates that removing initial strokes from Chinese characters makes them harder to read than removing final or internal ones. In the present study, we examined the contribution of important components to character configuration via singular value decomposition. The results indicated that when the least important segments, which…
Interpretation of In situ Testing of Cohesive Soils Using Rational Methods
1990-11-01
a circular steel diaphragm located on the flat face of the device. Section 5 describes a comprehensive analysis of the contact pressure for...9o II Z * 0~ 00 0 o- V 0 - V ( Cl c~0 114 U+ e 0 00 0 cn 0 - t 00 r- UI en tn m i > 0 3 115 WATER SEAL- SOIL SEAL UFRICTION SLEEVE BEARING LOAD CELL...difference with OCR (Sully et al., 1987) 127 5 0 1.0/F IV / r 4 Intact M50 0.4 0j 1 0 1 05 0 OCR * Figure 4.1 a Correlation of pore pressure ratio with OCR
NASA Technical Reports Server (NTRS)
Hong, J. P.
1971-01-01
Technique operates regardless of pattern rotation, translation or magnification and successfully detects out-of-register patterns. It improves accuracy and reduces cost of various optical character recognition devices and page readers and provides data input to computer.
Spatiotemporal Pixelization to Increase the Recognition Score of Characters for Retinal Prostheses
Kim, Hyun Seok; Park, Kwang Suk
2017-01-01
Most of the retinal prostheses use a head-fixed camera and a video processing unit. Some studies proposed various image processing methods to improve visual perception for patients. However, previous studies only focused on using spatial information. The present study proposes a spatiotemporal pixelization method mimicking fixational eye movements to generate stimulation images for artificial retina arrays by combining spatial and temporal information. Input images were sampled with a resolution that was four times higher than the number of pixel arrays. We subsampled this image and generated four different phosphene images. We then evaluated the recognition scores of characters by sequentially presenting phosphene images with varying pixel array sizes (6 × 6, 8 × 8 and 10 × 10) and stimulus frame rates (10 Hz, 15 Hz, 20 Hz, 30 Hz, and 60 Hz). The proposed method showed the highest recognition score at a stimulus frame rate of approximately 20 Hz. The method also significantly improved the recognition score for complex characters. This method provides a new way to increase practical resolution over restricted spatial resolution by merging the higher resolution image into high-frame time slots. PMID:29073735
Hypercorticism blunts circadian variations of osteocalcin regardless of nutritional status.
Vergély, N; Lafage-Proust, M-H; Caillot-Augusseau, A; Millot, L; Lang, F; Estour, B
2002-02-01
Anorexia nervosa (AN) and Cushing's syndrome (CS) are both responsible for osteoporosis. The mechanisms leading to osteoporosis in AN include hypogonadism, nutritional depletion, and in some cases hypercorticism. Osteocalcin circulating level is a serum marker of osteoblastic activity that follows a circadian rhythm (OCR). Serum osteocalcin is decreased in both CS and AN and can be increased with treatment. In this study we analyzed the influence of combined cortisol and nutritional status on osteocalcin levels and its circadian rhythm in these two different models of hypercorticism, one nutritionally replete (CS) and one nutritionally deplete (AN), and we evaluated the effects of their treatment (surgical cure and weight gain, respectively). Before treatment, osteocalcin levels were lower in CS (n = 16) and AN (n = 42) than in controls and in the AN patient subgroup with hypercorticism (n = 13) compared to those without (n = 29). OCR was absent in CS and in AN patients with hypercorticism, whereas their circadian cortisol cycle was maintained. In CS, successful surgical treatment increased osteocalcin levels (n = 5) and restored OCR. In AN, weight gain (n = 13) induced a significant decrease in cortisol levels in hypercortisolic AN patients, and restored normal osteocalcin levels and OCR. In conclusion, we found that hypercorticism was associated with a decrease in osteocalcin levels in nutritionally replete or deplete patients and that OCR was more affected by cortisol levels than by cortisol cycle.
A Functional Nuclear Localization Sequence in the C. elegans TRPV Channel OCR-2
Ezak, Meredith J.; Ferkey, Denise M.
2011-01-01
The ability to modulate gene expression in response to sensory experience is critical to the normal development and function of the nervous system. Calcium is a key activator of the signal transduction cascades that mediate the process of translating a cellular stimulus into transcriptional changes. With the recent discovery that the mammalian Cav1.2 calcium channel can be cleaved, enter the nucleus and act as a transcription factor to control neuronal gene expression, a more direct role for the calcium channels themselves in regulating transcription has begun to be appreciated. Here we report the identification of a nuclear localization sequence (NLS) in the C. elegans transient receptor potential vanilloid (TRPV) cation channel OCR-2. TRPV channels have previously been implicated in transcriptional regulation of neuronal genes in the nematode, although the precise mechanism remains unclear. We show that the NLS in OCR-2 is functional, being able to direct nuclear accumulation of a synthetic cargo protein as well as the carboxy-terminal cytosolic tail of OCR-2 where it is endogenously found. Furthermore, we discovered that a carboxy-terminal portion of the full-length channel can localize to the nucleus of neuronal cells. These results suggest that the OCR-2 TRPV cation channel may have a direct nuclear function in neuronal cells that was not previously appreciated. PMID:21957475
reCAPTCHA: human-based character recognition via Web security measures.
von Ahn, Luis; Maurer, Benjamin; McMillen, Colin; Abraham, David; Blum, Manuel
2008-09-12
CAPTCHAs (Completely Automated Public Turing test to tell Computers and Humans Apart) are widespread security measures on the World Wide Web that prevent automated programs from abusing online services. They do so by asking humans to perform a task that computers cannot yet perform, such as deciphering distorted characters. Our research explored whether such human effort can be channeled into a useful purpose: helping to digitize old printed material by asking users to decipher scanned words from books that computerized optical character recognition failed to recognize. We showed that this method can transcribe text with a word accuracy exceeding 99%, matching the guarantee of professional human transcribers. Our apparatus is deployed in more than 40,000 Web sites and has transcribed over 440 million words.
Ensuring Equal Access to High-Quality Education. Revised
ERIC Educational Resources Information Center
Office for Civil Rights, US Department of Education, 2011
2011-01-01
The Office for Civil Rights (OCR) in the U.S. Department of Education (Department) is a law enforcement agency charged with enforcing federal civil rights laws to ensure that educational institutions receiving federal financial assistance do not engage in discriminatory conduct. OCR enforces the federal civil rights laws that prohibit…
Cost-Effective CNC Part Program Verification Development for Laboratory Instruction.
ERIC Educational Resources Information Center
Chen, Joseph C.; Chang, Ted C.
2000-01-01
Describes a computer numerical control program verification system that checks a part program before its execution. The system includes character recognition, word recognition, a fuzzy-nets system, and a tool path viewer. (SK)
ERIC Educational Resources Information Center
Yeh, Su-Ling; Li, Jing-Ling
2004-01-01
Repetition blindness (RB) refers to the failure to detect the second occurrence of a repeated item in rapid serial visual presentation (RSVP). In two experiments using RSVP, the ability to report two critical characters was found to be impaired when these two characters were identical (Experiment 1) or similar by sharing one repeated component…
Cognitive Processing Hardware Elements
2005-01-31
characters. Results will be presented below. 1 4. Recognition of human faces. There are many other possible applications such as facial recognition and...For the experiments in facial recognition , we have used a 3-layer autoassociative neural network having the following specifications: "* The input...using the facial recognition system described in the section above as an example. This system uses an autoassociative neural network containing over 10
Recognition is Used as One Cue Among Others in Judgment and Decision Making
ERIC Educational Resources Information Center
Richter, Tobias; Spath, Pamela
2006-01-01
Three experiments with paired comparisons were conducted to test the noncompensatory character of the recognition heuristic (D. G. Goldstein & G. Gigerenzer, 2002) in judgment and decision making. Recognition and knowledge about the recognized alternative were manipulated. In Experiment 1, participants were presented pairs of animal names where…
Cheng, Juan; Chen, Xun; Liu, Aiping; Peng, Hu
2015-01-01
Sign language recognition (SLR) is an important communication tool between the deaf and the external world. It is highly necessary to develop a worldwide continuous and large-vocabulary-scale SLR system for practical usage. In this paper, we propose a novel phonology- and radical-coded Chinese SLR framework to demonstrate the feasibility of continuous SLR using accelerometer (ACC) and surface electromyography (sEMG) sensors. The continuous Chinese characters, consisting of coded sign gestures, are first segmented into active segments using EMG signals by means of moving average algorithm. Then, features of each component are extracted from both ACC and sEMG signals of active segments (i.e., palm orientation represented by the mean and variance of ACC signals, hand movement represented by the fixed-point ACC sequence, and hand shape represented by both the mean absolute value (MAV) and autoregressive model coefficients (ARs)). Afterwards, palm orientation is first classified, distinguishing “Palm Downward” sign gestures from “Palm Inward” ones. Only the “Palm Inward” gestures are sent for further hand movement and hand shape recognition by dynamic time warping (DTW) algorithm and hidden Markov models (HMM) respectively. Finally, component recognition results are integrated to identify one certain coded gesture. Experimental results demonstrate that the proposed SLR framework with a vocabulary scale of 223 characters can achieve an averaged recognition accuracy of 96.01% ± 0.83% for coded gesture recognition tasks and 92.73% ± 1.47% for character recognition tasks. Besides, it demonstrats that sEMG signals are rather consistent for a given hand shape independent of hand movements. Hence, the number of training samples will not be significantly increased when the vocabulary scale increases, since not only the number of the completely new proposed coded gestures is constant and limited, but also the transition movement which connects successive signs needs no training samples to model even though the same coded gesture performed in different characters. This work opens up a possible new way to realize a practical Chinese SLR system. PMID:26389907
Cheng, Juan; Chen, Xun; Liu, Aiping; Peng, Hu
2015-09-15
Sign language recognition (SLR) is an important communication tool between the deaf and the external world. It is highly necessary to develop a worldwide continuous and large-vocabulary-scale SLR system for practical usage. In this paper, we propose a novel phonology- and radical-coded Chinese SLR framework to demonstrate the feasibility of continuous SLR using accelerometer (ACC) and surface electromyography (sEMG) sensors. The continuous Chinese characters, consisting of coded sign gestures, are first segmented into active segments using EMG signals by means of moving average algorithm. Then, features of each component are extracted from both ACC and sEMG signals of active segments (i.e., palm orientation represented by the mean and variance of ACC signals, hand movement represented by the fixed-point ACC sequence, and hand shape represented by both the mean absolute value (MAV) and autoregressive model coefficients (ARs)). Afterwards, palm orientation is first classified, distinguishing "Palm Downward" sign gestures from "Palm Inward" ones. Only the "Palm Inward" gestures are sent for further hand movement and hand shape recognition by dynamic time warping (DTW) algorithm and hidden Markov models (HMM) respectively. Finally, component recognition results are integrated to identify one certain coded gesture. Experimental results demonstrate that the proposed SLR framework with a vocabulary scale of 223 characters can achieve an averaged recognition accuracy of 96.01% ± 0.83% for coded gesture recognition tasks and 92.73% ± 1.47% for character recognition tasks. Besides, it demonstrats that sEMG signals are rather consistent for a given hand shape independent of hand movements. Hence, the number of training samples will not be significantly increased when the vocabulary scale increases, since not only the number of the completely new proposed coded gestures is constant and limited, but also the transition movement which connects successive signs needs no training samples to model even though the same coded gesture performed in different characters. This work opens up a possible new way to realize a practical Chinese SLR system.
Recognition of degraded handwritten digits using dynamic Bayesian networks
NASA Astrophysics Data System (ADS)
Likforman-Sulem, Laurence; Sigelle, Marc
2007-01-01
We investigate in this paper the application of dynamic Bayesian networks (DBNs) to the recognition of handwritten digits. The main idea is to couple two separate HMMs into various architectures. First, a vertical HMM and a horizontal HMM are built observing the evolving streams of image columns and image rows respectively. Then, two coupled architectures are proposed to model interactions between these two streams and to capture the 2D nature of character images. Experiments performed on the MNIST handwritten digit database show that coupled architectures yield better recognition performances than non-coupled ones. Additional experiments conducted on artificially degraded (broken) characters demonstrate that coupled architectures better cope with such degradation than non coupled ones and than discriminative methods such as SVMs.
Longcamp, Marieke; Boucard, Céline; Gilhodes, Jean-Claude; Velay, Jean-Luc
2006-10-01
Recent data support the idea that movements play a crucial role in letter representation and suggest that handwriting knowledge contributes to visual recognition of letters. If so, using different motor activities while subjects are learning to write should affect their subsequent recognition performances. In order to test this hypothesis, we trained adult participants to write new characters either by copying them or by typing them on a keyboard. After three weeks of training we ran a series of tests requiring visual processing of the characters' orientation. Tests were ran immediately, one week after, and three weeks after the end of the training period. Results showed that when the characters had been learned by typing, they were more frequently confused with their mirror images than when they had been written by hand. This handwriting advantage did not appear immediately, but mostly three weeks after the end of the training. Our results therefore suggest that the stability of the characters' representation in memory depends on the nature of the motor activity produced during learning.
Artificial neural networks for document analysis and recognition.
Marinai, Simone; Gori, Marco; Soda, Giovanni; Society, Computer
2005-01-01
Artificial neural networks have been extensively applied to document analysis and recognition. Most efforts have been devoted to the recognition of isolated handwritten and printed characters with widely recognized successful results. However, many other document processing tasks, like preprocessing, layout analysis, character segmentation, word recognition, and signature verification, have been effectively faced with very promising results. This paper surveys the most significant problems in the area of offline document image processing, where connectionist-based approaches have been applied. Similarities and differences between approaches belonging to different categories are discussed. A particular emphasis is given on the crucial role of prior knowledge for the conception of both appropriate architectures and learning algorithms. Finally, the paper provides a critical analysis on the reviewed approaches and depicts the most promising research guidelines in the field. In particular, a second generation of connectionist-based models are foreseen which are based on appropriate graphical representations of the learning environment.
Address entry while driving: speech recognition versus a touch-screen keyboard.
Tsimhoni, Omer; Smith, Daniel; Green, Paul
2004-01-01
A driving simulator experiment was conducted to determine the effects of entering addresses into a navigation system during driving. Participants drove on roads of varying visual demand while entering addresses. Three address entry methods were explored: word-based speech recognition, character-based speech recognition, and typing on a touch-screen keyboard. For each method, vehicle control and task measures, glance timing, and subjective ratings were examined. During driving, word-based speech recognition yielded the shortest total task time (15.3 s), followed by character-based speech recognition (41.0 s) and touch-screen keyboard (86.0 s). The standard deviation of lateral position when performing keyboard entry (0.21 m) was 60% higher than that for all other address entry methods (0.13 m). Degradation of vehicle control associated with address entry using a touch screen suggests that the use of speech recognition is favorable. Speech recognition systems with visual feedback, however, even with excellent accuracy, are not without performance consequences. Applications of this research include the design of in-vehicle navigation systems as well as other systems requiring significant driver input, such as E-mail, the Internet, and text messaging.
ERIC Educational Resources Information Center
Office for Civil Rights, US Department of Education, 2016
2016-01-01
In FY 2016, the Office for Civil Rights (OCR) promoted equity and excellence in education through conducting investigations and monitoring schools under resolution agreements, promoting greater understanding of how OCR interprets and enforces civil rights laws through the release of policy guidance, providing technical assistance and outreach to…
Office for Civil Rights. Annual Report to Congress, Fiscal Year 1994.
ERIC Educational Resources Information Center
Office for Civil Rights (ED), Washington, DC.
The Office of Civil Rights (OCR) enforces the laws that prevent discrimination on the basis of race, national origin, sex, disability, and age in America's schools, colleges, and universities. To gauge how the department is fulfilling that mission, a description of the OCR's policy-guidance efforts, complaint investigations, and enforcement…
Office for Civil Rights Annual Report to Congress, Fiscal Years 2007-08
ERIC Educational Resources Information Center
US Department of Education, 2009
2009-01-01
This report details the U.S. Department of Education Office for Civil Rights' (OCRs') accomplishments in enforcing the civil rights laws under which OCR has been granted jurisdiction to address and remedy discrimination. These enforcement efforts include complaint investigation and resolution, compliance reviews and technical assistance, as well…
A Stirred Microchamber for Oxygen Consumption Rate Measurements With Pancreatic Islets
Papas, Klearchos K.; Pisania, Anna; Wu, Haiyan; Weir, Gordon C.; Colton, Clark K.
2010-01-01
Improvements in pancreatic islet transplantation for treatment of diabetes are hindered by the absence of meaningful islet quality assessment methods. Oxygen consumption rate (OCR) has previously been used to assess the quality of organs and primary tissue for transplantation. In this study, we describe and characterize a stirred microchamber for measuring OCR with small quantities of islets. The device has a titanium body with a chamber volume of about 200 µL and is magnetically stirred and water jacketed for temperature control. Oxygen partial pressure (pO2) is measured by fluorescence quenching with a fiber optic probe, and OCR is determined from the linear decrease of pO2 with time. We demonstrate that measurements can be made rapidly and with high precision. Measurements with βTC3 cells and islets show that OCR is directly proportional to the number of viable cells in mixtures of live and dead cells and correlate linearly with membrane integrity measurements made with cells that have been cultured for 24 h under various stressful conditions. PMID:17497731
Document image improvement for OCR as a classification problem
NASA Astrophysics Data System (ADS)
Summers, Kristen M.
2003-01-01
In support of the goal of automatically selecting methods of enhancing an image to improve the accuracy of OCR on that image, we consider the problem of determining whether to apply each of a set of methods as a supervised classification problem for machine learning. We characterize each image according to a combination of two sets of measures: a set that are intended to reflect the degree of particular types of noise present in documents in a single font of Roman or similar script and a more general set based on connected component statistics. We consider several potential methods of image improvement, each of which constitutes its own 2-class classification problem, according to whether transforming the image with this method improves the accuracy of OCR. In our experiments, the results varied for the different image transformation methods, but the system made the correct choice in 77% of the cases in which the decision affected the OCR score (in the range [0,1]) by at least .01, and it made the correct choice 64% of the time overall.
Reduction of ocular counter-rolling by adaptation to space
NASA Technical Reports Server (NTRS)
Dai, Mingjia; Mcgarvie, Leigh; Kozlovskaya, Inessa; Sirota, Mischa; Raphan, Theodore; Cohen, Bernard
1993-01-01
We studied the three-dimensional vestibulo-ocular reflex (VOR) of rhesus monkeys before and after the COSMOS Biosatellite 2229 Mission of 1992-1993. This included tests of ocular counter-rolling (OCR), the gain of the vestibulo-ocular reflex (VOR), and spatial orientation of velocity storage. A four-axis vestibular and oculomotor stimulator was transported to the Institute of Biomedical Problems in Moscow for the pre- and postflight ground-based testing. Twelve normal juvenile male rhesus monkey were implanted surgically with eye coils and tested 60-90 days before spaceflight. Two monkey (7906 and 6151), selected from the twelve as flight animals, flew from 12/29/92 to 1/10/93. Upon recovery, they were tested for 11 days postflight along with three control animals. Compensatory ocular torsion was produced in two ways: (1) Lateral head tilts evoked OCR through otolith-ocular reflexes. OCR was also measured dynamically during off-vertical axis rotation (OVAR). (2) Rotation about a naso-occipital axis that was either vertical of horizontal elicited torsional nystagmus through semicircular canal-ocular reflexes (roll VOR). OCR from the otoliths was substantially reduced (70 percent) for 11 days after reentry on both modes of testing. The gain of the roll VOR was also decreased, but less than OCR. These data demonstrate that there was a long-lasting depression of torsional or roll eye movements after adaptation to microgravity in these monkeys, especially those movements produced by the otolith organs.
Vogel, Markus; Kaisers, Wolfgang; Wassmuth, Ralf; Mayatepek, Ertan
2015-11-03
Clinical documentation has undergone a change due to the usage of electronic health records. The core element is to capture clinical findings and document therapy electronically. Health care personnel spend a significant portion of their time on the computer. Alternatives to self-typing, such as speech recognition, are currently believed to increase documentation efficiency and quality, as well as satisfaction of health professionals while accomplishing clinical documentation, but few studies in this area have been published to date. This study describes the effects of using a Web-based medical speech recognition system for clinical documentation in a university hospital on (1) documentation speed, (2) document length, and (3) physician satisfaction. Reports of 28 physicians were randomized to be created with (intervention) or without (control) the assistance of a Web-based system of medical automatic speech recognition (ASR) in the German language. The documentation was entered into a browser's text area and the time to complete the documentation including all necessary corrections, correction effort, number of characters, and mood of participant were stored in a database. The underlying time comprised text entering, text correction, and finalization of the documentation event. Participants self-assessed their moods on a scale of 1-3 (1=good, 2=moderate, 3=bad). Statistical analysis was done using permutation tests. The number of clinical reports eligible for further analysis stood at 1455. Out of 1455 reports, 718 (49.35%) were assisted by ASR and 737 (50.65%) were not assisted by ASR. Average documentation speed without ASR was 173 (SD 101) characters per minute, while it was 217 (SD 120) characters per minute using ASR. The overall increase in documentation speed through Web-based ASR assistance was 26% (P=.04). Participants documented an average of 356 (SD 388) characters per report when not assisted by ASR and 649 (SD 561) characters per report when assisted by ASR. Participants' average mood rating was 1.3 (SD 0.6) using ASR assistance compared to 1.6 (SD 0.7) without ASR assistance (P<.001). We conclude that medical documentation with the assistance of Web-based speech recognition leads to an increase in documentation speed, document length, and participant mood when compared to self-typing. Speech recognition is a meaningful and effective tool for the clinical documentation process.
Comparing the minimum spatial-frequency content for recognizing Chinese and alphabet characters
Wang, Hui; Legge, Gordon E.
2018-01-01
Visual blur is a common problem that causes difficulty in pattern recognition for normally sighted people under degraded viewing conditions (e.g., near the acuity limit, when defocused, or in fog) and also for people with impaired vision. For reliable identification, the spatial frequency content of an object needs to extend up to or exceed a minimum value in units of cycles per object, referred to as the critical spatial frequency. In this study, we investigated the critical spatial frequency for alphabet and Chinese characters, and examined the effect of pattern complexity. The stimuli were divided into seven categories based on their perimetric complexity, including the lowercase and uppercase alphabet letters, and five groups of Chinese characters. We found that the critical spatial frequency significantly increased with complexity, from 1.01 cycles per character for the simplest group to 2.00 cycles per character for the most complex group of Chinese characters. A second goal of the study was to test a space-bandwidth invariance hypothesis that would represent a tradeoff between the critical spatial frequency and the number of adjacent patterns that can be recognized at one time. We tested this hypothesis by comparing the critical spatial frequencies in cycles per character from the current study and visual-span sizes in number of characters (measured by Wang, He, & Legge, 2014) for sets of characters with different complexities. For the character size (1.2°) we used in the study, we found an invariant product of approximately 10 cycles, which may represent a capacity limitation on visual pattern recognition. PMID:29297056
Elevated nitrate alters the metabolic activity of embryonic zebrafish.
Conlin, Sarah M; Tudor, M Scarlett; Shim, Juyoung; Gosse, Julie A; Neilson, Andrew; Hamlin, Heather J
2018-04-01
Nitrate accumulation in aquatic reservoirs from agricultural pollution has often been overlooked as a water quality hazard, yet a growing body of literature suggests negative effects on human and wildlife health following nitrate exposure. This research seeks to understand differences in oxygen consumption rates between different routes of laboratory nitrate exposure, whether via immersion or injection, in zebrafish (Danio rerio) embryos. Embryos were exposed within 1 h post fertilization (hpf) to 0, 10, and 100 mg/L NO 3 -N with sodium nitrate, or to counter ion control (CIC) treatments using sodium chloride. Embryos in the immersion treatments received an injection of 4 nL of appropriate treatment solution into the perivitelline space. At 24 hpf, Oxygen Consumption Rates (OCR) were measured and recorded in vivo using the Agilent Technologies XF e 96 Extracellular Flux Analyzer and Spheroid Microplate. Immersion exposures did not induce significant changes in OCR, yet nitrate induced significant changes when injected through the embryo chorion. Injection of 10 and 100 mg/L NO 3 -N down-regulated OCR compared to the control treatment group. Injection of the 100 mg/L CIC also significantly down-regulated OCR compared to the control treatment group. Interestingly, the 100 mg/L NO 3 -N treatment further down-regulated OCR compared to the 100 mg/L CIC treatment, suggesting the potential for additive effects between the counter ion and the ion of interest. These data support that elevated nitrate exposure can alter normal metabolic activity by changing OCR in 24 hpf embryos. These results highlight the need for regularly examining the counter ion of laboratory nitrate compounds while conducting research with developing zebrafish, and justify examining different routes of laboratory nitrate exposure, as the chorion may act as an effective barrier to nitrate penetration in zebrafish, which may lead to conservative estimates of significant effects in other species for which nitrate more readily penetrates the chorion. Copyright © 2017 Elsevier Ltd. All rights reserved.
New horizons for multiple sclerosis therapeutics: milestones in the development of ocrelizumab.
Frau, Jessica; Coghe, Giancarlo; Lorefice, Lorena; Fenu, Giuseppe; Cocco, Eleonora
2018-01-01
Multiple Sclerosis (MS) is an inflammatory and neurodegenerative disease of the central nervous system, and both T and B cells are involved in its pathogenesis. The vast majority of disease-modifying drugs used for MS act on the inflammatory component of the disease and are approved for use in relapsing-remitting (RR) patients. Ocrelizumab (OCR) is the only MS drug that has been approved by the US Food and Drug Administration (FDA) not only for patients with RRMS but also for patients with primary progressive (PP) MS. OCR is a humanized anti-CD20 monoclonal antibody that can deplete the targeted B cells through antibody-dependent cellular cytotoxicity. Treatment involves administration by intravenous infusion every 6 months. OCR can cause long-lasting B-cell depletion and change the pool of reconstituted B cells. Phase III clinical trials have confirmed the results of previous Phase II studies. In particular, OPERA I and II trials, which were performed in patients with RRMS, showed a reduction in the annualized relapse rate, the risk of disability progression, and the number of new/enlarging T2 lesions and enhancing lesions measured using brain magnetic resonance. The ORATORIO trial, performed in PP subjects, showed that OCR can reduce disability progression, improve performance on the timed 25-foot walk, and decrease the total volume of T2 lesions and the mean number of new or enlarging T2 lesions. The most frequent adverse events were the infusion-related reactions and infections. Infections were mostly nasopharyngitis, as well as upper respiratory and urinary tract infections. OCR gives no indication for severe or opportunistic infections. There is not a clear increased risk of malignancies. Nevertheless, it could not be excluded. Real-life registries will provide more information about the long-term safety, the risk of exposure during pregnancy, and the risk of rare adverse events. In this review, we analyze the evidence regarding the efficacy and the safety of OCR.
Endorsing the Practical Endorsement? OCR's Approach to Practical Assessment in Science A-Levels
ERIC Educational Resources Information Center
Evans, Steve; Wade, Neil
2015-01-01
This article summarises the practical requirements for new science A-levels in biology, chemistry and physics for first teaching from September 2015. It discusses the background to how the new approach was reached and how OCR has seen this taking shape in our assessment models. The opportunities presented by this new approach to practical…
Office for Civil Rights. Annual Report to Congress, Fiscal Year 1991.
ERIC Educational Resources Information Center
Office for Civil Rights (ED), Washington, DC.
The Office for Civil Rights (OCR) is a law enforcement agency primarily responsible for ensuring that recipients of federal assistance do not discriminate against students, faculty, or other individuals on the basis of race, color, national origin, sex, handicap, or age. OCR is responsible for enforcing the following Federal civil rights laws: the…
Delivering Justice: Report to the President and Secretary of Education, Fiscal Year 2015
ERIC Educational Resources Information Center
Office for Civil Rights, US Department of Education, 2015
2015-01-01
For decades, the Office for Civil Rights (OCR) has enforced civil rights in our nation's preschool through 12th grade (P-12) and post-secondary schools, which now serve almost 80 million students annually. In Fiscal Year 2015, the OCR promoted excellence in education through conducting investigations and monitoring schools under resolution…
Ensuring Equal Access to High-Quality Education. Revised.
ERIC Educational Resources Information Center
Office for Civil Rights (ED), Washington, DC.
This brochure describes the activities of the Office for Civil Rights (OCR) in the U.S. Department of Education. The OCR is a law-enforcement agency charged with upholding the federal civil-rights laws that prohibit discrimination on the basis of race, color, national origin, sex, disability, and age in programs and activities that receive federal…
Effect of word familiarity on visually evoked magnetic fields.
Harada, N; Iwaki, S; Nakagawa, S; Yamaguchi, M; Tonoike, M
2004-11-30
This study investigated the effect of word familiarity of visual stimuli on the word recognizing function of the human brain. Word familiarity is an index of the relative ease of word perception, and is characterized by facilitation and accuracy on word recognition. We studied the effect of word familiarity, using "Hiragana" (phonetic characters in Japanese orthography) characters as visual stimuli, on the elicitation of visually evoked magnetic fields with a word-naming task. The words were selected from a database of lexical properties of Japanese. The four "Hiragana" characters used were grouped and presented in 4 classes of degree of familiarity. The three components were observed in averaged waveforms of the root mean square (RMS) value on latencies at about 100 ms, 150 ms and 220 ms. The RMS value of the 220 ms component showed a significant positive correlation (F=(3/36); 5.501; p=0.035) with the value of familiarity. ECDs of the 220 ms component were observed in the intraparietal sulcus (IPS). Increments in the RMS value of the 220 ms component, which might reflect ideographical word recognition, retrieving "as a whole" were enhanced with increments of the value of familiarity. The interaction of characters, which increased with the value of familiarity, might function "as a large symbol"; and enhance a "pop-out" function with an escaping character inhibiting other characters and enhancing the segmentation of the character (as a figure) from the ground.
Liu, Tianyin; Yeh, Su-Ling
2018-01-01
The left-side bias (LSB) effect observed in face and expert Chinese character perception is suggested to be an expertise marker for visual object recognition. However, in character perception this effect is limited to characters printed in a familiar font (font-sensitive LSB effect). Here we investigated whether the LSB and font-sensitive LSB effects depend on participants’ familiarity with global structure or local component information of the stimuli through examining their transfer effects across simplified and traditional Chinese scripts: the two Chinese scripts share similar overall structures but differ in the visual complexity of local components in general. We found that LSB in expert Chinese character processing could be transferred to the Chinese script that the readers are unfamiliar with. In contrast, the font-sensitive LSB effect did not transfer, and was limited to characters with the visual complexity the readers were most familiar with. These effects suggest that the LSB effect may be generalized to another visual category with similar overall structures; in contrast, effects of within-category variations such as fonts may depend on familiarity with local component information of the stimuli, and thus may be limited to the exemplars of the category that experts are typically exposed to. PMID:29608570
Signature Verification Based on Handwritten Text Recognition
NASA Astrophysics Data System (ADS)
Viriri, Serestina; Tapamo, Jules-R.
Signatures continue to be an important biometric trait because it remains widely used primarily for authenticating the identity of human beings. This paper presents an efficient text-based directional signature recognition algorithm which verifies signatures, even when they are composed of special unconstrained cursive characters which are superimposed and embellished. This algorithm extends the character-based signature verification technique. The experiments carried out on the GPDS signature database and an additional database created from signatures captured using the ePadInk tablet, show that the approach is effective and efficient, with a positive verification rate of 94.95%.
ERIC Educational Resources Information Center
Matsumoto, Kazumi
2013-01-01
This study investigated whether learners of Japanese with different first language (L1) writing systems use different recognition strategies and whether second language (L2) exposure affects L2 kanji recognition. The study used a computerized lexical judgment task with 3 types of kanji characters to investigate these questions: (a)…
Koçer, O; Wachter, M; Zellweger, M; Piazzalonga, S; Hoffmann, A
2011-07-30
Depression is an important independent prognostic variable in cardiac patients. The prevalence and predictors of depressive symptoms up to nine years after cardiac rehabilitation were studied. Follow-up questionnaires were sent to 2199 patients who had completed a 12-week exercise-based outpatient cardiac rehabilitation (OCR) programme between June 1999 and March 2006. Medical outcome, general wellbeing, and depressive symptoms were assessed, the latter by using two screening questions according to Arrol. Patients with incomplete data due to language problems, lack of compliance and non-response were excluded. Complete data for analysis was available for 710 patients. The median follow up period was 46 months (Interquartile range (IQR) 22-71, min. 6 months). At follow-up, 132 patients (19%) indicated low wellbeing, whereas 81 (11%) were having depressive symptoms. Multivariate analyses revealed impaired quality of life (p <0.001), diabetes (p = 0.013) and low exercise capacity after OCR (p = 0.003) to be independent predictors of low wellbeing at follow-up. Persistent smoking (p = 0.045) as well as negative mood (p = 0.022) at the end of OCR were independent predictors of depressive symptoms at follow-up. In a selected patient population a mean of four years after OCR, persistent smoking, diabetes, low exercise capacity and impaired quality of life at the end of OCR were independent long term predictors of low wellbeing and depressive symptoms, rather than specific cardiac variables. This highlights the need for close cooperation between cardiovascular and psychological specialists in cardiac rehabilitation.
Reading Machines for Blind People.
ERIC Educational Resources Information Center
Fender, Derek H.
1983-01-01
Ten stages of developing reading machines for blind people are analyzed: handling of text material; optics; electro-optics; pattern recognition; character recognition; storage; speech synthesizers; browsing and place finding; computer indexing; and other sources of input. Cost considerations of the final product are emphasized. (CL)
The Impact of the Native American Languages Act on Public School Curriculum: A Different View.
ERIC Educational Resources Information Center
Littlejohn, Jim
2000-01-01
Responds to Scott Ferrin's argument (EJ 583 598) and, as the former policy director for the U.S. Office for Civil Rights (OCR), decries both Ferrin and OCR for their relentless and, in his view, unwarranted promotion of bilingual education. Contends that no civil-rights laws, including the Native American Languages Act (NALA), support or allow…
ERIC Educational Resources Information Center
Mansfield, Wendy; Farris, Elizabeth
This report provides results of a Fast Response Survey System (FRSS) study conducted by the National Center for Education Statistics for the Office for Civil Rights (OCR). The OCR wanted input for their decision-making process on possible modifications to their biennial survey of a national sample of public school districts (PSDs). The survey, the…
ERIC Educational Resources Information Center
Office for Civil Rights (ED), Washington, DC.
This annual report details the responsibilities of the Office of Civil Rights (OCR). It outlines the main purpose of the OCR--the agency within the United States Department of Education that regulates and enforces civil-rights laws against discrimination in education programs and activities receiving federal financial assistance. The focus of the…
ERIC Educational Resources Information Center
Larson, Meredith A.; And Others
This two part study was designed to determine whether working relationships can be expanded between the Office for Civil Rights (OCR) and various State agencies with similar missions. First, a survey of State laws and enforcement mechanisms identifies those States with current laws or policies similar to those that OCR administers. Results…
Decreased otolith-mediated vestibular response in 25 astronauts induced by long-duration spaceflight
Hallgren, Emma; Kornilova, Ludmila; Fransen, Erik; Glukhikh, Dmitrii; Moore, Steven T.; Clément, Gilles; Van Ombergen, Angelique; MacDougall, Hamish; Naumov, Ivan
2016-01-01
The information coming from the vestibular otolith organs is important for the brain when reflexively making appropriate visual and spinal corrections to maintain balance. Symptoms related to failed balance control and navigation are commonly observed in astronauts returning from space. To investigate the effect of microgravity exposure on the otoliths, we studied the otolith-mediated responses elicited by centrifugation in a group of 25 astronauts before and after 6 mo of spaceflight. Ocular counterrolling (OCR) is an otolith-driven reflex that is sensitive to head tilt with regard to gravity and tilts of the gravito-inertial acceleration vector during centrifugation. When comparing pre- and postflight OCR, we found a statistically significant decrease of the OCR response upon return. Nine days after return, the OCR was back at preflight level, indicating a full recovery. Our large study sample allows for more general physiological conclusions about the effect of prolonged microgravity on the otolith system. A deconditioned otolith system is thought to be the cause of several of the negative effects seen in returning astronauts, such as spatial disorientation and orthostatic intolerance. This knowledge should be taken into account for future long-term space missions. PMID:27009158
Effects on Learning Logographic Character Formation in Computer-Assisted Handwriting Instruction
ERIC Educational Resources Information Center
Tsai, Chen-hui; Kuo, Chin-Hwa; Horng, Wen-Bing; Chen, Chun-Wen
2012-01-01
This paper reports on a study that investigates how different learning methods might affect the learning process of character handwriting among beginning college learners of Chinese, as measured by tests of recognition, approximate production, precise production, and awareness of conventional stroke sequence. Two methodologies were examined during…
The Inversion Effect for Chinese Characters Is Modulated by Radical Organization
ERIC Educational Resources Information Center
Luo, Canhuang; Chen, Wei; Zhang, Ye
2017-01-01
In studies of visual object recognition, strong inversion effects accompany the acquisition of expertise and imply the involvement of configural processing. Chinese literacy results in sensitivity to the orthography of Chinese characters. While there is some evidence that this orthographic sensitivity results in an inversion effect, and thus…
Training a whole-book LSTM-based recognizer with an optimal training set
NASA Astrophysics Data System (ADS)
Soheili, Mohammad Reza; Yousefi, Mohammad Reza; Kabir, Ehsanollah; Stricker, Didier
2018-04-01
Despite the recent progress in OCR technologies, whole-book recognition, is still a challenging task, in particular in case of old and historical books, that the unknown font faces or low quality of paper and print contributes to the challenge. Therefore, pre-trained recognizers and generic methods do not usually perform up to required standards, and usually the performance degrades for larger scale recognition tasks, such as of a book. Such reportedly low error-rate methods turn out to require a great deal of manual correction. Generally, such methodologies do not make effective use of concepts such redundancy in whole-book recognition. In this work, we propose to train Long Short Term Memory (LSTM) networks on a minimal training set obtained from the book to be recognized. We show that clustering all the sub-words in the book, and using the sub-word cluster centers as the training set for the LSTM network, we can train models that outperform any identical network that is trained with randomly selected pages of the book. In our experiments, we also show that although the sub-word cluster centers are equivalent to about 8 pages of text for a 101- page book, a LSTM network trained on such a set performs competitively compared to an identical network that is trained on a set of 60 randomly selected pages of the book.
ERIC Educational Resources Information Center
Raskind, Marshall
1993-01-01
This article describes assistive technologies for persons with learning disabilities, including word processing, spell checking, proofreading programs, outlining/"brainstorming" programs, abbreviation expanders, speech recognition, speech synthesis/screen review, optical character recognition systems, personal data managers, free-form databases,…
The EPA eXcats is an enterprise-level data tracking application that provides management complaint tracking information for the EPA's Office of Civil Rights (OCR) External Compliance Program. EPA's OCR is responsible for enforcing several federal civil rights laws that together prohibit discrimination on the bases of race, color, national origin (including limited English proficiency), disability, sex and age in programs or activities that receive federal financial assistance from the EPA.
Engineering Irisin for Understanding Its Benefits to Obesity
2018-03-01
measurement of oxygen consumption rate (OCR) and extracellular acidification rates (ECAR). Following basal respiration, the mitochondrial effectors...mitochondrial respiration, respectively. Effects of irisin on cellular oxygen consumption rate (OCR; A and B) and ECAR (extracellular acidification rates; C...irisin alanine variants for 60 min at room temperature . The cells were then washed and resuspended in PBS/0.5% BSA. Fifty thousand events per sample
ERIC Educational Resources Information Center
Wilson, Frances; Wade, Neil; Evans, Steve
2016-01-01
As part of a wider qualification reform at upper secondary level in England, the assessment of practical work is changing, following concerns that the legacy model of assessment was having a negative impact on teaching and learning. The OCR examination board is running a longitudinal survey of teachers' views on practical work, focusing on the…
Aspects of quality insurance in digitizing historical climate data in Germany
NASA Astrophysics Data System (ADS)
Mächel, H.; Behrends, J.; Kapala, A.
2010-09-01
This contribution presents some of the problems and offers solutions regarding the digitization of historical meteorological data, and explains the need for verification and quality control. For the assessment of changes in climate extremes, long-term and complete observational records with a high temporal resolution are needed. However, in most countries, including Germany, such climate data are rare. Therefore, in 2005, the German Weather Service launched a project to inventory and digitize historical daily climatic records in cooperation with the Meteorological Institute of the University of Bonn. Experience with Optical Character Recognition (OCR) show that it is only of very limited use, as even printed tables (e.g. yearbooks) are not sufficiently recognized (10-20% error). In hand-written records, the recognition rate is about 50%. By comparing daily and monthly values, it is possible to auto-detect errors, but they can not be automatically corrected, since there is often more than one error per month. These erroneous data must then be controlled manually on an individual basis, which is significantly more error-prone than direct manual input. Therefore, both precipitation and climate station data are digitized manually. The time required to digitize one year of precipitation data (including the recording of daily precipitation amount and type, snow amount and type, and weather events such as thunder storms, fog, etc.) is equivalent to about five hours for one year of data. This involves manually typing, reformatting and quality control of the digitized data, as well as creating a digital photograph. For climate stations with three observations per day, the working time is 30-50 hours for one year of data, depending on the number of parameters and the condition of the documents. Several other problems occur when creating the digital records from historical observational data, some of which are listed below. Older records often used varying units and different conventions. For example, a value of 100 was added to the observed temperatures to avoid negative values. Furthermore, because standardization of the observations was very low when measurements began up to 200 years ago, the data often reflect a greater part of non-climatic influences. Varying daily observation times make it difficult to calculate a representative daily value. Even unconventional completed tables cost labor and requires experienced and trained staff. Data homogenization as well as both manual and automatic quality control may address some of these problems.
Biswas, Mithun; Islam, Rafiqul; Shom, Gautam Kumar; Shopon, Md; Mohammed, Nabeel; Momen, Sifat; Abedin, Anowarul
2017-06-01
BanglaLekha-Isolated, a Bangla handwritten isolated character dataset is presented in this article. This dataset contains 84 different characters comprising of 50 Bangla basic characters, 10 Bangla numerals and 24 selected compound characters. 2000 handwriting samples for each of the 84 characters were collected, digitized and pre-processed. After discarding mistakes and scribbles, 1,66,105 handwritten character images were included in the final dataset. The dataset also includes labels indicating the age and the gender of the subjects from whom the samples were collected. This dataset could be used not only for optical handwriting recognition research but also to explore the influence of gender and age on handwriting. The dataset is publicly available at https://data.mendeley.com/datasets/hf6sf8zrkc/2.
Integrative Lifecourse and Genetic Analysis of Military Working Dogs
2012-10-01
Intelligent Character Recognition) and HWR ( Handwriting Recognition). A number of various software packages were evaluated and we have settled on a...third-party software is able to recognize check-boxes and columns and do a reasonable job with handwriting – which is does. This workflow will
76 FR 39757 - Filing Procedures
Federal Register 2010, 2011, 2012, 2013, 2014
2011-07-06
... an optical character recognition process, such a document may contain recognition errors. CAUTION... network speed e-filing of these documents may be difficult. Pursuant to section II(C) above, the Secretary... optical scan format or a typed ``electronic signature,'' e.g., ``/s/Jane Doe.'' (3) In the case of a...
Tracker: Image-Processing and Object-Tracking System Developed
NASA Technical Reports Server (NTRS)
Klimek, Robert B.; Wright, Theodore W.
1999-01-01
Tracker is an object-tracking and image-processing program designed and developed at the NASA Lewis Research Center to help with the analysis of images generated by microgravity combustion and fluid physics experiments. Experiments are often recorded on film or videotape for analysis later. Tracker automates the process of examining each frame of the recorded experiment, performing image-processing operations to bring out the desired detail, and recording the positions of the objects of interest. It can load sequences of images from disk files or acquire images (via a frame grabber) from film transports, videotape, laser disks, or a live camera. Tracker controls the image source to automatically advance to the next frame. It can employ a large array of image-processing operations to enhance the detail of the acquired images and can analyze an arbitrarily large number of objects simultaneously. Several different tracking algorithms are available, including conventional threshold and correlation-based techniques, and more esoteric procedures such as "snake" tracking and automated recognition of character data in the image. The Tracker software was written to be operated by researchers, thus every attempt was made to make the software as user friendly and self-explanatory as possible. Tracker is used by most of the microgravity combustion and fluid physics experiments performed by Lewis, and by visiting researchers. This includes experiments performed on the space shuttles, Mir, sounding rockets, zero-g research airplanes, drop towers, and ground-based laboratories. This software automates the analysis of the flame or liquid s physical parameters such as position, velocity, acceleration, size, shape, intensity characteristics, color, and centroid, as well as a number of other measurements. It can perform these operations on multiple objects simultaneously. Another key feature of Tracker is that it performs optical character recognition (OCR). This feature is useful in extracting numerical instrumentation data that are embedded in images. All the results are saved in files for further data reduction and graphing. There are currently three Tracking Systems (workstations) operating near the laboratories and offices of Lewis Microgravity Science Division researchers. These systems are used independently by students, scientists, and university-based principal investigators. The researchers bring their tapes or films to the workstation and perform the tracking analysis. The resultant data files generated by the tracking process can then be analyzed on the spot, although most of the time researchers prefer to transfer them via the network to their offices for further analysis or plotting. In addition, many researchers have installed Tracker on computers in their office for desktop analysis of digital image sequences, which can be digitized by the Tracking System or some other means. Tracker has not only provided a capability to efficiently and automatically analyze large volumes of data, saving many hours of tedious work, but has also provided new capabilities to extract valuable information and phenomena that was heretofore undetected and unexploited.
Fuzzy Logic Module of Convolutional Neural Network for Handwritten Digits Recognition
NASA Astrophysics Data System (ADS)
Popko, E. A.; Weinstein, I. A.
2016-08-01
Optical character recognition is one of the important issues in the field of pattern recognition. This paper presents a method for recognizing handwritten digits based on the modeling of convolutional neural network. The integrated fuzzy logic module based on a structural approach was developed. Used system architecture adjusted the output of the neural network to improve quality of symbol identification. It was shown that proposed algorithm was flexible and high recognition rate of 99.23% was achieved.
Hawley, Alana; Mercuri, Mathew; Hogg, Kerstin; Hanel, Erich
2017-03-01
The growing popularity of obstacle course runs (OCRs) has led to significant concerns regarding their safety. The influx of injuries and illnesses in rural areas where OCRs are often held can impose a large burden on emergency medical services (EMS) and local EDs. Literature concerning the safety of these events is minimal and mostly consists of media reports. We sought to characterise the injury and illness profile of OCRs and the level of medical care required. This study analysed OCR events occurring in eight locations across Canada from May to August 2015 (total 45 285 participants). Data were extracted from event medical charts of patients presenting to the onsite medical team, including injury or illness type, onsite treatment and disposition. There were 557 race participants treated at eight OCR events (1.2% of all participants). There were 609 medical complaints in total. Three quarters of injuries were musculoskeletal in nature. Eighty-nine per cent returned to the event with no need for further medical care. The majority of treatments were completed with first aid and basic medical equipment. Eleven patients (2% of patients) required transfer to hospital by EMS for presentations including fracture, dislocation, head injury, chest pain, fall from height, and abdominal pain. We found that 1.2% of race participants presented to onsite medical services. The majority of complaints were minor and musculoskeletal in nature. Only 2% of those treated were transferred to hospital through EMS. This is consistent with other types of mass gathering events. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://www.bmj.com/company/products-services/rights-and-licensing/.
Huang, Shih-Hao; Hsu, Yu-Hsuan; Wu, Chih-Wei; Wu, Chang-Jer
2012-01-01
A digital light modulation system that utilizes a modified commercial digital micromirror device (DMD) projector, which is equipped with a UV light-emitting diode as a light modulation source, has been developed to spatially direct excited light toward a microwell array device to detect the oxygen consumption rate (OCR) of single cells via phase-based phosphorescence lifetime detection. The microwell array device is composed of a combination of two components: an array of glass microwells containing Pt(II) octaethylporphine (PtOEP) as the oxygen-sensitive luminescent layer and a microfluidic module with pneumatically actuated glass lids set above the microwells to controllably seal the microwells of interest. By controlling the illumination pattern on the DMD, the modulated excitation light can be spatially projected to only excite the sealed microwell for cellular OCR measurements. The OCR of baby hamster kidney-21 fibroblast cells cultivated on the PtOEP layer within a sealed microwell has been successfully measured at 104 ± 2.96 amol s−1 cell−1. Repeatable and consistent measurements indicate that the oxygen measurements did not adversely affect the physiological state of the measured cells. The OCR of the cells exhibited a good linear relationship with the diameter of the microwells, ranging from 400 to 1000 μm and containing approximately 480 to 1200 cells within a microwell. In addition, the OCR variation of single cells in situ infected by Dengue virus with a different multiplicity of infection was also successfully measured in real-time. This proposed platform provides the potential for a wide range of biological applications in cell-based biosensing, toxicology, and drug discovery. PMID:24348889
Effect of temperature on the standard metabolic rates of juvenile and adult Exopalaemon carinicauda
NASA Astrophysics Data System (ADS)
Zhang, Chengsong; Li, Fuhua; Xiang, Jianhai
2015-03-01
Ridgetail white prawn ( Exopalaemon carinicauda) are of significant economic importance in China where they are widely cultured. However, there is little information on the basic biology of this species. We evaluated the effect of temperature (16, 19, 22, 25, 28, 31, and 34°C) on the standard metabolic rates (SMRs) of juvenile and adult E. carinicauda in the laboratory under static conditions. The oxygen consumption rate (OCR), ammonia-N excretion rate (AER), and atomic ratio of oxygen consumed to nitrogen consumed (O:N ratio) of juvenile and adult E. carinicauda were significantly influenced by temperature ( P < 0.05). Both the OCR and AER of juveniles increased significantly with increasing temperature from 16 to 34°C, but the maximum OCR for adults was at 31°C. Juvenile shrimp exhibited a higher OCR than the adults from 19 to 34°C. There was no significant difference between the AERs of the two life-stages from 16 to 31°C ( P >0.05). The O:N ratio in juveniles was significantly higher than that in the adults over the entire temperature range ( P <0.05). The temperature coefficient ( Q 10) of OCR and AER ranged from 5.03 to 0.86 and 6.30 to 0.85 for the adults, respectively, and from 6.09-1.03 and 3.66-1.80 for the juveniles, respectively. The optimal temperature range for growth of the juvenile and adult shrimp was from 28 to 31°C, based on Q 10 and SMR values. Results from the present study may be used to guide pond culture production of E. carinicauda.
Sosa, Ivan; Estrada, Amara H; Winter, Brandy D; Erger, Kirsten E; Conlon, Thomas J
2016-02-01
To compare mitochondrial oxygen consumption rate (OCR) of fibroblasts from Doberman Pinschers with and without dilated cardiomyopathy (DCM) and mutation of the gene for pyruvate dehydrogenase kinase isozyme 4 (PDK4) and to evaluate in vitro whether treatment with adeno-associated virus (AAV) vector (i.e., gene therapy) would alter metabolic efficiency. 10 Doberman Pinschers screened for DCM and PDK4 mutation. PROCEDURES Fibroblasts were harvested from skin biopsy specimens obtained from Doberman Pinschers, and dogs were classified as without DCM or PDK4 mutation (n = 3) or with occult DCM and heterozygous (4) or homozygous (3) for PDK4 mutation. Fibroblasts were or were not treated with tyrosine mutant AAV type 2 vector containing PDK4 at multiplicities of infection of 1,000. Mitochondrial OCR was measured to evaluate mitochondrial metabolism. The OCR was compared among dog groups and between untreated and treated fibroblasts within groups. Mean ± SD basal OCR of fibroblasts from heterozygous (74 ± 8 pmol of O2/min) and homozygous (58 ± 12 pmol of O2/min) dogs was significantly lower than that for dogs without PDK4 mutation (115 ± 9 pmol of O2/min). After AAV transduction, OCR did not increase significantly in any group (mutation-free group, 121 ± 26 pmol of O2/min; heterozygous group, 88 ± 6 pmol of O2/min; homozygous group, 59 ± 3 pmol of O2/min). Mitochondrial function was altered in skin fibroblasts of Doberman Pinschers with DCM and PDK4 mutation. Change in mitochondrial function after in vitro gene therapy at the multiplicities of infection used in this study was not significant.
Su, Alvin W; Chen, Yunchan; Wailes, Dustin H; Wong, Van W; Cai, Shengqiang; Chen, Albert C; Bugbee, William D; Sah, Robert L
2018-01-01
An osteochondral graft (OCG) is an effective treatment for articular cartilage and osteochondral defects. Impact of an OCG during insertion into the osteochondral recipient site (OCR) can cause chondrocyte death and matrix damage. The aim of the present study was to analyze the effects of graft-host interference fit and a modified OCG geometry on OCG insertion biomechanics and cartilage damage. The effects of interference fit (radius of OCG - radius of OCR), loose (0.00 mm), moderate (0.05 mm), tight (0.10 mm), and of a tight fit with OCG geometry modification (central region of decreased radius), were analyzed for OCG cylinders and OCR blocks from adult bovine knee joints with an instrumented drop tower apparatus. An increasingly tight (OCG - OCR) interference fit led to increased taps for insertion, peak axial force, graft cartilage axial compression, cumulative and total energy delivery to cartilage, lower time of peak axial force, lesser graft advancement during each tap, higher total crack length in the cartilage surface, and lower chondrocyte viability. The modified OCG, with reduction of diameter in the central area, altered the biomechanical insertion variables and biological consequences to be similar to those of the moderate interference fit scenario. Micro-computed tomography confirmed structural interference between the OCR bone and both the proximal and distal bone segments of the OCGs, with the central regions being slightly separated for the modified OCGs. These results clarify OCG insertion biomechanics and mechanobiology, and introduce a simple modification of OCGs that facilitates insertion with reduced energy while maintaining a structural interference fit. © 2017 Orthopaedic Research Society. Published by Wiley Periodicals, Inc. J Orthop Res 36:377-386, 2018. © 2017 Orthopaedic Research Society. Published by Wiley Periodicals, Inc.
Why Learning to Write Chinese Is a Waste of Time: A Modest Proposal
ERIC Educational Resources Information Center
Allen, Joseph R.
2008-01-01
This article argues that for students of Chinese and Japanese, learning to write Chinese characters ("hanzi/kanji") by hand from memory is an inefficient use of resources. Rather, beginning students should focus on character/word recognition (reading) and electronic writing. Although electronic technologies have diminished the usefulness of…
The Role of Orthographic Neighborhood Size Effects in Chinese Word Recognition
ERIC Educational Resources Information Center
Li, Meng-Feng; Lin, Wei-Chun; Chou, Tai-Li; Yang, Fu-Ling; Wu, Jei-Tun
2015-01-01
Previous studies about the orthographic neighborhood size (NS) in Chinese have overlooked the morphological processing, and the co-variation between the character frequency and the the NS. The present study manipulated the word frequency and the NS simultaneously, with the leading character frequency controlled, to explore their influences on word…
Chinese Characters Elicit Face-Like N170 Inversion Effects
ERIC Educational Resources Information Center
Wang, Man-Ying; Kuo, Bo-Cheng; Cheng, Shih-Kuen
2011-01-01
Recognition of both faces and Chinese characters is commonly believed to rely on configural information. While faces typically exhibit behavioral and N170 inversion effects that differ from non-face stimuli (Rossion, Joyce, Cottrell, & Tarr, 2003), the current study examined whether a similar reliance on configural processing may result in similar…
Hu, Zhishan; Zhang, Juan; Couto, Tania Alexandra; Xu, Shiyang; Luan, Ping; Yuan, Zhen
2018-06-22
In this study, functional near-infrared spectroscopy (fNIRS) was used to examine the brain activation and connectivity in occipitotemporal cortex during Chinese character recognition (CCR). Eighteen healthy participants were recruited to perform a well-designed task with three categories of stimuli (real characters, pseudo characters, and checkerboards). By inspecting the brain activation difference and its relationship with behavioral data, the left laterality during CCR was clearly identified in the Brodmann area (BA) 18 and 19. In addition, our novel findings also demonstrated that the bilateral superior temporal gyrus (STG), bilateral BA 19, and left fusiform gyrus were also involved in high-level lexical information processing such as semantic and phonological ones. Meanwhile, by examining functional brain networks, we discovered that the right BA 19 exhibited enhanced brain connectivity. In particular, the connectivity in the right fusiform gyrus, right BA 19, and left STG showed significant correlation with the performance of CCR. Consequently, the combination of fNIRS technique with functional network analysis paves a new avenue for improved understanding of the cognitive mechanism underlying CCR.
Producing good font attribute determination using error-prone information
NASA Astrophysics Data System (ADS)
Cooperman, Robert
1997-04-01
A method to provide estimates of font attributes in an OCR system, using detectors of individual attributes that are error-prone. For an OCR system to preserve the appearance of a scanned document, it needs accurate detection of font attributes. However, OCR environments have noise and other sources of errors, tending to make font attribute detection unreliable. Certain assumptions about font use can greatly enhance accuracy. Attributes such as boldness and italics are more likely to change between neighboring words, while attributes such as serifness are less likely to change within the same paragraph. Furthermore, the document as a whole, tends to have a limited number of sets of font attributes. These assumptions allow a better use of context than the raw data, or what would be achieved by simpler methods that would oversmooth the data.
Assaying Mitochondrial Respiration as an Indicator of Cellular Metabolism and Fitness.
Smolina, Natalia; Bruton, Joseph; Kostareva, Anna; Sejersen, Thomas
2017-01-01
Mitochondrial respiration is the most important generator of cellular energy under most circumstances. It is a process of energy conversion of substrates into ATP. The Seahorse equipment allows measuring oxygen consumption rate (OCR) in living cells and estimates key parameters of mitochondrial respiration in real-time mode. Through use of mitochondrial inhibitors, four key mitochondrial respiration parameters can be measured: basal, ATP production-linked, maximal, and proton leak-linked OCR. This approach requires application of mitochondrial inhibitors-oligomycin to block ATP synthase, FCCP-to make the inner mitochondrial membrane permeable for protons and allow maximum electron flux through the electron transport chain, and rotenone and antimycin A-to inhibit complexes I and III, respectively. This chapter describes the protocol of OCR assessment in the culture of primary myotubes obtained upon satellite cell fusion.
ERIC Educational Resources Information Center
Guan, Connie Qun; Liu, Ying; Chan, Derek Ho Leung; Ye, Feifei; Perfetti, Charles A.
2011-01-01
Learning to write words may strengthen orthographic representations and thus support word-specific recognition processes. This hypothesis applies especially to Chinese because its writing system encourages character-specific recognition that depends on accurate representation of orthographic form. We report 2 studies that test this hypothesis in…
ERIC Educational Resources Information Center
Defense Documentation Center, Alexandria, VA.
This unclassified-unlimited bibliography contains 183 references, with abstracts, dealing specifically with optical or graphic information processing. Citations are grouped under three headings: display devices and theory, character recognition, and pattern recognition. Within each group, they are arranged in accession number (AD-number) sequence.…
Sadeghi, Zahra; Testolin, Alberto
2017-08-01
In humans, efficient recognition of written symbols is thought to rely on a hierarchical processing system, where simple features are progressively combined into more abstract, high-level representations. Here, we present a computational model of Persian character recognition based on deep belief networks, where increasingly more complex visual features emerge in a completely unsupervised manner by fitting a hierarchical generative model to the sensory data. Crucially, high-level internal representations emerging from unsupervised deep learning can be easily read out by a linear classifier, achieving state-of-the-art recognition accuracy. Furthermore, we tested the hypothesis that handwritten digits and letters share many common visual features: A generative model that captures the statistical structure of the letters distribution should therefore also support the recognition of written digits. To this aim, deep networks trained on Persian letters were used to build high-level representations of Persian digits, which were indeed read out with high accuracy. Our simulations show that complex visual features, such as those mediating the identification of Persian symbols, can emerge from unsupervised learning in multilayered neural networks and can support knowledge transfer across related domains.
BACS: The Brussels Artificial Character Sets for studies in cognitive psychology and neuroscience.
Vidal, Camille; Content, Alain; Chetail, Fabienne
2017-12-01
Written symbols such as letters have been used extensively in cognitive psychology, whether to understand their contributions to written word recognition or to examine the processes involved in other mental functions. Sometimes, however, researchers want to manipulate letters while removing their associated characteristics. A powerful solution to do so is to use new characters, devised to be highly similar to letters, but without the associated sound or name. Given the growing use of artificial characters in experimental paradigms, the aim of the present study was to make available the Brussels Artificial Character Sets (BACS): two full, strictly controlled, and portable sets of artificial characters for a broad range of experimental situations.
Adversity, emotion recognition, and empathic concern in high-risk youth.
Quas, Jodi A; Dickerson, Kelli L; Matthew, Richard; Harron, Connor; Quas, Catherine M
2017-01-01
Little is known about how emotion recognition and empathy jointly operate in youth growing up in contexts defined by persistent adversity. We investigated whether adversity exposure in two groups of youth was associated with reduced empathy and whether deficits in emotion recognition mediated this association. Foster, rural poor, and comparison youth from Swaziland, Africa identified emotional expressions and rated their empathic concern for characters depicted in images showing positive, ambiguous, and negative scenes. Rural and foster youth perceived greater anger and happiness in the main characters in ambiguous and negative images than did comparison youth. Rural children also perceived less sadness. Youth's perceptions of sadness in the negative and ambiguous expressions mediated the relation between adversity and empathic concern, but only for the rural youth, who perceived less sadness, which then predicted less empathy. Findings provide new insight into processes that underlie empathic tendencies in adversity-exposed youth and highlight potential directions for interventions to increase empathy.
Fast cat-eye effect target recognition based on saliency extraction
NASA Astrophysics Data System (ADS)
Li, Li; Ren, Jianlin; Wang, Xingbin
2015-09-01
Background complexity is a main reason that results in false detection in cat-eye target recognition. Human vision has selective attention property which can help search the salient target from complex unknown scenes quickly and precisely. In the paper, we propose a novel cat-eye effect target recognition method named Multi-channel Saliency Processing before Fusion (MSPF). This method combines traditional cat-eye target recognition with the selective characters of visual attention. Furthermore, parallel processing enables it to achieve fast recognition. Experimental results show that the proposed method performs better in accuracy, robustness and speed compared to other methods.
[The present state and progress of researches on gait recognition].
Xue, Zhaojun; Jin, Jingna; Ming, Dong; Wan, Baikun
2008-10-01
Recognition by gait is a new field for the biometric recognition technology. Its aim is to recognize people and detect physiological, pathological and mental characters by their walk style. The use of gait as a biometric for human identification is promising. The technique of gait recognition, as an attractive research area of biomedical information detection, attracts more and more attention. In this paper is introduced a survey of the basic theory, existing gait recognition methods and potential prospects. The latest progress and key factors of research difficulties are analyzed, and future researches are envisaged.
ERIC Educational Resources Information Center
Yeari, Menahem; Isser, Michal; Schiff, Rachel
2017-01-01
A controversy has recently developed regarding the hypothesis that developmental dyslexia may be caused, in some cases, by a reduced visual attention span (VAS). To examine this hypothesis, independent of phonological abilities, researchers tested the ability of dyslexic participants to recognize arrays of unfamiliar visual characters. Employing…
ERIC Educational Resources Information Center
Oyeleye, Omobola Awosika
2013-01-01
The purpose of the study was to examine the circumstances and practices that led to OCR and DOJ investigations in seven selected school districts, and to determine the emerging themes from the details of the settlement agreements between the school districts and the United States. The themes developed through this study were aimed at providing a…
Real-Time Measurement of Host Bioenergetics During Mycobacterium Tuberculosis Infection
2015-05-01
antimycobacterial drugs on Mtb bioenergetics. We focused on Clofazimine (CFZ, targets Complex I), Bedaquiline (BDQ/TMC207, targets Complex V) and Q203 (targets... Complex III). Firstly we investigated the effect of CFZ and BDQ on the OCR profiles of Mtb mc2 6230 (Figure 3). These experiments were done in...addition with of CFZ. The decrease in OCR is consistent with ETC complex inhibition. BDQ caused a very surprising concentration-depended increase
NASA Astrophysics Data System (ADS)
Liasi, Faezeh Talebi; Samatham, Ravikant; Jacques, Steven L.
2017-11-01
Assessing the metabolic activity of a tissue, whether normal, damaged, aged, or pathologic, is useful for diagnosis and evaluating the effects of drugs. This report describes a handheld optical fiber probe that contacts the skin, applies pressure to blanch the superficial vascular plexus of the skin, then releases the pressure to allow refill of the plexus. The optical probe uses white light spectroscopy to record the time dynamics of blanching and refilling. The magnitude and dynamics of changes in blood content and hemoglobin oxygen saturation yield an estimate of the oxygen consumption rate (OCR) in units of attomoles per cell per second. The average value of OCR on nine forearm sites on five subjects was 10±5 (amol/cell/s). This low-cost, portable, rapid, noninvasive optical probe can characterize the OCR of a skin site to assess the metabolic activity of the epidermis or a superficial lesion.
Cho, Zang-Hee; Kim, Nambeom; Bae, Sungbong; Chi, Je-Geun; Park, Chan-Woong; Ogawa, Seiji; Kim, Young-Bo
2014-10-01
The two basic scripts of the Korean writing system, Hanja (the logography of the traditional Korean character) and Hangul (the more newer Korean alphabet), have been used together since the 14th century. While Hanja character has its own morphemic base, Hangul being purely phonemic without morphemic base. These two, therefore, have substantially different outcomes as a language as well as different neural responses. Based on these linguistic differences between Hanja and Hangul, we have launched two studies; first was to find differences in cortical activation when it is stimulated by Hanja and Hangul reading to support the much discussed dual-route hypothesis of logographic and phonological routes in the brain by fMRI (Experiment 1). The second objective was to evaluate how Hanja and Hangul affect comprehension, therefore, recognition memory, specifically the effects of semantic transparency and morphemic clarity on memory consolidation and then related cortical activations, using functional magnetic resonance imaging (fMRI) (Experiment 2). The first fMRI experiment indicated relatively large areas of the brain are activated by Hanja reading compared to Hangul reading. The second experiment, the recognition memory study, revealed two findings, that is there is only a small difference in recognition memory for semantic transparency, while for the morphemic clarity was much larger between Hanja and Hangul. That is the morphemic clarity has significantly more effect than semantic transparency on recognition memory when studies by fMRI in correlation with behavioral study.
Comparing the Frequency Effect Between the Lexical Decision and Naming Tasks in Chinese
Wu, Jei-Tun
2016-01-01
In psycholinguistic research, the frequency effect can be one of the indicators for eligible experimental tasks that examine the nature of lexical access. Usually, only one of those tasks is chosen to examine lexical access in a study. Using two exemplar experiments, this paper introduces an approach to include both the lexical decision task and the naming task in a study. In the first experiment, the stimuli were Chinese characters with frequency and regularity manipulated. In the second experiment, the stimuli were switched to Chinese two-character words, in which the word frequency and the regularity of the leading character were manipulated. The logic of these two exemplar experiments was to explore some important issues such as the role of phonology on recognition by comparing the frequency effect between both the tasks. The results revealed different patterns of lexical access from those reported in the alphabetic systems. The results of Experiment 1 manifested a larger frequency effect in the naming task as compared to the LDT, when the stimuli were Chinese characters. And it is noteworthy that, in Experiment 1, when the stimuli were regular Chinese characters, the frequency effect observed in the naming task was roughly equivalent to that in the LDT. However, a smaller frequency effect was shown in the naming task as compared to the LDT, when the stimuli were switched to Chinese two-character words in Experiment 2. Taking advantage of the respective demands and characteristics in both tasks, researchers can obtain a more complete and precise picture of character/word recognition. PMID:27077703
Speech Recognition Technology for Disabilities Education
ERIC Educational Resources Information Center
Tang, K. Wendy; Kamoua, Ridha; Sutan, Victor; Farooq, Omer; Eng, Gilbert; Chu, Wei Chern; Hou, Guofeng
2005-01-01
Speech recognition is an alternative to traditional methods of interacting with a computer, such as textual input through a keyboard. An effective system can replace or reduce the reliability on standard keyboard and mouse input. This can especially assist dyslexic students who have problems with character or word use and manipulation in a textual…
Neighborhood Frequency Effect in Chinese Word Recognition: Evidence from Naming and Lexical Decision
ERIC Educational Resources Information Center
Li, Meng-Feng; Gao, Xin-Yu; Chou, Tai-Li; Wu, Jei-Tun
2017-01-01
Neighborhood frequency is a crucial variable to know the nature of word recognition. Different from alphabetic scripts, neighborhood frequency in Chinese is usually confounded by component character frequency and neighborhood size. Three experiments were designed to explore the role of the neighborhood frequency effect in Chinese and the stimuli…
ERIC Educational Resources Information Center
Murphy, Harry; Higgins, Eleanor
This final report describes the activities and accomplishments of a 3-year study on the compensatory effectiveness of three assistive technologies, optical character recognition, speech synthesis, and speech recognition, on postsecondary students (N=140) with learning disabilities. These technologies were investigated relative to: (1) immediate…
Is Syntactic-Category Processing Obligatory in Visual Word Recognition? Evidence from Chinese
ERIC Educational Resources Information Center
Wong, Andus Wing-Kuen; Chen, Hsuan-Chih
2012-01-01
Three experiments were conducted to investigate how syntactic-category and semantic information is processed in visual word recognition. The stimuli were two-character Chinese words in which semantic and syntactic-category ambiguities were factorially manipulated. A lexical decision task was employed in Experiment 1, whereas a semantic relatedness…
Federal Register 2010, 2011, 2012, 2013, 2014
2011-10-17
... such as logos and special characters. Identifying information that you provide, such as phone numbers... are further made in recognition of the position, set out in the revisions to proposed question and...-day notice period. However, in recognition of standard provisions in many contracts entered into...
Offline handwritten word recognition using MQDF-HMMs
NASA Astrophysics Data System (ADS)
Ramachandrula, Sitaram; Hambarde, Mangesh; Patial, Ajay; Sahoo, Dushyant; Kochar, Shaivi
2015-01-01
We propose an improved HMM formulation for offline handwriting recognition (HWR). The main contribution of this work is using modified quadratic discriminant function (MQDF) [1] within HMM framework. In an MQDF-HMM the state observation likelihood is calculated by a weighted combination of MQDF likelihoods of individual Gaussians of GMM (Gaussian Mixture Model). The quadratic discriminant function (QDF) of a multivariate Gaussian can be rewritten by avoiding the inverse of covariance matrix by using the Eigen values and Eigen vectors of it. The MQDF is derived from QDF by substituting few of badly estimated lower-most Eigen values by an appropriate constant. The estimation errors of non-dominant Eigen vectors and Eigen values of covariance matrix for which the training data is insufficient can be controlled by this approach. MQDF has been successfully shown to improve the character recognition performance [1]. The usage of MQDF in HMM improves the computation, storage and modeling power of HMM when there is limited training data. We have got encouraging results on offline handwritten character (NIST database) and word recognition in English using MQDF HMMs.
Syntactic/semantic techniques for feature description and character recognition
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gonzalez, R.C.
1983-01-01
The Pattern Analysis Branch, Mapping, Charting and Geodesy (MC/G) Division, of the Naval Ocean Research and Development Activity (NORDA) has been involved over the past several years in the development of algorithms and techniques for computer recognition of free-form handprinted symbols as they appear on the Defense Mapping Agency (DMA) maps and charts. NORDA has made significant contributions to the automation of MC/G through advancing the state of the art in such information extraction techniques. In particular, new concepts in character (symbol) skeletonization, rugged feature measurements, and expert system-oriented decision logic have allowed the development of a very high performancemore » Handprinted Symbol Recognition (HSR) system for identifying depth soundings from naval smooth sheets (accuracies greater than 99.5%). The study reported in this technical note is part of NORDA's continuing research and development in pattern and shape analysis as it applies to Navy and DMA ocean/environment problems. The issue addressed in this technical note deals with emerging areas of syntactic and semantic techniques in pattern recognition as they might apply to the free-form symbol problem.« less
Experiments on Urdu Text Recognition
NASA Astrophysics Data System (ADS)
Mukhtar, Omar; Setlur, Srirangaraj; Govindaraju, Venu
Urdu is a language spoken in the Indian subcontinent by an estimated 130-270 million speakers. At the spoken level, Urdu and Hindi are considered dialects of a single language because of shared vocabulary and the similarity in grammar. At the written level, however, Urdu is much closer to Arabic because it is written in Nastaliq, the calligraphic style of the Persian-Arabic script. Therefore, a speaker of Hindi can understand spoken Urdu but may not be able to read written Urdu because Hindi is written in Devanagari script, whereas an Arabic writer can read the written words but may not understand the spoken Urdu. In this chapter we present an overview of written Urdu. Prior research in handwritten Urdu OCR is very limited. We present (perhaps) the first system for recognizing handwritten Urdu words. On a data set of about 1300 handwritten words, we achieved an accuracy of 70% for the top choice, and 82% for the top three choices.
Lossef, S V; Schwartz, L H
1990-09-01
A computerized reference system for radiology journal articles was developed by using an IBM-compatible personal computer with a hand-held optical scanner and optical character recognition software. This allows direct entry of scanned text from printed material into word processing or data-base files. Additionally, line diagrams and photographs of radiographs can be incorporated into these files. A text search and retrieval software program enables rapid searching for keywords in scanned documents. The hand scanner and software programs are commercially available, relatively inexpensive, and easily used. This permits construction of a personalized radiology literature file of readily accessible text and images requiring minimal typing or keystroke entry.
Text Line Detection from Rectangle Traffic Panels of Natural Scene
NASA Astrophysics Data System (ADS)
Wang, Shiyuan; Huang, Linlin; Hu, Jian
2018-01-01
Traffic sign detection and recognition is very important for Intelligent Transportation. Among traffic signs, traffic panel contains rich information. However, due to low resolution and blur in the rectangular traffic panel, it is difficult to extract the character and symbols. In this paper, we propose a coarse-to-fine method to detect the Chinese character on traffic panels from natural scenes. Given a traffic panel Color Quantization is applied to extract candidate regions of Chinese characters. Second, a multi-stage filter based on learning is applied to discard the non-character regions. Third, we aggregate the characters for text lines by Distance Metric Learning method. Experimental results on real traffic images from Baidu Street View demonstrate the effectiveness of the proposed method.
Adjunctive Local Application of Lidocaine during Scleral Buckling under General Anesthesia.
Dehghani, Alireza; Montazeri, Kamran; Masjedi, Amin; Karbasi, Najmeh; Ashrafi, Leila; Saeedian, Behrooz
2011-07-01
To evaluate the effect of local lidocaine application on the incidence of the oculocardiac reflex (OCR) during scleral buckling (SB) for rhegmatogenous retinal detachment (RRD) under general anesthesia. In a randomized clinical trial, eyes with RRD scheduled for SB under general anesthesia were randomized to adjunctive local application of 1 ml lidocaine 2% versus normal saline to the muscles after conjunctival opening. Surgical stimulation was initiated 5 minutes afterwards. Additionally, 100 mg of lidocaine 2% was added to 50 ml of normal saline in the treatment group which was used for irrigation during surgery; control eyes were irrigated with normal saline. The incidence of the OCR, rate of postoperative nausea/vomiting (PONV), total intravenous (IV) analgesic dose, duration of surgery, and period of hospitalization were compared between the study groups. Thirty eyes of 30 patients including 22 (73.3%) male and 8 (26.7%) subjects with mean age of 49.4±16.3 years were operated. OCR and PONV occurred less frequently, and total intravenous analgesic dose was significantly lower in the lidocaine group (P < 0.05 for all comparisons). However, no significant difference was noted between the study groups in terms of duration of surgery and period of hospitalization. Adjunctive local application of lidocaine during SB under GA for RRD decreases the rate of OCR and PONV, reduces the intravenous analgesic dose, but does not affect the duration of surgery or hospitalization.
Einstein, Samuel A; Weegman, Bradley P; Kitzmann, Jennifer P; Papas, Klearchos K; Garwood, Michael
2017-05-01
Transplantation of macroencapsulated tissue-engineered grafts (TEGs) is being investigated as a treatment for type 1 diabetes, but there is a critical need to measure TEG viability both in vitro and in vivo. Oxygen deficiency is the most critical issue preventing widespread implementation of TEG transplantation and delivery of supplemental oxygen (DSO) has been shown to enhance TEG survival and function in vivo. In this study, we demonstrate the first use of oxygen-17 magnetic resonance spectroscopy ( 17 O-MRS) to measure the oxygen consumption rate (OCR) of TEGs and show that in addition to providing therapeutic benefits to TEGs, DSO with 17 O 2 can also enable measurements of TEG viability. Macroencapsulated TEGs containing βTC3 murine insulinoma cells were prepared with three fractional viabilities and provided with 17 O 2 . Cellular metabolism of 17 O 2 into nascent mitochondrial water (H 2 17 O) was monitored by 17 O-MRS and, from the measured data, OCR was calculated. For comparison, OCR was simultaneously measured on a separate, but equivalent sample of cells with a well-established stirred microchamber technique. OCR measured by 17 O-MRS agreed well with measurements made in the stirred microchamber device. These studies confirm that 17 O-MRS can quantify TEG viability noninvasively. Biotechnol. Bioeng. 2017;114: 1118-1121. © 2016 Wiley Periodicals, Inc. © 2016 Wiley Periodicals, Inc.
Differentiation of perceptual and semantic subsequent memory effects using an orthographic paradigm.
Kuo, Michael C C; Liu, Karen P Y; Ting, Kin Hung; Chan, Chetwyn C H
2012-11-27
This study aimed to differentiate perceptual and semantic encoding processes using subsequent memory effects (SMEs) elicited by the recognition of orthographs of single Chinese characters. Participants studied a series of Chinese characters perceptually (by inspecting orthographic components) or semantically (by determining the object making sounds), and then made studied or unstudied judgments during the recognition phase. Recognition performance in terms of d-prime measure in the semantic condition was higher, though not significant, than that of the perceptual condition. The between perceptual-semantic condition differences in SMEs at P550 and late positive component latencies (700-1000ms) were not significant in the frontal area. An additional analysis identified larger SME in the semantic condition during 600-1000ms in the frontal pole regions. These results indicate that coordination and incorporation of orthographic information into mental representation is essential to both task conditions. The differentiation was also revealed in earlier SMEs (perceptual>semantic) at N3 (240-360ms) latency, which is a novel finding. The left-distributed N3 was interpreted as more efficient processing of meaning with semantically learned characters. Frontal pole SMEs indicated strategic processing by executive functions, which would further enhance memory. Copyright © 2012 Elsevier B.V. All rights reserved.
Feature-extracted joint transform correlation.
Alam, M S
1995-12-10
A new technique for real-time optical character recognition that uses a joint transform correlator is proposed. This technique employs feature-extracted patterns for the reference image to detect a wide range of characters in one step. The proposed technique significantly enhances the processing speed when compared with the presently available joint transform correlator architectures and shows feasibility for multichannel joint transform correlation.
Younger and Older Users’ Recognition of Virtual Agent Facial Expressions
Beer, Jenay M.; Smarr, Cory-Ann; Fisk, Arthur D.; Rogers, Wendy A.
2015-01-01
As technology advances, robots and virtual agents will be introduced into the home and healthcare settings to assist individuals, both young and old, with everyday living tasks. Understanding how users recognize an agent’s social cues is therefore imperative, especially in social interactions. Facial expression, in particular, is one of the most common non-verbal cues used to display and communicate emotion in on-screen agents (Cassell, Sullivan, Prevost, & Churchill, 2000). Age is important to consider because age-related differences in emotion recognition of human facial expression have been supported (Ruffman et al., 2008), with older adults showing a deficit for recognition of negative facial expressions. Previous work has shown that younger adults can effectively recognize facial emotions displayed by agents (Bartneck & Reichenbach, 2005; Courgeon et al. 2009; 2011; Breazeal, 2003); however, little research has compared in-depth younger and older adults’ ability to label a virtual agent’s facial emotions, an import consideration because social agents will be required to interact with users of varying ages. If such age-related differences exist for recognition of virtual agent facial expressions, we aim to understand if those age-related differences are influenced by the intensity of the emotion, dynamic formation of emotion (i.e., a neutral expression developing into an expression of emotion through motion), or the type of virtual character differing by human-likeness. Study 1 investigated the relationship between age-related differences, the implication of dynamic formation of emotion, and the role of emotion intensity in emotion recognition of the facial expressions of a virtual agent (iCat). Study 2 examined age-related differences in recognition expressed by three types of virtual characters differing by human-likeness (non-humanoid iCat, synthetic human, and human). Study 2 also investigated the role of configural and featural processing as a possible explanation for age-related differences in emotion recognition. First, our findings show age-related differences in the recognition of emotions expressed by a virtual agent, with older adults showing lower recognition for the emotions of anger, disgust, fear, happiness, sadness, and neutral. These age-related difference might be explained by older adults having difficulty discriminating similarity in configural arrangement of facial features for certain emotions; for example, older adults often mislabeled the similar emotions of fear as surprise. Second, our results did not provide evidence for the dynamic formation improving emotion recognition; but, in general, the intensity of the emotion improved recognition. Lastly, we learned that emotion recognition, for older and younger adults, differed by character type, from best to worst: human, synthetic human, and then iCat. Our findings provide guidance for design, as well as the development of a framework of age-related differences in emotion recognition. PMID:25705105
Character-level neural network for biomedical named entity recognition.
Gridach, Mourad
2017-06-01
Biomedical named entity recognition (BNER), which extracts important named entities such as genes and proteins, is a challenging task in automated systems that mine knowledge in biomedical texts. The previous state-of-the-art systems required large amounts of task-specific knowledge in the form of feature engineering, lexicons and data pre-processing to achieve high performance. In this paper, we introduce a novel neural network architecture that benefits from both word- and character-level representations automatically, by using a combination of bidirectional long short-term memory (LSTM) and conditional random field (CRF) eliminating the need for most feature engineering tasks. We evaluate our system on two datasets: JNLPBA corpus and the BioCreAtIvE II Gene Mention (GM) corpus. We obtained state-of-the-art performance by outperforming the previous systems. To the best of our knowledge, we are the first to investigate the combination of deep neural networks, CRF, word embeddings and character-level representation in recognizing biomedical named entities. Copyright © 2017 Elsevier Inc. All rights reserved.
Analog design of a new neural network for optical character recognition.
Morns, I P; Dlay, S S
1999-01-01
An electronic circuit is presented for a new type of neural network, which gives a recognition rate of over 100 kHz. The network is used to classify handwritten numerals, presented as Fourier and wavelet descriptors, and has been shown to train far quicker than the popular backpropagation network while maintaining classification accuracy.
Detailed Phonetic Labeling of Multi-language Database for Spoken Language Processing Applications
2015-03-01
which contains about 60 interfering speakers as well as background music in a bar. The top panel is again clean training /noisy testing settings, and...recognition system for Mandarin was developed and tested. Character recognition rates as high as 88% were obtained, using an approximately 40 training ...Tool_ComputeFeat.m) .............................................................................................................. 50 6.3. Training
26 CFR 1.367(a)-6T - Transfer of foreign branch with previously deducted losses (temporary).
Code of Federal Regulations, 2011 CFR
2011-04-01
... the recognition of the gain realized on the transfer. Paragraph (c) of this section sets forth rules concerning the character of, and limitations on, the gain required to be recognized. Paragraph (d) of this... section. Finally, paragraph (g) of this section defines the term foreign branch. (b) Recognition of gain...
Synthesis of Common Arabic Handwritings to Aid Optical Character Recognition Research.
Dinges, Laslo; Al-Hamadi, Ayoub; Elzobi, Moftah; El-Etriby, Sherif
2016-03-11
Document analysis tasks such as pattern recognition, word spotting or segmentation, require comprehensive databases for training and validation. Not only variations in writing style but also the used list of words is of importance in the case that training samples should reflect the input of a specific area of application. However, generation of training samples is expensive in the sense of manpower and time, particularly if complete text pages including complex ground truth are required. This is why there is a lack of such databases, especially for Arabic, the second most popular language. However, Arabic handwriting recognition involves different preprocessing, segmentation and recognition methods. Each requires particular ground truth or samples to enable optimal training and validation, which are often not covered by the currently available databases. To overcome this issue, we propose a system that synthesizes Arabic handwritten words and text pages and generates corresponding detailed ground truth. We use these syntheses to validate a new, segmentation based system that recognizes handwritten Arabic words. We found that a modification of an Active Shape Model based character classifiers-that we proposed earlier-improves the word recognition accuracy. Further improvements are achieved, by using a vocabulary of the 50,000 most common Arabic words for error correction.
Synthesis of Common Arabic Handwritings to Aid Optical Character Recognition Research
Dinges, Laslo; Al-Hamadi, Ayoub; Elzobi, Moftah; El-etriby, Sherif
2016-01-01
Document analysis tasks such as pattern recognition, word spotting or segmentation, require comprehensive databases for training and validation. Not only variations in writing style but also the used list of words is of importance in the case that training samples should reflect the input of a specific area of application. However, generation of training samples is expensive in the sense of manpower and time, particularly if complete text pages including complex ground truth are required. This is why there is a lack of such databases, especially for Arabic, the second most popular language. However, Arabic handwriting recognition involves different preprocessing, segmentation and recognition methods. Each requires particular ground truth or samples to enable optimal training and validation, which are often not covered by the currently available databases. To overcome this issue, we propose a system that synthesizes Arabic handwritten words and text pages and generates corresponding detailed ground truth. We use these syntheses to validate a new, segmentation based system that recognizes handwritten Arabic words. We found that a modification of an Active Shape Model based character classifiers—that we proposed earlier—improves the word recognition accuracy. Further improvements are achieved, by using a vocabulary of the 50,000 most common Arabic words for error correction. PMID:26978368
Enhancement and character recognition of the erased colophon of a 15th-century Hebrew prayer book
NASA Astrophysics Data System (ADS)
Walvoord, Derek J.; Easton, Roger L., Jr.; Knox, Keith T.; Heimbueger, Matthew
2005-01-01
A handwritten codex often included an inscription that listed facts about its publication, such as the names of the scribe and patron, date of publication, the city where the book was copied, etc. These facts obviously provide essential information to a historian studying the provenance of the codex. Unfortunately, this page was sometimes erased after the sale of the book to a new owner, often by scraping off the original ink. The importance of recovering this information would be difficult to overstate. This paper reports on the methods of imaging, image enhancement, and character recognition that were applied to this page in a Hebrew prayer book copied in Florence in the 15th century.
Enhancement and character recognition of the erased colophon of a 15th-century Hebrew prayer book
NASA Astrophysics Data System (ADS)
Walvoord, Derek J.; Easton, Roger L., Jr.; Knox, Keith T.; Heimbueger, Matthew
2004-12-01
A handwritten codex often included an inscription that listed facts about its publication, such as the names of the scribe and patron, date of publication, the city where the book was copied, etc. These facts obviously provide essential information to a historian studying the provenance of the codex. Unfortunately, this page was sometimes erased after the sale of the book to a new owner, often by scraping off the original ink. The importance of recovering this information would be difficult to overstate. This paper reports on the methods of imaging, image enhancement, and character recognition that were applied to this page in a Hebrew prayer book copied in Florence in the 15th century.
Liquid lens: advances in adaptive optics
NASA Astrophysics Data System (ADS)
Casey, Shawn Patrick
2010-12-01
'Liquid lens' technologies promise significant advancements in machine vision and optical communications systems. Adaptations for machine vision, human vision correction, and optical communications are used to exemplify the versatile nature of this technology. Utilization of liquid lens elements allows the cost effective implementation of optical velocity measurement. The project consists of a custom image processor, camera, and interface. The images are passed into customized pattern recognition and optical character recognition algorithms. A single camera would be used for both speed detection and object recognition.
NASA Technical Reports Server (NTRS)
Knasel, T. Michael
1996-01-01
The primary goal of the Adaptive Vision Laboratory Research project was to develop advanced computer vision systems for automatic target recognition. The approach used in this effort combined several machine learning paradigms including evolutionary learning algorithms, neural networks, and adaptive clustering techniques to develop the E-MOR.PH system. This system is capable of generating pattern recognition systems to solve a wide variety of complex recognition tasks. A series of simulation experiments were conducted using E-MORPH to solve problems in OCR, military target recognition, industrial inspection, and medical image analysis. The bulk of the funds provided through this grant were used to purchase computer hardware and software to support these computationally intensive simulations. The payoff from this effort is the reduced need for human involvement in the design and implementation of recognition systems. We have shown that the techniques used in E-MORPH are generic and readily transition to other problem domains. Specifically, E-MORPH is multi-phase evolutionary leaming system that evolves cooperative sets of features detectors and combines their response using an adaptive classifier to form a complete pattern recognition system. The system can operate on binary or grayscale images. In our most recent experiments, we used multi-resolution images that are formed by applying a Gabor wavelet transform to a set of grayscale input images. To begin the leaming process, candidate chips are extracted from the multi-resolution images to form a training set and a test set. A population of detector sets is randomly initialized to start the evolutionary process. Using a combination of evolutionary programming and genetic algorithms, the feature detectors are enhanced to solve a recognition problem. The design of E-MORPH and recognition results for a complex problem in medical image analysis are described at the end of this report. The specific task involves the identification of vertebrae in x-ray images of human spinal columns. This problem is extremely challenging because the individual vertebra exhibit variation in shape, scale, orientation, and contrast. E-MORPH generated several accurate recognition systems to solve this task. This dual use of this ATR technology clearly demonstrates the flexibility and power of our approach.
Adversity, emotion recognition, and empathic concern in high-risk youth
Quas, Jodi A.; Matthew, Richard; Harron, Connor; Quas, Catherine M.
2017-01-01
Little is known about how emotion recognition and empathy jointly operate in youth growing up in contexts defined by persistent adversity. We investigated whether adversity exposure in two groups of youth was associated with reduced empathy and whether deficits in emotion recognition mediated this association. Foster, rural poor, and comparison youth from Swaziland, Africa identified emotional expressions and rated their empathic concern for characters depicted in images showing positive, ambiguous, and negative scenes. Rural and foster youth perceived greater anger and happiness in the main characters in ambiguous and negative images than did comparison youth. Rural children also perceived less sadness. Youth’s perceptions of sadness in the negative and ambiguous expressions mediated the relation between adversity and empathic concern, but only for the rural youth, who perceived less sadness, which then predicted less empathy. Findings provide new insight into processes that underlie empathic tendencies in adversity-exposed youth and highlight potential directions for interventions to increase empathy. PMID:28738074
A Dynamic Bayesian Network Based Structural Learning towards Automated Handwritten Digit Recognition
NASA Astrophysics Data System (ADS)
Pauplin, Olivier; Jiang, Jianmin
Pattern recognition using Dynamic Bayesian Networks (DBNs) is currently a growing area of study. In this paper, we present DBN models trained for classification of handwritten digit characters. The structure of these models is partly inferred from the training data of each class of digit before performing parameter learning. Classification results are presented for the four described models.
Biochemical and physical factors affecting discoloration characteristics of 19 bovine muscles.
McKenna, D R; Mies, P D; Baird, B E; Pfeiffer, K D; Ellebracht, J W; Savell, J W
2005-08-01
Steaks from muscles (n=19 from nine beef carcasses) were evaluated over the course of retail display (0-, 1-, 2-, 3-, 4- or 5-d) for objective measures of discoloration (metmyoglobin, oxymyoglobin, L*-, a*-, and b*-values), reducing ability (metmyoglobin reductase activity (MRA), resistance to induced metmyoglobin formation (RIMF), and nitric oxide metmyoglobin reducing ability (NORA)), oxygen consumption rate (OCR), oxygen penetration depth, myoglobin content, oxidative rancidity, and pH. Muscles were grouped according to objective color measures of discoloration. M. longissimus lumborum, M. longissimus thoracis, M. semitendinosus, and M. tensor fasciae latae were grouped as "high" color stability muscles, M. semimembranosus, M. rectus femoris, and M. vastus lateralis were grouped as "moderate" color stability muscles, M. trapezius, M. gluteus medius, and M. latissimus dorsi were grouped as "intermediate" color stability muscles, M. triceps brachi - long head, M. biceps femoris, M. pectoralis profundus, M. adductor, M. triceps brachi - lateral head, and M. serratus ventralis were grouped as "low" color stability muscles, and M. supraspinatus, M. infraspinatus, and M. psoas major were grouped as "very low" color stability muscles. Generally, muscles of high color stability had high RIMF, nitric oxide reducing ability, and oxygen penetration depth and possessed low OCRs, myoglobin content, and oxidative rancidity. In contrast, muscles of low color stability had high MRA, OCRs, myoglobin content, and oxidative rancidity and low RIMF, NORA, and oxygen penetration depth. Data indicate that discoloration differences between muscles are related to the amount of reducing activity relative to the OCR.
Adjunctive Local Application of Lidocaine during Scleral Buckling under General Anesthesia
Dehghani, Alireza; Montazeri, Kamran; Masjedi, Amin; Karbasi, Najmeh; Ashrafi, Leila; Saeedian, Behrooz
2011-01-01
Purpose To evaluate the effect of local lidocaine application on the incidence of the oculocardiac reflex (OCR) during scleral buckling (SB) for rhegmatogenous retinal detachment (RRD) under general anesthesia. Methods In a randomized clinical trial, eyes with RRD scheduled for SB under general anesthesia were randomized to adjunctive local application of 1 ml lidocaine 2% versus normal saline to the muscles after conjunctival opening. Surgical stimulation was initiated 5 minutes afterwards. Additionally, 100 mg of lidocaine 2% was added to 50 ml of normal saline in the treatment group which was used for irrigation during surgery; control eyes were irrigated with normal saline. The incidence of the OCR, rate of postoperative nausea/vomiting (PONV), total intravenous (IV) analgesic dose, duration of surgery, and period of hospitalization were compared between the study groups. Results Thirty eyes of 30 patients including 22 (73.3%) male and 8 (26.7%) subjects with mean age of 49.4±16.3 years were operated. OCR and PONV occurred less frequently, and total intravenous analgesic dose was significantly lower in the lidocaine group (P < 0.05 for all comparisons). However, no significant difference was noted between the study groups in terms of duration of surgery and period of hospitalization. Conclusion Adjunctive local application of lidocaine during SB under GA for RRD decreases the rate of OCR and PONV, reduces the intravenous analgesic dose, but does not affect the duration of surgery or hospitalization. PMID:22454732
Altered Glycolysis and Mitochondrial Respiration in a Zebrafish Model of Dravet Syndrome.
Kumar, Maneesh G; Rowley, Shane; Fulton, Ruth; Dinday, Matthew T; Baraban, Scott C; Patel, Manisha
2016-01-01
Altered metabolism is an important feature of many epileptic syndromes but has not been reported in Dravet syndrome (DS), a catastrophic childhood epilepsy associated with mutations in a voltage-activated sodium channel, Nav1.1 (SCN1A). To address this, we developed novel methodology to assess real-time changes in bioenergetics in zebrafish larvae between 4 and 6 d postfertilization (dpf). Baseline and 4-aminopyridine (4-AP) stimulated glycolytic flux and mitochondrial respiration were simultaneously assessed using a Seahorse Biosciences extracellular flux analyzer. Scn1Lab mutant zebrafish showed a decrease in baseline glycolytic rate and oxygen consumption rate (OCR) compared to controls. A ketogenic diet formulation rescued mutant zebrafish metabolism to control levels. Increasing neuronal excitability with 4-AP resulted in an immediate increase in glycolytic rates in wild-type zebrafish, whereas mitochondrial OCR increased slightly and quickly recovered to baseline values. In contrast, scn1Lab mutant zebrafish showed a significantly slower and exaggerated increase of both glycolytic rates and OCR after 4-AP. The underlying mechanism of decreased baseline OCR in scn1Lab mutants was not because of altered mitochondrial DNA content or dysfunction of enzymes in the electron transport chain or tricarboxylic acid cycle. Examination of glucose metabolism using a PCR array identified five glycolytic genes that were downregulated in scn1Lab mutant zebrafish. Our findings in scn1Lab mutant zebrafish suggest that glucose and mitochondrial hypometabolism contribute to the pathophysiology of DS.
Neural basis of hierarchical visual form processing of Japanese Kanji characters.
Higuchi, Hiroki; Moriguchi, Yoshiya; Murakami, Hiroki; Katsunuma, Ruri; Mishima, Kazuo; Uno, Akira
2015-12-01
We investigated the neural processing of reading Japanese Kanji characters, which involves unique hierarchical visual processing, including the recognition of visual components specific to Kanji, such as "radicals." We performed functional MRI to measure brain activity in response to hierarchical visual stimuli containing (1) real Kanji characters (complete structure with semantic information), (2) pseudo Kanji characters (subcomponents without complete character structure), (3) artificial characters (character fragments), and (4) checkerboard (simple photic stimuli). As we expected, the peaks of the activation in response to different stimulus types were aligned within the left occipitotemporal visual region along the posterior-anterior axis in order of the structural complexity of the stimuli, from fragments (3) to complete characters (1). Moreover, only the real Kanji characters produced functional connectivity between the left inferotemporal area and the language area (left inferior frontal triangularis), while pseudo Kanji characters induced connectivity between the left inferotemporal area and the bilateral cerebellum and left putamen. Visual processing of Japanese Kanji takes place in the left occipitotemporal cortex, with a clear hierarchy within the region such that the neural activation differentiates the elements in Kanji characters' fragments, subcomponents, and semantics, with different patterns of connectivity to remote regions among the elements.
When is the right hemisphere holistic and when is it not? The case of Chinese character recognition.
Chung, Harry K S; Leung, Jacklyn C Y; Wong, Vienne M Y; Hsiao, Janet H
2018-05-15
Holistic processing (HP) has long been considered a characteristic of right hemisphere (RH) processing. Indeed, holistic face processing is typically associated with left visual field (LVF)/RH processing advantages. Nevertheless, expert Chinese character recognition involves reduced HP and increased RH lateralization, presenting a counterexample. Recent modeling research suggests that RH processing may be associated with an increase or decrease in HP, depending on whether spacing or component information was used respectively. Since expert Chinese character recognition involves increasing sensitivity to components while deemphasizing spacing information, RH processing in experts may be associated with weaker HP than novices. Consistent with this hypothesis, in a divided visual field paradigm, novices exhibited HP only in the LVF/RH, whereas experts showed no HP in either visual field. This result suggests that the RH may flexibly switch between part-based and holistic representations, consistent with recent fMRI findings. The RH's advantage in global/low spatial frequency processing is suggested to be relative to the task relevant frequency range. Thus, its use of holistic and part-based representations may depend on how attention is allocated for task relevant information. This study provides the first behavioral evidence showing how type of information used for processing modulates perceptual representations in the RH. Copyright © 2018 Elsevier B.V. All rights reserved.
Diard, Julien; Rynik, Vincent; Lorenceau, Jean
2013-01-01
This research involves a novel apparatus, in which the user is presented with an illusion inducing visual stimulus. The user perceives illusory movement that can be followed by the eye, so that smooth pursuit eye movements can be sustained in arbitrary directions. Thus, free-flow trajectories of any shape can be traced. In other words, coupled with an eye-tracking device, this apparatus enables "eye writing," which appears to be an original object of study. We adapt a previous model of reading and writing to this context. We describe a probabilistic model called the Bayesian Action-Perception for Eye On-Line model (BAP-EOL). It encodes probabilistic knowledge about isolated letter trajectories, their size, high-frequency components of the produced trajectory, and pupil diameter. We show how Bayesian inference, in this single model, can be used to solve several tasks, like letter recognition and novelty detection (i.e., recognizing when a presented character is not part of the learned database). We are interested in the potential use of the eye writing apparatus by motor impaired patients: the final task we solve by Bayesian inference is disability assessment (i.e., measuring and tracking the evolution of motor characteristics of produced trajectories). Preliminary experimental results are presented, which illustrate the method, showing the feasibility of character recognition in the context of eye writing. We then show experimentally how a model of the unknown character can be used to detect trajectories that are likely to be new symbols, and how disability assessment can be performed by opportunistically observing characteristics of fine motor control, as letter are being traced. Experimental analyses also help identify specificities of eye writing, as compared to handwriting, and the resulting technical challenges.
Diard, Julien; Rynik, Vincent; Lorenceau, Jean
2013-01-01
This research involves a novel apparatus, in which the user is presented with an illusion inducing visual stimulus. The user perceives illusory movement that can be followed by the eye, so that smooth pursuit eye movements can be sustained in arbitrary directions. Thus, free-flow trajectories of any shape can be traced. In other words, coupled with an eye-tracking device, this apparatus enables “eye writing,” which appears to be an original object of study. We adapt a previous model of reading and writing to this context. We describe a probabilistic model called the Bayesian Action-Perception for Eye On-Line model (BAP-EOL). It encodes probabilistic knowledge about isolated letter trajectories, their size, high-frequency components of the produced trajectory, and pupil diameter. We show how Bayesian inference, in this single model, can be used to solve several tasks, like letter recognition and novelty detection (i.e., recognizing when a presented character is not part of the learned database). We are interested in the potential use of the eye writing apparatus by motor impaired patients: the final task we solve by Bayesian inference is disability assessment (i.e., measuring and tracking the evolution of motor characteristics of produced trajectories). Preliminary experimental results are presented, which illustrate the method, showing the feasibility of character recognition in the context of eye writing. We then show experimentally how a model of the unknown character can be used to detect trajectories that are likely to be new symbols, and how disability assessment can be performed by opportunistically observing characteristics of fine motor control, as letter are being traced. Experimental analyses also help identify specificities of eye writing, as compared to handwriting, and the resulting technical challenges. PMID:24273525
ERIC Educational Resources Information Center
Haight, Larry
1989-01-01
Types of specialty software that can help in computer editing are discussed, including programs for file transformation, optical character recognition, facsimile transmission, spell-checking, style assistance, editing, indexing, and headline-writing. (MSE)
Invariant approach to the character classification
NASA Astrophysics Data System (ADS)
Šariri, Kristina; Demoli, Nazif
2008-04-01
Image moments analysis is a very useful tool which allows image description invariant to translation and rotation, scale change and some types of image distortions. The aim of this work was development of simple method for fast and reliable classification of characters by using Hu's and affine moment invariants. Measure of Eucleidean distance was used as a discrimination feature with statistical parameters estimated. The method was tested in classification of Times New Roman font letters as well as sets of the handwritten characters. It is shown that using all Hu's and three affine invariants as discrimination set improves recognition rate by 30%.
ERIC Educational Resources Information Center
Congress of the U.S., Washington, DC. House Committee on Education and Labor.
Since 1981, the Office for Civil Rights (OCR) of the United States Department of Education has been accused of failing to enforce the civil rights laws according to its mandate. OCR is responsible for enforcing Federal laws prohibiting discrimination on the basis of race, sex, national origin, handicap, or age in educational programs or activities…
NASA Astrophysics Data System (ADS)
Tanioka, Toshimasa; Egashira, Hiroyuki; Takata, Mayumi; Okazaki, Yasuhisa; Watanabe, Kenzi; Kondo, Hiroki
We have designed and implemented a PC operation support system for a physically disabled person with a speech impediment via voice. Voice operation is an effective method for a physically disabled person with involuntary movement of the limbs and the head. We have applied a commercial speech recognition engine to develop our system for practical purposes. Adoption of a commercial engine reduces development cost and will contribute to make our system useful to another speech impediment people. We have customized commercial speech recognition engine so that it can recognize the utterance of a person with a speech impediment. We have restricted the words that the recognition engine recognizes and separated a target words from similar words in pronunciation to avoid misrecognition. Huge number of words registered in commercial speech recognition engines cause frequent misrecognition for speech impediments' utterance, because their utterance is not clear and unstable. We have solved this problem by narrowing the choice of input down in a small number and also by registering their ambiguous pronunciations in addition to the original ones. To realize all character inputs and all PC operation with a small number of words, we have designed multiple input modes with categorized dictionaries and have introduced two-step input in each mode except numeral input to enable correct operation with small number of words. The system we have developed is in practical level. The first author of this paper is physically disabled with a speech impediment. He has been able not only character input into PC but also to operate Windows system smoothly by using this system. He uses this system in his daily life. This paper is written by him with this system. At present, the speech recognition is customized to him. It is, however, possible to customize for other users by changing words and registering new pronunciation according to each user's utterance.
Naz, Saeeda; Umar, Arif Iqbal; Ahmed, Riaz; Razzak, Muhammad Imran; Rashid, Sheikh Faisal; Shafait, Faisal
2016-01-01
The recognition of Arabic script and its derivatives such as Urdu, Persian, Pashto etc. is a difficult task due to complexity of this script. Particularly, Urdu text recognition is more difficult due to its Nasta'liq writing style. Nasta'liq writing style inherits complex calligraphic nature, which presents major issues to recognition of Urdu text owing to diagonality in writing, high cursiveness, context sensitivity and overlapping of characters. Therefore, the work done for recognition of Arabic script cannot be directly applied to Urdu recognition. We present Multi-dimensional Long Short Term Memory (MDLSTM) Recurrent Neural Networks with an output layer designed for sequence labeling for recognition of printed Urdu text-lines written in the Nasta'liq writing style. Experiments show that MDLSTM attained a recognition accuracy of 98% for the unconstrained Urdu Nasta'liq printed text, which significantly outperforms the state-of-the-art techniques.
NASA Technical Reports Server (NTRS)
Liu, Hua-Kuang (Editor); Schenker, Paul (Editor)
1987-01-01
The papers presented in this volume provide an overview of current research in both optical and digital pattern recognition, with a theme of identifying overlapping research problems and methodologies. Topics discussed include image analysis and low-level vision, optical system design, object analysis and recognition, real-time hybrid architectures and algorithms, high-level image understanding, and optical matched filter design. Papers are presented on synthetic estimation filters for a control system; white-light correlator character recognition; optical AI architectures for intelligent sensors; interpreting aerial photographs by segmentation and search; and optical information processing using a new photopolymer.
Automated extraction of radiation dose information from CT dose report images.
Li, Xinhua; Zhang, Da; Liu, Bob
2011-06-01
The purpose of this article is to describe the development of an automated tool for retrieving texts from CT dose report images. Optical character recognition was adopted to perform text recognitions of CT dose report images. The developed tool is able to automate the process of analyzing multiple CT examinations, including text recognition, parsing, error correction, and exporting data to spreadsheets. The results were precise for total dose-length product (DLP) and were about 95% accurate for CT dose index and DLP of scanned series.
PIMS-Universal Payload Information Management
NASA Technical Reports Server (NTRS)
Elmore, Ralph; McNair, Ann R. (Technical Monitor)
2002-01-01
As the overall manager and integrator of International Space Station (ISS) science payloads and experiments, the Payload Operations Integration Center (POIC) at Marshall Space Flight Center had a critical need to provide an information management system for exchange and management of ISS payload files as well as to coordinate ISS payload related operational changes. The POIC's information management system has a fundamental requirement to provide secure operational access not only to users physically located at the POIC, but also to provide collaborative access to remote experimenters and International Partners. The Payload Information Management System (PIMS) is a ground based electronic document configuration management and workflow system that was built to service that need. Functionally, PIMS provides the following document management related capabilities: 1. File access control, storage and retrieval from a central repository vault. 2. Collect supplemental data about files in the vault. 3. File exchange with a PMS GUI client, or any FTP connection. 4. Files placement into an FTP accessible dropbox for pickup by interfacing facilities, included files transmitted for spacecraft uplink. 5. Transmission of email messages to users notifying them of new version availability. 6. Polling of intermediate facility dropboxes for files that will automatically be processed by PIMS. 7. Provide an API that allows other POIC applications to access PIMS information. Functionally, PIMS provides the following Change Request processing capabilities: 1. Ability to create, view, manipulate, and query information about Operations Change Requests (OCRs). 2. Provides an adaptable workflow approval of OCRs with routing through developers, facility leads, POIC leads, reviewers, and implementers. Email messages can be sent to users either involving them in the workflow process or simply notifying them of OCR approval progress. All PIMS document management and OCR workflow controls are coordinated through and routed to individual user's "to do" list tasks. A user is given a task when it is their turn to perform some action relating to the approval of the Document or OCR. The user's available actions are restricted to only functions available for the assigned task. Certain actions, such as review or action implementation by non-PIMS users, can also be coordinated through automated emails.
The Profiles in Science Digital Library: Behind the Scenes.
Gallagher, Marie E; Moffatt, Christie
2012-01-01
This demonstration shows the Profiles in Science ® digital library. Profiles in Science contains digitized selections from the personal manuscript collections of prominent biomedical researchers, medical practitioners, and those fostering science and health. The Profiles in Science Web site is the delivery mechanism for content derived from the digital library system. The system is designed according to our basic principles for digital library development [1]. The digital library includes the rules and software used for digitizing items, creating and editing database records and performing quality control as well as serving the digital content to the public. Among the types of data managed by the digital library are detailed item-level, collection-level and cross-collection metadata, digitized photographs, papers, audio clips, movies, born-digital electronic files, optical character recognized (OCR) text, and annotations (see Figure 1). The digital library also tracks the status of each item, including digitization quality, sensitivity of content, and copyright. Only items satisfying all required criteria are released to the public through the World Wide Web. External factors have influenced all aspects of the digital library's infrastructure.
Offline Arabic handwriting recognition: a survey.
Lorigo, Liana M; Govindaraju, Venu
2006-05-01
The automatic recognition of text on scanned images has enabled many applications such as searching for words in large volumes of documents, automatic sorting of postal mail, and convenient editing of previously printed documents. The domain of handwriting in the Arabic script presents unique technical challenges and has been addressed more recently than other domains. Many different methods have been proposed and applied to various types of images. This paper provides a comprehensive review of these methods. It is the first survey to focus on Arabic handwriting recognition and the first Arabic character recognition survey to provide recognition rates and descriptions of test data for the approaches discussed. It includes background on the field, discussion of the methods, and future research directions.
Murdison, T Scott; Paré-Bingley, Chanel A; Blohm, Gunnar
2013-08-01
To compute spatially correct smooth pursuit eye movements, the brain uses both retinal motion and extraretinal signals about the eyes and head in space (Blohm and Lefèvre 2010). However, when smooth eye movements rely solely on memorized target velocity, such as during anticipatory pursuit, it is unknown if this velocity memory also accounts for extraretinal information, such as head roll and ocular torsion. To answer this question, we used a novel behavioral updating paradigm in which participants pursued a repetitive, spatially constant fixation-gap-ramp stimulus in series of five trials. During the first four trials, participants' heads were rolled toward one shoulder, inducing ocular counterroll (OCR). With each repetition, participants increased their anticipatory pursuit gain, indicating a robust encoding of velocity memory. On the fifth trial, they rolled their heads to the opposite shoulder before pursuit, also inducing changes in ocular torsion. Consequently, for spatially accurate anticipatory pursuit, the velocity memory had to be updated across changes in head roll and ocular torsion. We tested how the velocity memory accounted for head roll and OCR by observing the effects of changes to these signals on anticipatory trajectories of the memory decoding (fifth) trials. We found that anticipatory pursuit was updated for changes in head roll; however, we observed no evidence of compensation for OCR, representing the absence of ocular torsion signals within the velocity memory. This indicated that the directional component of the memory must be coded retinally and updated to account for changes in head roll, but not OCR.
Clark, Robyn A; Tideman, Phillip; Tirimacco, Rosy; Wanguhu, Ken; Poulsen, Vanessa; Simpson, Paul; Duncan, Mitch J; Van Itallie, Anetta; Corry, Kelly; Vandelanotte, Corneel; Mummery, W Kerry
2013-05-01
Interventions that facilitate access to cardiac rehabilitation and secondary prevention programs are in demand. This pilot study used a mixed methods design to evaluate the feasibility of an Internet-based, electronic Outpatient Cardiac Rehabilitation (eOCR). Patients who had suffered a cardiac event and their case managers were recruited from rural primary practices. Feasibility was evaluated in terms of the number of patients enrolled and patient and case manager engagement with the eOCR website. Four rural general practices, 16 health professionals (cardiologists, general practitioners, nurses and allied health) and 24 patients participated in the project and 11 (46%) completed the program. Utilisation of the website during the 105 day evaluation period by participating health professionals was moderate to low (mean of 8.25 logins, range 0-28 logins). The mean login rate for patients was 16 (range 1-77 logins), mean time from first login to last (days using the website) was 51 (range 1-105 days). Each patient monitored at least five risk factors and read at least one of the secondary prevention articles. There was low utilisation of other tools such as weekly workbooks and discussion boards. It was important to evaluate how an eOCR website would be used within an existing healthcare setting. These results will help to guide the implementation of future internet based cardiac rehabilitation programs considering barriers such as access and appropriate target groups of participants. Crown Copyright © 2012. Published by Elsevier B.V. All rights reserved.
Big Data Quality Case Study Preliminary Findings, U.S. Army MEDCOM MODS
2013-09-01
captured in electronic form is relatively small, on the order of hundreds of thousands of health profiles at say around 500K per profile, or in the...in electronic form, then different language identification, handwriting recognition, and Natural Language Processing (NLP) techniques could be used...and patterns” [15]. Volume - The free text fields vary in length from say ten characters to several hundred characters. Other materials can be much
ERIC Educational Resources Information Center
Physics Education, 1986
1986-01-01
Describes (1) computer graphics for the coefficient of restitution; (2) an experiment on the optical processing of images; and (3) a simple, coherent optical system for character recognition using Polaroid (Type 665) negative film. (JN)
Who was that masked man? Conjoint representations of intrinsic motions with actor appearance.
Kersten, Alan W; Earles, Julie L; Negri, Leehe
2018-09-01
Motion plays an important role in recognising animate creatures. This research supports a distinction between intrinsic and extrinsic motions in their relationship to identifying information about the characters performing the motions. Participants viewed events involving costumed human characters. Intrinsic motions involved relative movements of a character's body parts, whereas extrinsic motions involved movements with respect to external landmarks. Participants were later tested for recognition of the motions and who had performed them. The critical test items involved familiar characters performing motions that had previously been performed by other characters. Participants falsely recognised extrinsic conjunction items, in which characters followed the paths of other characters, more often than intrinsic conjunction items, in which characters moved in the manner of other characters. In contrast, participants falsely recognised new extrinsic motions less often than new intrinsic motions, suggesting that they remembered extrinsic motions but had difficulty remembering who had performed them. Modelling of receiver operating characteristics indicated that participants discriminated old items from intrinsic conjunction items via familiarity, consistent with conjoint representations of intrinsic motion and identity information. In contrast, participants used recollection to distinguish old items from extrinsic conjunction items, consistent with separate but associated representations of extrinsic motion and identity information.
Miwa, Koji; Libben, Gary; Dijkstra, Ton; Baayen, Harald
2014-01-01
This lexical decision study with eye tracking of Japanese two-kanji-character words investigated the order in which a whole two-character word and its morphographic constituents are activated in the course of lexical access, the relative contributions of the left and the right characters in lexical decision, the depth to which semantic radicals are processed, and how nonlinguistic factors affect lexical processes. Mixed-effects regression analyses of response times and subgaze durations (i.e., first-pass fixation time spent on each of the two characters) revealed joint contributions of morphographic units at all levels of the linguistic structure with the magnitude and the direction of the lexical effects modulated by readers' locus of attention in a left-to-right preferred processing path. During the early time frame, character effects were larger in magnitude and more robust than radical and whole-word effects, regardless of the font size and the type of nonwords. Extending previous radical-based and character-based models, we propose a task/decision-sensitive character-driven processing model with a level-skipping assumption: Connections from the feature level bypass the lower radical level and link up directly to the higher character level.
De Novo Design of Protein Mimics of B-DNA
Yüksel, Deniz; Bianco, Piero R.; Kumar, Krishna
2015-01-01
Structural mimicry of DNA is utilized in nature as a strategy to evade molecular defences mounted by host organisms. One such example is the protein Ocr – the first translation product to be expressed as the bacteriophage T7 infects E. coli. The structure of Ocr reveals an intricate and deliberate arrangement of negative charges that endows it with the ability to mimic ∼24 base pair stretches of B–DNA. This uncanny resemblance to DNA enables Ocr to compete in binding the type I restriction modification (R/M) system, and neutralizes the threat of hydrolytic cleavage of viral genomic material. Here, we report the de novo design and biophysical characterization of DNA mimicking peptides, and describe the inhibitory action of the designed helical bundles on a type I R/M enzyme, EcoR124I. This work validates the use of charge patterning as a design principle for creation of protein mimics of DNA, and serves as a starting point for development of therapeutic peptide inhibitors against human pathogens that employ molecular camouflage as part of their invasion stratagem. PMID:26568416
Goal-oriented rectification of camera-based document images.
Stamatopoulos, Nikolaos; Gatos, Basilis; Pratikakis, Ioannis; Perantonis, Stavros J
2011-04-01
Document digitization with either flatbed scanners or camera-based systems results in document images which often suffer from warping and perspective distortions that deteriorate the performance of current OCR approaches. In this paper, we present a goal-oriented rectification methodology to compensate for undesirable document image distortions aiming to improve the OCR result. Our approach relies upon a coarse-to-fine strategy. First, a coarse rectification is accomplished with the aid of a computationally low cost transformation which addresses the projection of a curved surface to a 2-D rectangular area. The projection of the curved surface on the plane is guided only by the textual content's appearance in the document image while incorporating a transformation which does not depend on specific model primitives or camera setup parameters. Second, pose normalization is applied on the word level aiming to restore all the local distortions of the document image. Experimental results on various document images with a variety of distortions demonstrate the robustness and effectiveness of the proposed rectification methodology using a consistent evaluation methodology that encounters OCR accuracy and a newly introduced measure using a semi-automatic procedure.
Yang, Cheng-Huei; Luo, Ching-Hsing; Yang, Cheng-Hong; Chuang, Li-Yeh
2004-01-01
Morse code is now being harnessed for use in rehabilitation applications of augmentative-alternative communication and assistive technology, including mobility, environmental control and adapted worksite access. In this paper, Morse code is selected as a communication adaptive device for disabled persons who suffer from muscle atrophy, cerebral palsy or other severe handicaps. A stable typing rate is strictly required for Morse code to be effective as a communication tool. This restriction is a major hindrance. Therefore, a switch adaptive automatic recognition method with a high recognition rate is needed. The proposed system combines counter-propagation networks with a variable degree variable step size LMS algorithm. It is divided into five stages: space recognition, tone recognition, learning process, adaptive processing, and character recognition. Statistical analyses demonstrated that the proposed method elicited a better recognition rate in comparison to alternative methods in the literature.
Learning and Inductive Inference
1982-07-01
a set of graph grammars to describe visual scenes . Other researchers have applied graph grammars to the pattern recognition of handwritten characters...345 1. Issues / 345 2. Mostows’ operationalizer / 350 0. Learning from ezamples / 360 1. Issues / 3t60 2. Learning in control and pattern recognition ...art.icleis on rote learntinig and ailvice- tAik g. K(ennieth Clarkson contributed Ltte article on grmvit atical inference, anid Geoff’ lroiney wrote
Semi-automated contour recognition using DICOMautomaton
NASA Astrophysics Data System (ADS)
Clark, H.; Wu, J.; Moiseenko, V.; Lee, R.; Gill, B.; Duzenli, C.; Thomas, S.
2014-03-01
Purpose: A system has been developed which recognizes and classifies Digital Imaging and Communication in Medicine contour data with minimal human intervention. It allows researchers to overcome obstacles which tax analysis and mining systems, including inconsistent naming conventions and differences in data age or resolution. Methods: Lexicographic and geometric analysis is used for recognition. Well-known lexicographic methods implemented include Levenshtein-Damerau, bag-of-characters, Double Metaphone, Soundex, and (word and character)-N-grams. Geometrical implementations include 3D Fourier Descriptors, probability spheres, boolean overlap, simple feature comparison (e.g. eccentricity, volume) and rule-based techniques. Both analyses implement custom, domain-specific modules (e.g. emphasis differentiating left/right organ variants). Contour labels from 60 head and neck patients are used for cross-validation. Results: Mixed-lexicographical methods show an effective improvement in more than 10% of recognition attempts compared with a pure Levenshtein-Damerau approach when withholding 70% of the lexicon. Domain-specific and geometrical techniques further boost performance. Conclusions: DICOMautomaton allows users to recognize contours semi-automatically. As usage increases and the lexicon is filled with additional structures, performance improves, increasing the overall utility of the system.
NASA Astrophysics Data System (ADS)
Mousas, Christos; Anagnostopoulos, Christos-Nikolaos
2017-06-01
This paper presents a hybrid character control interface that provides the ability to synthesize in real-time a variety of actions based on the user's performance capture. The proposed methodology enables three different performance interaction modules: the performance animation control that enables the direct mapping of the user's pose to the character, the motion controller that synthesizes the desired motion of the character based on an activity recognition methodology, and the hybrid control that lies within the performance animation and the motion controller. With the methodology presented, the user will have the freedom to interact within the virtual environment, as well as the ability to manipulate the character and to synthesize a variety of actions that cannot be performed directly by him/her, but which the system synthesizes. Therefore, the user is able to interact with the virtual environment in a more sophisticated fashion. This paper presents examples of different scenarios based on the three different full-body character control methodologies.
ERPs reveal sub-lexical processing in Chinese character recognition.
Wu, Yan; Mo, Deyuan; Tsang, Yiu-Kei; Chen, Hsuan-Chih
2012-04-18
The present study used ERPs and a lexical decision task to explore the roles of position-general and position-specific radicals and their relative time courses in processing Chinese characters. Two types of radical frequency were manipulated: the number of characters containing a specific radical irrespective of position (i.e., radical frequency or RF) and the number of characters containing a specific radical at a particular position (i.e., position-specific radical frequency or PRF). The PRF effect was found to be associated with P150, P200, and N400, whereas the RF effect was associated with P200. These results suggest that both position-general and position-specific radicals could influence character processing, but the effect of position-specific radicals appeared earlier and lasted longer than that of position-general radicals. These findings are interpreted in terms of the specific orthographic properties of the sub-lexical components of Chinese characters. Copyright © 2012 Elsevier Ireland Ltd. All rights reserved.
Character Recognition Using Novel Optoelectronic Neural Network
1993-04-01
interest will include machine learning and perception. Permanent Address: William M. Robinson c/o Dave and Judy Bartine 117 Westcliff Drive Harriman, TN 37748 This thesis was typed by William M. Robinson. 190 END
Keyboarding: An Important Skill for the Office of the Future.
ERIC Educational Resources Information Center
Burford, Anna M.
1980-01-01
Defines the components of the office of the future: data processing, micrographics, optical character recognition, telecommunications, and word processing. Also discusses teacher responsibility, student preparation, future challenges, and teacher awareness. (CT)
Enter Words and Pictures the Easy Way--Scan Them.
ERIC Educational Resources Information Center
Olivas, Jerry
1989-01-01
Discusses image scanning and optical character recognition. Describes how computer scanners work. Summarizes scan quality, scanning speed requirements, and hardware requirements for scanners. Surveys the range of scanners currently available. (MVL)
Multiscale characterization and analysis of shapes
Prasad, Lakshman; Rao, Ramana
2002-01-01
An adaptive multiscale method approximates shapes with continuous or uniformly and densely sampled contours, with the purpose of sparsely and nonuniformly discretizing the boundaries of shapes at any prescribed resolution, while at the same time retaining the salient shape features at that resolution. In another aspect, a fundamental geometric filtering scheme using the Constrained Delaunay Triangulation (CDT) of polygonized shapes creates an efficient parsing of shapes into components that have semantic significance dependent only on the shapes' structure and not on their representations per se. A shape skeletonization process generalizes to sparsely discretized shapes, with the additional benefit of prunability to filter out irrelevant and morphologically insignificant features. The skeletal representation of characters of varying thickness and the elimination of insignificant and noisy spurs and branches from the skeleton greatly increases the robustness, reliability and recognition rates of character recognition algorithms.
Reading as Active Sensing: A Computational Model of Gaze Planning in Word Recognition
Ferro, Marcello; Ognibene, Dimitri; Pezzulo, Giovanni; Pirrelli, Vito
2010-01-01
We offer a computational model of gaze planning during reading that consists of two main components: a lexical representation network, acquiring lexical representations from input texts (a subset of the Italian CHILDES database), and a gaze planner, designed to recognize written words by mapping strings of characters onto lexical representations. The model implements an active sensing strategy that selects which characters of the input string are to be fixated, depending on the predictions dynamically made by the lexical representation network. We analyze the developmental trajectory of the system in performing the word recognition task as a function of both increasing lexical competence, and correspondingly increasing lexical prediction ability. We conclude by discussing how our approach can be scaled up in the context of an active sensing strategy applied to a robotic setting. PMID:20577589
Reading as active sensing: a computational model of gaze planning in word recognition.
Ferro, Marcello; Ognibene, Dimitri; Pezzulo, Giovanni; Pirrelli, Vito
2010-01-01
WE OFFER A COMPUTATIONAL MODEL OF GAZE PLANNING DURING READING THAT CONSISTS OF TWO MAIN COMPONENTS: a lexical representation network, acquiring lexical representations from input texts (a subset of the Italian CHILDES database), and a gaze planner, designed to recognize written words by mapping strings of characters onto lexical representations. The model implements an active sensing strategy that selects which characters of the input string are to be fixated, depending on the predictions dynamically made by the lexical representation network. We analyze the developmental trajectory of the system in performing the word recognition task as a function of both increasing lexical competence, and correspondingly increasing lexical prediction ability. We conclude by discussing how our approach can be scaled up in the context of an active sensing strategy applied to a robotic setting.
Distorted Character Recognition Via An Associative Neural Network
NASA Astrophysics Data System (ADS)
Messner, Richard A.; Szu, Harold H.
1987-03-01
The purpose of this paper is two-fold. First, it is intended to provide some preliminary results of a character recognition scheme which has foundations in on-going neural network architecture modeling, and secondly, to apply some of the neural network results in a real application area where thirty years of effort has had little effect on providing the machine an ability to recognize distorted objects within the same object class. It is the author's belief that the time is ripe to start applying in ernest the results of over twenty years of effort in neural modeling to some of the more difficult problems which seem so hard to solve by conventional means. The character recognition scheme proposed utilizes a preprocessing stage which performs a 2-dimensional Walsh transform of an input cartesian image field, then sequency filters this spectrum into three feature bands. Various features are then extracted and organized into three sets of feature vectors. These vector patterns that are stored and recalled associatively. Two possible associative neural memory models are proposed for further investigation. The first being an outer-product linear matrix associative memory with a threshold function controlling the strength of the output pattern (similar to Kohonen's crosscorrelation approach [1]). The second approach is based upon a modified version of Grossberg's neural architecture [2] which provides better self-organizing properties due to its adaptive nature. Preliminary results of the sequency filtering and feature extraction preprocessing stage and discussion about the use of the proposed neural architectures is included.
Tong, Xiuli; McBride, Catherine
2017-07-01
Following a review of contemporary models of word-level processing for reading and their limitations, we propose a new hypothetical model of Chinese character reading, namely, the graded lexical space mapping model that characterizes how sublexical radicals and lexical information are involved in Chinese character reading development. The underlying assumption of this model is that Chinese character recognition is a process of competitive mappings of phonology, semantics, and orthography in both lexical and sublexical systems, operating as functions of statistical properties of print input based on the individual's specific level of reading. This model leads to several testable predictions concerning how the quasiregularity and continuity of Chinese-specific radicals are organized in memory for both child and adult readers at different developmental stages of reading.
Yeh, Su-Ling; Chou, Wei-Lun; Ho, Pokuan
2017-11-17
Most Chinese characters are compounds consisting of a semantic radical indicating semantic category and a phonetic radical cuing the pronunciation of the character. Controversy surrounds whether radicals also go through the same lexical processing as characters and, critically, whether phonetic radicals involve semantic activation since they can also be characters when standing alone. Here we examined these issues using the Stroop task whereby participants responded to the ink color of the character. The key finding was that Stroop effects were found when the character itself had a meaning unrelated to color, but contained a color name phonetic radical (e.g., "guess", with the phonetic radical "cyan", on the right) or had a meaning associated with color (e.g., "pity", with the phonetic radical "blood" on the right which has a meaning related to "red"). Such Stroop effects from the phonetic radical within a character unrelated to color support that Chinese character recognition involves decomposition of characters into their constituent radicals; with each of their meanings including phonetic radicals activated independently, even though it would inevitably interfere with that of the whole character. Compared with the morphological decomposition in English whereby the semantics of the morphemes are not necessarily activated, the unavoidable semantic activation of phonetic radicals represents a unique feature in Chinese character processing.
ERIC Educational Resources Information Center
Pan, Jinger; McBride-Chang, Catherine; Shu, Hua; Liu, Hongyun; Zhang, Yuping; Li, Hong
2011-01-01
Among 262 Chinese children, syllable awareness and rapid automatized naming (RAN) at age 5 years and invented spelling of Pinyin at age 6 years independently predicted subsequent Chinese character recognition and English word reading at ages 8 years and 10 years, even with initial Chinese character reading ability statistically controlled. In…
Container-code recognition system based on computer vision and deep neural networks
NASA Astrophysics Data System (ADS)
Liu, Yi; Li, Tianjian; Jiang, Li; Liang, Xiaoyao
2018-04-01
Automatic container-code recognition system becomes a crucial requirement for ship transportation industry in recent years. In this paper, an automatic container-code recognition system based on computer vision and deep neural networks is proposed. The system consists of two modules, detection module and recognition module. The detection module applies both algorithms based on computer vision and neural networks, and generates a better detection result through combination to avoid the drawbacks of the two methods. The combined detection results are also collected for online training of the neural networks. The recognition module exploits both character segmentation and end-to-end recognition, and outputs the recognition result which passes the verification. When the recognition module generates false recognition, the result will be corrected and collected for online training of the end-to-end recognition sub-module. By combining several algorithms, the system is able to deal with more situations, and the online training mechanism can improve the performance of the neural networks at runtime. The proposed system is able to achieve 93% of overall recognition accuracy.
Rantala, J; Raisamo, R; Lylykangas, J; Surakka, V; Raisamo, J; Salminen, K; Pakkanen, T; Hippula, A
2009-01-01
Three novel interaction methods were designed for reading six-dot Braille characters from the touchscreen of a mobile device. A prototype device with a piezoelectric actuator embedded under the touchscreen was used to create tactile feedback. The three interaction methods, scan, sweep, and rhythm, enabled users to read Braille characters one at a time either by exploring the characters dot by dot or by sensing a rhythmic pattern presented on the screen. The methods were tested with five blind Braille readers as a proof of concept. The results of the first experiment showed that all three methods can be used to convey information as the participants could accurately (91-97 percent) recognize individual characters. In the second experiment the presentation rate of the most efficient and preferred method, the rhythm, was varied. A mean recognition accuracy of 70 percent was found when the speed of presenting a single character was nearly doubled from the first experiment. The results showed that temporal tactile feedback and Braille coding can be used to transmit single-character information while further studies are still needed to evaluate the presentation of serial information, i.e., multiple Braille characters.
Rasmussen, Tyler P.; Wu, Yuejin; Joiner, Mei-ling A.; Koval, Olha M.; Wilson, Nicholas R.; Luczak, Elizabeth D.; Wang, Qinchuan; Chen, Biyi; Gao, Zhan; Zhu, Zhiyong; Wagner, Brett A.; Soto, Jamie; McCormick, Michael L.; Kutschke, William; Weiss, Robert M.; Yu, Liping; Boudreau, Ryan L.; Abel, E. Dale; Zhan, Fenghuang; Spitz, Douglas R.; Buettner, Garry R.; Song, Long-Sheng; Zingman, Leonid V.; Anderson, Mark E.
2015-01-01
Myocardial mitochondrial Ca2+ entry enables physiological stress responses but in excess promotes injury and death. However, tissue-specific in vivo systems for testing the role of mitochondrial Ca2+ are lacking. We developed a mouse model with myocardial delimited transgenic expression of a dominant negative (DN) form of the mitochondrial Ca2+ uniporter (MCU). DN-MCU mice lack MCU-mediated mitochondrial Ca2+ entry in myocardium, but, surprisingly, isolated perfused hearts exhibited higher O2 consumption rates (OCR) and impaired pacing induced mechanical performance compared with wild-type (WT) littermate controls. In contrast, OCR in DN-MCU–permeabilized myocardial fibers or isolated mitochondria in low Ca2+ were not increased compared with WT, suggesting that DN-MCU expression increased OCR by enhanced energetic demands related to extramitochondrial Ca2+ homeostasis. Consistent with this, we found that DN-MCU ventricular cardiomyocytes exhibited elevated cytoplasmic [Ca2+] that was partially reversed by ATP dialysis, suggesting that metabolic defects arising from loss of MCU function impaired physiological intracellular Ca2+ homeostasis. Mitochondrial Ca2+ overload is thought to dissipate the inner mitochondrial membrane potential (ΔΨm) and enhance formation of reactive oxygen species (ROS) as a consequence of ischemia-reperfusion injury. Our data show that DN-MCU hearts had preserved ΔΨm and reduced ROS during ischemia reperfusion but were not protected from myocardial death compared with WT. Taken together, our findings show that chronic myocardial MCU inhibition leads to previously unanticipated compensatory changes that affect cytoplasmic Ca2+ homeostasis, reprogram transcription, increase OCR, reduce performance, and prevent anticipated therapeutic responses to ischemia-reperfusion injury. PMID:26153425
Thickness and marking quality of different occlusal contact registration strips
TOLEDO, Maria Fernanda de Souza Mauá Serapião; JÓIAS, Renata Pilli; MARQUES-IASI, Yves Santini; NEVES, Ana Christina Claro; RODE, Sigmar de Mello
2014-01-01
Objectives Evaluate the thickness and the marking quality of different occlusal contact registration strips (OCRS) and a possible correlation between them. Material and Methods The following OCRS were selected: Accufilm II, BK20, BK21, BK22, BK23, BK28, and BK31. The thickness was measured in three points of the OCRS with an electronic measuring device (TESA), and the mean was calculated. To produce the marks on the strips, composite resin specimens were adapted to a universal testing machine (Versat 2000) with 40 kgf load cell at a speed of 1.0 mm/min. The mark images were photographed with a stereoscopic microscope (Stemi SV11) and processed and analyzed by the 550-Leica Qwin® analyzer. Results Values (μm) found in the 1st and 2nd thickness measurements were: Accufilm II - 16.4 and 14.2; BK20 - 10.0 and 8.1; BK21 - 9.5 and 8.0; BK22 - 9.7 and 8.7; BK23 - 9.8 and 7.9; BK28 - 12.8 and 10.0; and BK31 - 8.4 and 8.0, respectively. The mean (mm2) values found in the mark areas were: Accufilm II - 0.078; BK20 - 0.035; BK21 - 0.045; BK22 - 0.012; BK23 - 0.022; BK28 - 0.024; and BK31 - 0.024. The results were submitted to the Kruskal-Wallis (p<0.05) and Pearson’s correlation tests. Conclusions Only in the 2nd measurement, the OCRS thickness observed was similar to the value indicated by the manufacturers; the Accufilm II and the BK28 strips showed the better marks; and no correlation was found between the thickness and the marking area. PMID:25591020