You are on page 1of 19

Pattern Recognition Pattern Recognition, it does not a new sentence that exists in human world.

However the Pattern Recognition was quite familiar for today technologies research, no matter is in the biomedical field, digital area and Artificial Intelligence learning area. Pattern Recognition usually has been understood as the subject of researching the object, description and classification method in analysis. It is also known as a technique of collection of mathematical, statistical and heuristic of fundamental role in operating the task like human being on computer or individual machine system. Pattern Recognition is a long history for humans civilization; the word of Pattern Recognition had existed since 1960s and defines as is a field concerned with machine recognition of meaning regularities in noisy of complex environments. by Duda and Hart at year 1973, in the meaning of the act of taking in raw data and taking an action based on the category of the pattern. However there have some different definitions for the Pattern Recognition, Devroye, Gyrfi and Lugosi claim that the pattern recognition is about predict and guessing the unknown nature of an observation, a discrete quantity such as black or white, one or zero, sick or healthy, real or fake. In year 2003, Sergios Theodoridis had defined that Pattern recognition is a scientific discipline whose aim is the classification of the objects into a lot of categories or classes. Pattern recognition is also an integral part in most machine intelligence system built for decision making. Pattern Recognition system can be explain as the steps or process that allow to cope with actual and noisy data. A pattern recognition system is mainly includes three mutual-associate and differentiated processes. One is data building; the other two are pattern analysis and pattern classification. Data building is converting the original information into vector which can be dealt with by computer. Pattern analysis task is to process the data (vector), such as feature selection, feature extraction, datadimension compress and so on. The aim of pattern classification is to utilize the information acquired from pattern analysis to discipline the computer in order to accomplish the classification. General Composition of the pattern recognition system is given below:

Feature Extraction is a function of the original measurement from the sensor, the feature extraction in the Pattern Recognition generally guided by domain expertise in an attempt to find good discriminators. Classification is a Pattern Recognition problem of assigning an object to a class, the output of the Pattern Recognition system is an integer label, such as classifying a product as 1 or 0 in a quality control test. Regression is a generalization of a classification task, and the output of the Pattern Recognition system is a real-valued number, such as predicting the share value of a firm based on past performance and stock market indicators. Description is the problem of representing an object in terms of a series of primitives, and the Pattern Recognition system produces a structural or linguistic description.

Each processes of the pattern recognition system also can be defines in its own group:

In conclusion, Pattern Recognition System giving many possible for today technologies, it not just a concept but really works in some advance devices such as Facial Recognition System which using by the USA national security agency, Motion Capture or recognition that usually apply in the 3-D movies and computer games, and the Imagine and information processing in medical diagnosis. However the pattern recognition technologies by today still not advance than an expected, but it does mean wont it bring a great improve in the future.

KINECT for XBOX 360: The Natal Project Have you ever imagine that to be a controller to control the character inside the video games without holding any devices? Good news for you, this dream isnt far away from you. June 1, 2009, a Project called as The Natal Project had been announced by the Microsoft director Alex Kipman. The Natal Project is a project of the represent the idea about "controller-free gaming and entertainment experience" by Microsoft for the Xbox 360 video game platform. The KINECT enables users to control and interact with the Xbox 360 without the need to touch a game controller through a natural user interface using gestures, spoken commands, or presented objects and images. The project is aimed at broadening the Xbox 360's audience beyond its typical gamer base. Furthermore, Project Natal also been designed to be connectivity centre use the motion sensor to navigate between system windows when not playing a game. User can explore Xbox live, and browse everything that would normally do with a controller. Talking to friends would be as simple as pointing at their name and talking. There has been rumour that the Natal project will also have a built in microphone, so the user can escape from the horrible headphone devices. What technologies make the Project Natal so special and attractive? When the new KINECT for XBOX 360 had announced, people are concerned that how this devices able to function and how the devices can let the concept of the imagination become reality and exist for the people. The technology applied at the Project Natal actually is not a fresh in the digital world; however it is very familiar in digital world by today. The Project Natal is using the pattern recognition system to capture human motion and the voice from the human, let the user themselves be the controller to control the program that play by the XBOX device.

KINECT is a kind of sensor with horizontal bar connected to a small base with a motorized pivot, and is designed to be positioned lengthwise below the video display. The device features an "RGB camera, depth sensor and multi-array microphone running proprietary software", which provides full-body 3D motion capture, facial recognition, and voice recognition capabilities. The Kinect sensor's microphone array enables the Xbox 360 to conduct acoustic source localization and ambient noise suppression, allowing for things such as headset-free party chat over Xbox Live. The depth sensor consists of an infrared projector combined with a monochrome CMOS sensor, and allows the Kinect sensor to see in 3D under any ambient light conditions. The sensing range of the depth sensor is adjustable, with the Kinect software capable of automatically calibrating the sensor based on gameplay and the player's physical environment, such as the presence of chairs.

Kinect is based on software technology developed internally by Microsoft and range camera technology by Israeli developer PrimeSense, which interprets 3D scene information from a continuously-projected infrared pattern. Described by Microsoft personnel as the primary innovation of Kinect, the software technology enables advanced gesture recognition, facial recognition, and voice recognition. According to information supplied to retailers, the Kinect is capable of simultaneously tracking up to six people, including two active players for motion analysis with a feature extraction of 20 joints per player. According to information, the Kinect sensor outputs video at a frame rate of 30 Hz, with the RGB video stream at 32-bit colour VGA resolution (640480 pixels), and the monochrome video stream used for depth sensing at 16-bit QVGA resolution (320240 pixels with 65,536 levels of sensitivity). The Kinect sensor has a practical ranging limit of 1.23.5 metres (3.911 ft) distance. The sensor has an angular field of view of 57 horizontally and a 43 vertically, while the motorized pivot is capable of tilting the sensor as much as 27 either up or down. The microphone array features four microphone capsules, and operates with each channel processing 16-bit audio at a sampling rate of 16 kHz. Gesture recognition is a topic in computer science and language technology with the goal of interpreting human gestures via mathematical algorithms. Gestures can originate from any bodily motion or state but commonly originate from the face or hand. Current focuses in the field include emotion recognition from the face and hand gesture recognition. Many approaches have been made using cameras and computer vision algorithms to interpret sign language. However, the identification

and recognition of posture, gait, proxemics, and human behaviours is also the subject of gesture recognition techniques. Gesture recognition can be seen as a way for computers to begin to understand human body language, thus building a richer bridge between machines and humans than primitive text user interfaces or even GUIs (graphical user interfaces), which still limit the majority of input to keyboard and mouse.

Gesture recognition enables humans to interface with the machine (HMI) and interact naturally without any mechanical devices. Using the concept of gesture recognition, it is possible to point a finger at the computer screen so that the cursor will move accordingly. This could potentially make conventional input devices such as mouse, keyboards and even touch-screens redundant. Gesture recognition can be conducted with techniques from computer vision and image processing.

Face Recognition System In recent years face recognition has received substantial attention from researchers in biometrics, pattern recognition, and computer vision communities. The machine learning and computer graphics communities are also increasingly involved in face recognition. This common interest among researchers working in diverse fields is motivated by our remarkable ability to recognize people and the fact that human activity is a primary concern in everyday life. Besides, there are large numbers of commercial, security, and forensic applications requiring the use of face recognition technologies. These applications include automated crowd surveillance, access control, mugshot identification (e.g., for issuing driver licenses), face reconstruction, design of human computer interface (HCI), and content-based image database management. A number of commercial face recognition systems have been deployed, such as Cognitec, Eyematic, Viisage, and Identix . Automated face recognition is a relatively new concept. Developed in the 1960s, the first semi-automated system for face recognition required the administrator to locate features (such as eyes, ears, nose, and mouth) on the photographs before it calculated distances and ratios to a common reference point, which were then compared to reference data. In 1988, Kirby and Sirovich applied principle component analysis, a standard linear algebra technique, to the face recognition problem. This was considered somewhat of a milestone as it showed that less than one hundred values were required to accurately code a suitably aligned and normalized face image. In 1991, Turk and Pentland discovered that while using the eigenfaces techniques, the residual error could be used to detect faces in images to discovery that enabled reliable real-time automated face recognition systems. The technology first

captured the publics attention from the media reaction to a trial implementation at the January 2001 Super Bowl, which captured surveillance images and compared them to a database of digital mugshots. This demonstration initiated much-needed analysis on how to use the technology to support national needs while being considerate of the publics social and privacy concerns. Today, face recognition technology is being used to combat passport fraud, support law enforcement, identify missing children, and minimize benefit/identity fraud. Facial recognition research and Facial recognition Technology is a subfield in a larger field of pattern recognition research and technology. Pattern recognition technology uses statistical techniques to detect and extract patterns from data in order to match it with patterns stored in a database. The data upon which the recognition system works (such as a photo of a face) is no more than a set of discernable pixel level patterns for the system. Nevertheless, it is very important for these systems to be able to locate or detect a face in a field of vision so that it is only the image pattern of the face (and not the background noise) can be processed and analyzed. There are two predominant approaches to the face recognition problem: geometric (feature based) and photometric (view based). As researcher interest in face recognition continued, many different algorithms were developed, three of which have been well studied in face recognition literature: Principal Components Analysis (PCA), Linear Discriminant Analysis (LDA), and Elastic Bunch Graph Matching (EBGM).

Principal Components Analysis (PCA)

Principal Components Analysis commonly referred to as the use of eigenfaces, is the technique pioneered by Kirby and Sirivich in 1988. With PCA, the probe and gallery images must be the same size and must first be normalized to line up the eyes and mouth of the subjects within the images. The PCA approach is then used to reduce the dimension of the data by means of data compression basics and reveals the most effective low dimensional structure of facial patterns. This reduction in dimensions removes information that is not useful and precisely decomposes the face structure into orthogonal (uncorrelated) components known as eigenfaces. Each face image may be represented as a weighted sum (feature vector) of the eigenfaces, which are stored in a 1D array. A probe image is compared against a gallery image by

measuring the distance between their respective feature vectors. The PCA approach typically requires the full frontal face to be presented each time; otherwise the image results in poor performance. The primary advantage of this technique is that it can reduce the data needed to identify the individual to 1/1000th of the data presented.

Linear Discriminant Analysis LDA is a statistical approach for classifying samples of unknown classes based on training samples with known classes. This technique aims to maximize between-class (i.e., across users) variance and minimize within-class (i.e., within user) variance. In Figure 2 where each block represents a class, there are large variances between classes, but little variance within classes. When dealing with high dimensional face data, this technique faces the small sample size problem that arises where there are a small number of available training samples compared to the

dimensionality of the sample space.

Elastic Bunch Graph Matching EBGM relies on the concept that real face images have many non-linear characteristics that are not addressed by the linear analysis methods discussed earlier, such as variations in illumination (outdoor lighting vs. indoor fluorescents), pose (standing straight vs. leaning over) and expression (smile vs. frown). A Gabor wavelet transform creates a dynamic link architecture that projects the face onto an elastic grid. The Gabor jet is a node on the elastic grid, notated by circles on the image below, which describes the image behaviour around a given pixel. It is the result of a convolution of the image with a Gabor filter, which is used to detect shapes and to extract features using image processing. A convolution expresses the amount of overlap from functions, blending the functions together.] Recognition is based on the similarity of the Gabor filter response at each Gabor node. This biologically-based method using Gabor filters is a process executed in the visual cortex of higher mammals. The difficulty with this method is the requirement of accurate landmark localization, which can sometimes be achieved by combining PCA and LDA methods.

The facial recognition process normally has four interrelated phases or steps. The first step is face detection, the second is normalization, the third is feature extraction, and the final cumulative step is face recognition. These steps depend on each other and often use similar techniques. They may also be described as separate components of a typical facial recognition system. Nevertheless, it is useful to keep them conceptually separate for the purposes of clarity. Each of these steps poses very significant challenges to the successful operation of a facial recognition system.

Detecting a face in a probe image may be a relatively simple task for humans, but it is not so for a computer. The computer has to decide which pixels in the image is part of the face and which are not. In a typical passport photo, where the background is clear, it is easy to do, but as soon as the background becomes cluttered with other objects, the problem becomes extremely complex. Traditionally, methods that focus on facial landmarks (such as eyes), that detect face-like colours in circular regions, or that use standard feature templates, were used to detect faces. Once the face has been detected (separated from its background), the face needs to be normalized. This means that the image must be standardized in terms of size, pose, illumination, etc., relative to the images in the gallery or reference database. To normalize a probe image, the key facial landmarks must be located accurately. Using these landmarks, the normalization algorithm can (to some degree) reorient the image for slight variations. 15 Facial landmarks are the key to all systems, irrespective of the overall method of recognition. If the facial landmarks cannot be located, then the recognition process will fail. Recognition can only succeed if the

probe image and the gallery images are the same in terms of pose orientation, rotation, scale, size, etc. Normalization ensures that this similarity is achievedto a greater or lesser degree. Once the face image has been normalized, the feature extraction and recognition of the face can take place. In feature extraction, a mathematical representation called a biometric template or biometric reference is generated, which is stored in the database and will form the basis of any recognition task. Facial recognition algorithms differ in the way they translate or transform a face image (represented at this point as gray scale pixels) into a simplified mathematical representation (the features) in order to perform the recognition. It is important for successful recognition that maximal information is retained in this transformation process so that the biometric template is sufficiently distinctive. If this cannot be achieved, the algorithm will not have the discriminating ability required for successful recognition.

Pros in using Face Recognition System:


1. The general hardware devices apply on the Face Recognition System are

cheap, The Face Recognition system may just need a simple camera that able to capture humans face.

2. Face recognition systems can be installed to require a person to explicitly step

up to a camera and get their picture taken, or to automatically survey people as they pass by a camera. The later mode allows for scanning of many people at the same time.
3. For the high level Face Recognition system such as using the Elastic Bunch Graph Matching method, this method can provide assist in the security area in order to help the user identify and recognize the crime or suspect easily. 4. It belongs to No Human Touch technology so very rare chances of frauds

are there as human need not touch system.

Cons in using the Face recognition system:


1. The hardware of the system might be cheap and easily to get but not the

software. The software of face recognition can be costly due to its functions and performances.
2. The accuracy of the system was very unstable and volatile. The system might

easily effect by the ornaments that wearing on the person as like as glasses, hat and etc.
3. Hard to handle pose variation while capturing, the system may take times to

recapture and detect the face while the person change a pose. In the conclusion, the system will face difficult to detect the human face while he or she is moving.

Computer Aided Diagnosis System Recently, computer-aided diagnosis (CAD) has become a part of the routine clinical work for detection of breast cancer on mammograms at many screening sites and hospitals in the United States. This seems to indicate that CAD is beginning to be applied widely in the detection and differential diagnosis of many different types of abnormalities in medical images obtained in various examinations by use of different imaging modalities. In fact, CAD has become one of the major research subjects in medical imaging and diagnostic radiology. Although early attempts at computerized analysis of medical images were made in the 1960s, serious and systematic investigation on CAD began in the 1980s with a fundamental change in the concept for utilization of the computer output, from automated computer diagnosis to computer-aided diagnosis. Large-scale and systematic research and development of various CAD schemes were begun in the early 1980s at the Kurt Rossmann Laboratories for Radiologic Image Research in the Department of Radiology at the University of Chicago. In general a CAD unit consists of 3 main parts; there are the scanner, the software, and the viewer. The scanner is used to scan and digitize the mammogram,

similar to a desktop scanner used to digitally save photographs. Some mammograms are already captured digitally, in which case this step does not apply. The software includes sophisticated computer programs that analyze the film or image and prompt the radiologist to review areas that may suggest a lesion. An example is the CAD unit of the MAGIC-5 project, where a pc, a CCD linear scanner (with an 85 m pitch and 4096 grey levels), and a high resolution screen are used. The Graphical Users Interface (GUI), by means of a facility tool for image visualisation and elaboration, provides the support for medical diagnosis. The CAD system, in addition to the utility for the new patient acquisition, is also immediately usable in digital mammography, since it is compatible with the standard DICOM (Digital Imaging and Communications in Medicine) format. The portable version (with notebook) and the GUI are shown below:

A portable CAD unit of the MAGIC-5 project and the Graphical users interface ImageChecker DM is an early detection of breast cancer with CAD for both film-based and digital mammography. The ImageChecker uses algorithms of pattern recognition in order to find the abnormal areas and neural nets to classify abnormalities in benign and malignant regions. Radiologists have different reading approaches and preferences with CAD markings.

The ImageChecker DM system is designed to integrate with new or existing mammography.

A Computer-Aided Diagnosis (CAD-Dx) algorithm has been previously developed to assist radiologists in the diagnosis of mammographic clusters of calcifications with the modules: (a) detection of all calcification-like areas, (b) falsepositive reduction and segmentation of the detected calcifications, (c) selection of morphological and distributional features and (d) classification of the clusters. The purpose of this work was threefold: (a) optimize the existing algorithm and test on a large database, (b) rank classification features and select the best feature set, and (c) determine the impact of single and two-view feature estimation on classification and feature ranking. A Computer-Aided Diagnosis takes a digitized mammogram as an input. To extract useful lesion information from the image for the purpose of classification requires additional image processing and pattern recognition techniques. Figure shows a block diagram with the various modules that may be a part of a ComputerAided Diagnosis algorithm. Pre-processing and detection/segmentation are a part of

the detection block while feature extraction and classification are additional Computer-Aided Diagnosis functionalities.

Pre-processing, A pre-processing operation that suppresses and reduce the unimportant image features and artifacts simultaneously enhancing the features of interest often aids an accurate detection. Several methods have been proposed for the enhancement of mammographic images to facilitate detection and segmentation. Detection/Segmentation, various techniques for detection and segmentation of the calcification clusters have been described: wavelet-based techniques, linear discriminant analysis (LDA), k-nearest neighbour (kNN), fuzzy logic and gray-level thresholding and clustering. The process is about differentiation of different structures in the image and does the matching with anatomic databank. Feature extraction methodologies analyze objects and images to extract the most prominent features that are representative of the various classes of objects. Features are used as inputs to classifiers that assign them to the class that they represent.

Classification, The task of the classifier component of CAD-Dx system is to use the features provided by the feature extraction step to assign the objects of interest to a category or class. Pros
1. Computer-aided Diagnosis potential to improve diagnostic accuracy and

reduces variability in Radiologists Interpretations of Mammograms, thus reduce the mistakes and uncertainty issues occur in the analysis. 2. By the assist of Computer aided Diagnosis system, it helps reduce the risk of the patient dramatically. Doctor can follow the data analysis from the system to giving more correctly and better treatment to the patient. Cons
1. The system also has remains the uncertain while achieving high sensitivities,

system tend to have very low specificity and the benefits of using.

Reference: Pattern Recognition 1. Edward, J. & Jeffrey, L. Pattern Recognition. George Mason University. 2. Gigerenzer G. (1989). A General Algorithm for Pattern Recognition?

Behavioral and Brain Sciences, (764765). Cambridge University. 3. Liu, J., Sun J. G., & Wang, S. S. (2006). Pattern Recognition: An overview. International Journal of Computer Science and Network Security, VOL.6. College of Computer Science and Technology, Jilin University. Kinect for Xbox 360, Project Natal
1. http://www.xbox.com/en-US/community/events/e3/kinect.htm 2. http://en.wikipedia.org/wiki/Kinect 3. Eisenstein, J., Ghandeharizadeh, S., Golubchik, L., Shahabi, C., Yan, D. H., &

Zimmermann, R. Multi-Layer Gesture Recognition: An Experimental Evaluation. Department of Computer Science, University of Southern California. Face Recognition System

1. Lucas, D., & Nissenbaum, H. Facial Recognition Technology. Lancaster University, UK; Centre for the Study of Technology and Organization. & New York University; Department of Media, Culture, and Communication, Computer Science, and the Information Law Institute. 2. Lin, S. H., (2000) An Introduction to Face Recognition Technology. Informing Science Special Issue on Multimedia Informing Technologies,(vol. 3) 3. Brunelli, R., and Poggio, T., Face Recognition: Features versus Templates, IEEE Trans. Pattern Anal. Machine Intell., vol. 15,pp.1042-1052, 1993. Computer-aided Diagnosis 1. Tembey, M. (2003). Computer-Aided Diagnosis for Mammographic Microcalcification Clusters. Department of Computer Science and Engineering, College of Engineering, University of South Florida. 2. Masala, G. L. (2006). Computer Aided Detection on Mammography. World Academy of Science, Engineering and Technology

You might also like