Friday, March 29, 2019

answers to biometric question paper RBVRR


Section A                                                                                                                            4x3 marks =12 marks
1. Q1: WHAT ARE THE COMMON LIVELINESS TESTS USED IN INDIA [3 MARKS]

LIVENESS DETECTION:
One of the most frequently used PAD techniques is called liveness detection. The relevant aspects of this technique will be preliminarily elaborated and explained within this chapter. Subsequently, the obtained knowledge will be used to acquire and compare several different liveness detection techniques to recognize the previously described attack scenarios against biometric systems. The general task of liveness detection is to detect whether a biometric probe (e.g. a fingerprint) belongs to a living subject that is present at the point of biometric capture [Am15]. Using liveness detection techniques, a reliable recognition of dead fingers or photographed faces can be established, for example. Consequently, the risk of successful presentation attacks is significantly reduced. Thus, in addition to the regular biometric recognition, liveness detection is an important procedure aiming at an increased reliability of biometric systems. In the global market, several different methods verifying the liveness of biometric features are already established.

Among others these methods include an evaluation of anatomical characteristics, physiological processes of the human body and involuntary reactions to stimuli as well as various predictable human behaviors [Am15].

3.1 Hardware and Software-based Approaches

 Typically, liveness detection methods are divided into hardware and software-based approaches. Giving an example, special medical hardware can be used to perform an electrocardiogram or pulse oximetry to detect living subjects. For this purpose, the acquisition of additional sensors such as devices for measuring body temperature or pulse rate ([LJ09b], pp.924-925) that needs to be combined with the regular biometric test system is required. As a consequence, account should be taken of additional costs for acquisition as well as for routine maintenance. Software-based techniques, by contrast, make use of biometric data already being captured for biometric recognition of individuals. In general, these solutions are implemented as supplementary algorithms being integrated into a consisting biometric system. To give an example, these algorithms are then applied to the extracted biometric fingerprint probe in order to detect the deformation of a living finger that is pressed on the sensor ([LJ09b], p. 925). 

3.2 Passive and active Techniques 

Besides the distinction of hardware and software-based techniques another common attempt consists of a separation between passive (non-stimulating) and active (stimulating) automated liveness detection methods [Am15]. In general, passive detection techniques make use of biometric probes which were recorded through a biometric sensor. According to this, further interactions with the data subject are not necessary. For this, a typical example would be a temperature or pulse measurement taking place while the biometric probe is collected [MA14]. Active detection techniques normally require additional interaction of the biometric data subject with the biometric system. These further interactions should be requested using challenge response procedures. The different challenge response approaches can be read in ISO/IEC DIS 30107-1 as they cannot be discussed in this paper [Am15]. 

3.3 Defense against Presentation Attacks 

There are several different liveness detection techniques that have already been evaluated on the market and successfully used for presentation attack detection. Various techniques that behave as possible countermeasures against previously described attack scenarios will be elaborated first. Based on the main results of this paper, an overall evaluation of these scenarios will be carried out in chapter 3.4 subsequently. The table below contains liveness detection techniques that could be used as counter-measures for detecting various presentation attacks:

Type: Fingerprint scanner
Challenges: 2D print, dead finger, artificial finger, capacitive finger
LIVENESS TEST
Passive: pulse measurement* [Am15], temperature measurement** [MA14], sweat detection [Am15], skin resistance detection ** [KS13] Active (challenge response): Request of different fingers in random order [Am15]

Type: Vein scanner
Challenges: 2D print, dead finger, artificial finger,
Liveness test:
Passive: pulse measurement, temperature measurement, (sweat detection), skin resistance detection Active (challenge response): Request of different fingers in random order
Type: Face scanner
Challenges: 2D print, 3D face mask, video attack
Liveness test:
Passive:** natural eye blinking * [Am15], natural muscle movements while speaking [MA14] Active (challenge response): eye closing request [Am15], voice usage request** [MA14], head turning request** [Am15]

Type: Fingerprint, vein & face scanner
Challenges: 2D print, 3D face mask, dead body parts, artificial or digital fingers, veins & faces
Liveness test:
Passive: Infrared & ultraviolet light, thermal scans* [MA14], medical techniques like ECG, pulse oximetry or blood pressure reading [KS13]


Q2: WHAT ARE THE REASONS AND NEED FOR COMPRESSION OF DATA ON BIOMETRICS [3 MARKS]

COMPRESSSION IN BIOMETRIC DATABASE
Compression is a method for database protection used to be in line with legal requirements internationally and in india.

While there are hardly any legal provisions in the world that are specific to biometric data, despite the very specific character of such data, the French Data Protection Act of 1978, officially entitled the "Loi relative à l'informatique, aux fichiers et aux libertés " [English title: Act on Information Technology, Data Files and Civil Liberties] sets out specific requirements for biometric data.

 The "United Nations Resolution" of December 14, 1990, which sets out guidelines for the regulation of computerized personal data files does not have any binding force. 

On April 14, 2016, the draft General Data Protection Regulation was adopted by the European Parliament. Its provisions will be directly applicable in all 27 Member States of the European Union and the UK in May 2018.  And biometric data are clearly defined.
In a nutshell:
1.   it establishes a harmonized framework within the EU, 
2.   the right to be forgotten, 
3.   "clear" and "affirmative" consent

4.   and, amongst other things, serious penalties for failure to comply with these rules. 
Supreme court has dealt with the same in aadhar case:

With respect to the same, the majority in the Aadhaar Judgment has held as follows:-
(i)  Any authentication records are to be stored only for 6 months. The provision under Regulation 27(1) of the Aadhaar (Authentication) Regulation, 2016 allowed the Aadhaar authority to retain the transaction related data for 6 months and archive the same for 5 years. Now, the authority is only allowed to retain the data for 6 months.

(ii)  Creation of metabase, as provided for in Regulation 26 of the Aadhaar (Authentication) Regulation, 2016, has been held to be impermissible in its current form and has been suggested that the same needs suitable amendments.

(iii) Interestingly, Section 33(2) of the Aadhaar Act has also been struck down. Under this provision disclosure of information, including identity information and authentication records collected was allowed, when disclosed in the interest of national security.

(iv)  Section 57 of the Aadhaar Act, which enabled body corporate and individual to seek authentication, has been held to be unconstitutional.

(v)  Most importantly, the importance of bringing out a robust data protection regime in the form of an enactment, basis the Justice B.N.Krishna Committee Report (with necessary and appropriate modifications), has been stressed upon. For further reading on aspects of Data Protection Bill, 2018, read our earlier blog

As rightly stated by Justice R. F. Nariman, in the Indian context right to privacy will mainly cover three aspects “(i) Privacy that involves person i.e. physical privacy, (ii) informational privacy and (iii) privacy of choice. We can ground physical privacy and privacy relating to the body in Articles 19 (1) (d) and (e) read with Article 21; ground personal information privacy under Article 21; and the privacy of choice in Article 19 (1)(a) to (c), 20 (3), 21 and 25 of the Constitution of India”. We had written earlier on some aspects of this.

It is clearly laid down under the Aadhaar judgment referring to the judgment provided by nine-judge bench that “right to privacy cannot impinge without a just, fair and reasonable law: It has to fulfill the test of proportionality i.e. (i) existence of a law; (ii) must serve a legitimate State (as defined under the Indian Constitution) aim; and (iii) proportionality”.

The first requirement is an express requirement under Article 21, where no restrictions on right to privacy or no individual can be deprived of a right to privacy without a law being in existence. The second requirement of legitimate state aim makes ensures that the law enacted by the legislature which imposes restrictions on individual’s right to privacy, falls within the ambit of reasonable restriction as mandated under Article 14 of the Indian Constitution. The third requirement ensures that the means that are adopted by the legislature are proportional in nature to the object and needs sought to be fulfilled by the law.

In reaching a conclusion with respect to the above, reference has been placed to the statement of objects and reasons for Introducing the Aadhaar (Targeted Delivery of Financial and Other Subsidies, Benefits and Services) Bill, 2016 and preamble of the corresponding Act (2016), which emphasises on “a good governance, efficient, transparent, and targeted delivery of subsidies, benefits and services, the expenditure for which is incurred from the Consolidated Fund of India, to individuals residing in India through assigning of unique identity numbers to such individuals and for matters connected therewith or incidental thereto.

The judgement confirms that, there is a vital State interest in ensuring that scarce public resources are not dissipated by the diversion of resources to persons who do not qualify as recipients. Allocation of resources for human development is coupled with a legitimate concern that the utilization of resources should not be siphoned away for extraneous purposes.
Therefore, the first two requirements of the 3-part test are met by the Aadhaar scheme and the act i.e. existence of a law and the restrictions imposed on the fundamental right of an individual by the law serves legitimate state interests.

The Supreme Court also examined the third criteria of proportionality and held that the Aadhaar act meets the criteria of proportionality, as all the aspects of proportionality, stand satisfied.

In view of the above, it has been held that the Aadhaar act does not lead to any loss of privacy and is not unconstitutional, subject to certain provisions of the Aadhaar act which were struck down. The apex court resorted to the following grounds to establish that the Aadhaar act is constitutional. Firstly, only minimal information of the applicant is stored, who intends to have an Aadhaar. Secondly, there is no risk of loss of privacy as the information collected by the enrolment agencies is transmitted to the CIDR in an encrypted form. Finally, at the time of authentication, the information does not remain with the requesting entity.

Also, the Authority has mandated the use of a registered device (RD) for all authentication requests.  The RD rules out any possibility of the arbitrary use of the stored biometric, as the biometric captured is encrypted within a few seconds. The Authority gets only the registered device code for authentication and does not obtain any information relating to the IP address or the GPS location from where such information is collected. Further, the Authority or any entity is barred from collecting, keeping or maintain information for the purpose of the authentication under Section 32 (3) of the Act. Above all, there is an oversight by the Technology and Architecture Review Board (TARB) and Security Review Committee to all the information flow through the Aadhaar system. These committees endeavor to provide safeguards.

Lossy and lossless are the two methods for compression of images.

Lossy compression transform an image into another domain and also quantize and encode its coefficients. It is used to compress multimedia data such as video, audio and still images. It eliminates the redundant or unnecessary data.

Lossless compression is used to reconstruct original images from the compressed data and is used in source code, text documents and executable programs. Images can be compressed based on some transformations.

Discrete Cosine Transform (DCT) and Discrete Wavelet Transform (DWT) are the two common transformations.

DCT expresses a sequence of data points in terms of a sum of cosine functions oscillating at different frequencies. A DCT is a fourier related transform similar to the discrete fourier transform (DFT), but using only real numbers.

The discrete wavelet transform (DWT) uses a discrete set of the wavelet scales and translations obeying some defined rules. This transformation decompose the signal into mutually orthogonal set of wavelets. The wavelet is constructed from a scaling function which describes its scaling properties.
Wavelet Scalar Quantization is a DWT based algorithm. WSQ is an algorithm to compress gray scale fingerprint images. It is a Federal Bureau of Investigation (FBI) standard for the compression of fingerprint images.

In addition to WSQ, Contourlet Transform is another fingerprint compression algorithm.
Bezier curves are used in many fields such as industrial and computer-aided design, vector-based drawing, font design and 3D modeling and also used in computer graphics to model smooth curves.
Since the curve is in the convex hull of its control points and it is possible to graphically display the points and also the control points can be used to manipulate the curve intuitively.

In biometric data processing, data compression is the process of encoding information using fewer bits (or other information-bearing units) than an unencoded representation would use, through use of specific encoding schemes (compression). As with any communication, compressed data exchange only works and the message makes sense when both the sender and receiver of the information understand the encoding scheme applied to the message. Similarly, compressed biometric information can only be understood if the encoding and decoding method is agreed upon and known by all parties involved in the transaction.

The criminal justice communities throughout the world traditionally exchange fingerprint (friction ridge) imagery data primarily in 8-bit gray-scale and at 500 pixels per inch (19.7 pixels per millimeter) using the Wavelet Scalar Quantization (WSQ) fingerprint image compression algorithm.
An emerging class of next-generation systems now support friction ridge imagery at the resolution of 1000 pixels per inch (39.4 pixels per millimeter).

The NIST compression study formalizes a JPEG 2000 compression profile for 1000 ppi friction ridge imagery that can be used by stakeholders to ensure compatibility and utility of these images.


Q3. WHAT ARE DIFFEREENT TYPES OF FP SENSORS IN USE [3 marks]

Fingerprint Sensor Types
There are many types of Fingerprint Sensor Types, but most commonly used are Optical Fingerprint sensors and Capacitive fingerprint sensors.
Capacitive sensors
Capacitive sensors use an array capacitor plates to image the fingerprint. Skin is conductive enough to provide a capacitive coupling with an individual capacitive element on the array.
Ridges, being closer to the detector, have a higher capacitance and valleys have a lower capacitance.
Capacitive sensors can be sensitive to electrostatic discharge (shock) but they are insensitive to ambient lighting and are more resist contamination issues than some optical designs.
Capacitive sensors use a difference between skin-sensor and air-sensor contact in terms of capacitive values.
When a finger is placed on the sensor, an array of pixels measures the variation in capacity between the valleys and the ridges in the fingerprint.
Optical Sensors
Optical sensors use arrays of photodiode or phototransistor detectors to convert the energy in light incident on the detector into electrical charge.
The sensor package usually includes a light-emitting-diode (LED) to illuminate the finger.
With optical sensors, the finger is placed on a plate and illuminated by LED light sources. Through a prism and a system of lenses, the image is projected on a CMOS image sensor. Using frame grabber techniques, the image is stored and ready for analysis.
Capacitive
·         Greater miniaturization
·         Newer technology
·         Can be embedded into small devices
·         Prone to dirt etc since finger touches silicon
·         Relatively cheap

Optical Sensors
·         Larger sensing area since manufacturing large pure silicon chips is expensive
·         More robust. Longer life
·         More expensive
·         Better image quality and higher resolution
Fingerprint Sensor Types
A Comparison : Optical vs. Capacitive (Semiconductor) Fingerprint Sensors
Optical Sensors
Capacitive Sensors
Sensor Type
Optical
Capacitive
(Semiconductor or chip)
Sensor Surface
No special treatments or maintenance required
Usually needs surface treatments, including ESD and other protective coatingsCoatings may be uneven, wear out over time, degrade performance, and shorten product lifetime
Overall Durability
Scratch-proof, unbreakable glass platen made of material as hard as quartz. Resistant to shock, ESD, and extreme weather
Corrodes easily from repeated handling and everyday exposureSusceptible to damage by electrostatic discharge
Thin silicon chips are inherently fragile and susceptible to damage by hard external impact and scratches
Imaging Area, Resolution, and Size
Large imaging area (0.5 inch x 0.6 inch)High resolution (500 dpi)
Large image size (78,000 pixels)
Usually smaller imaging area, image size, and resolution due to greater cost of manufacturing larger, high-quality chips
Cost-Effectiveness
Low manufacturing cost, long life, no maintenance required
Consistent quality surface coatings may be expensive to produceReplacement, maintenance, and downtime costs can add up

Q4: WRITE ABOUT SIGNATURE GAIT AND KEYSTROKES [3 marks]
Signature
3.10 Signature
Signature defines the way in which one individual writes a specific word (mostly her name or a symbol). It is one of the oldest forms of biometric and has gained acceptance for several application scenarios. With the progress of technology, two different kinds of signature biometric emerged. Offline signature considers geometrical features of the signature biometric. Online signature provides cue about dynamic features of hand-written signature. These dynamic features include the speed and the acceleration of the pen movement, pen-up and pen-down times, etc
Two different feature extraction approaches are present for handwritten signatures as they aim to capture the static or the dynamic features. Geometrical features of the signature are considered for the static approach. Dynamic features of hand-written signature includes the speed and the acceleration of the pen movement, penup and pen-down times,
3.11Gait
 The posture and the way a person walks maintain uniqueness about the person. It is non-invasive and hard to conceal. It can be easily captured even at public places at a low resolution. Unlike other biometric systems, the individual is not required to pay any attention while her gait is being captured. As the gait can be captured from a distance, it becomes very useful for security. It requires detection of the subject, silhouette extraction, extraction of features, selection of features and classification. A lot of research has been carried out in gait recognition system. Body part (specially, feet and head) trajectories were introduced for extracting gait features.

A gait energy image will lead to a set of view intensive features [Liu and Sarkar, 2006]. Recognition is carried out by a match of similar patterns in the gait energy image. Silhouette quality quantification has a key role in the method in [Han and Bhanu, 2006]. A one-dimensional foreground sum signal modeling is used in [Han and Bhanu, 2006] to analyse the silhouettes. Segmenting the human into components and subsequent integration of the results to derive a common distance metric has lead to an improved performance for gait recognition [Vera-Rodriguez et al, 2013]. Using a population based generic walking model, a gait recognition system attempts to solve the challenges which the system encounters due to surface time, movement speed, and carrying condition [Liu and Sarkar, 2005]. The concept of point cloud registration has been proposed in [Lee et al, 2009] in order to avoid the problem of occlusion during gait recognition. Moreover, face and gait characteristics have been extracted using principal component analysis from a side face image and gait energy image, respectively. A set of synthetic features have been attained by integration of these features and application of multiple discriminant analysis. Performance improvement is achieved in comparison to individual biometric features [Tan et al, 2006].

3.13 Key Stroke
It is believed that there is a pattern about how a person types on a keyboard. This behavioral biometric, which is referred as key stroke dynamics, shows traits to identify or verify a person. But it is not considered to be as unique as several other biometric traits.


Section B                             9 marks x 2 questions = 18 marks
5a: WHAT ARE THE CHALLENGES IN Facial recognition
Facial recognition is the most natural means of biometric identification. The face recognition system does not require any contact with the person.The 1000 million electronic passports in circulation in mid 2017 provide a huge opportunity to implement face recognition at international borders. Guidelines to improve the quality of the reference picture embedded in the epassport micro-chip are provided by the ISO/IEC 19794-5 standard and used by the International Civil Aviation Organization 9303 standard for passport photographs.

According to the Keesing Journal of documents & Identity (March 2017) , 2 complementary topics have been identified by standardization groups.
·         Make sure the captured image has been done from a person and not from a mask, a photograph or a video screen, (liveliness check or liveness detection
·         Make sure that facial images (morphed portraits) or two or more individuals have not been joined into a reference document such as a passport.
The risks of error are related to very different factors. 

·         Take the example of a person with their biometric characteristics. We have noted that particular biometric techniques were more or less well suited to certain categories of persons. The difficulties are related to ergonomic factors of which we do not yet have a firm grasp or understanding. A certain system may work for women, but less well for men, or for young people but not for older people, for people with lighter skin, but less well for those with darker skin. 

·         Other difficulties arise in particular with facial recognition, when the person dyes or cuts their hair, changes the line of their eyebrows or grows a beard. We can imagine cases of "false acceptance" when the photo taken modifies distinctive character traits in such a way that they match another item of biometric data stored in the database. 

·         Other errors are also possible depending on the technologies used during the biometric enrollment phase. A verification photo taken with a low-quality model of camera can noticeably increase the risk of error. The accuracy of the identification relies entirely on the reliability of the equipment used to capture data. 

·         The risk of error also varies depending on the environment and the conditions of application. The light may differ from one place to another, and the same goes for the intensity or nature of background noise. The person's position may have changed. 

In the laboratory, under perfect conditions, in a controlled environment and using adapted technologies, the rate of error in detection of a face varies between 5 and 10 %. 

In addition, in a biometric control application, the rejection or acceptance rate are intertwined and can be tuned according to an acceptable level of risk. It is not possible to modify one without impact the other one. 
In the case of a nuclear plant access control application, the rate of false acceptance will be extremely reduced. You don't want ANYONE to enter by chance.

This will also impact the rate of false rejections because you will tune the system to be extremely accurate. You will probably use several authentication factors including a valid ID in addtional to biometrics (single mode or multimodal).

5b] WRITE ABOUT IRIS ACQUISITION, SEGMENTATION,BOUNDARY DETECTION AND POLAR MAPPING[ 9 marks]
3.2 Iris
Iris is considered to be another reliable biometric trait. The iris is the muscle in the eye which controls the pupil size in order to regulate the amount of light rays entering the eye. The iris can be identified as an annular region in between the sclera (white portion of the eye) and the pupil (Figure 3). The pattern of iris of an individual is also unique [Jain et al, 2004]. A set of twins also possess distinguishing iris patterns. The speed and accuracy of iris recognition have also caused widespread adoption of iris biometric. Registration of iris takes relatively longer time as it needs several iris images. 

A test template is generated upon scanning of an individual’s iris. Subsequently, the produced template is matched with the existing templates which were produced at the time of registration. Zero crossing representation of the one-dimensional wavelet transform has been proposed in [Radu et al, 2012] to encode the texture in iris. In [Sun et al, 2005], an integration of Local Feature Based Classifier (LFC) and an iris blob matcher increases the accuracy of iris recognition. The noisy images are detected by the iris blob matcher.

 Therefore, it helps in the situations where the LFC does not guarantee the performance. In [Dong et al, 2011], a set of training images is used to learn classspecific weight map for iris matching technique. Experiments have revealed the effectiveness of the technique. Moreover, a preprocessing method has been suggested to enhance the performance of iris biometric on mobile phones which usually are constrained by computing power. The system has demonstrated its capability to decide whether the person is dead or alive.

Question 6a] What should biometric vendors do to accelerate the deployment of biometric technologies? 9 marks
Answer
Biometric vendors must continue to integrate closely with three As solution vendors such as IBM, RSA, Netegrity and CA.
Biometric vendors should partner with online information solutions, including portals, online exchanges and other intranet infrastructure providers.
Targeting financial services and health care verticals will lead to early wins and long-term recurring revenue.
Targeting universities and health clubs will help educate and broaden the acceptance of biometric technologies among a diverse user base.
These systems should be sold at deep discounts in order to provide the technology as an educational and viral marketing activity.
Vendors should also focus on wireless technology providers. They will be integrating voice biometrics into cellphones to enable strong authenticated purchasing.

 Question 6b discuss FP FUSION TECHNIQUES BASED ON MINUTIAE, RIDGES, OR OTHER FEATURE EXTRACTION TECHNIQUES

With the increasing emphasis on the automatic personal identification applications, biometrics especially fingerprint identification is the most reliable and widely accepted technique. In this paper Fingerprint Verification based on fusion of Minutiae and Ridges using Strength Factors (FVMRSF) is presented.

 In the preprocessing stage the Fingerprint is Binarised and Thinned.
The Minutiae Matching Score is determined using Block Filter and Ridge matching score is estimated using Hough Transform.

The strength factors Alpha (α) and Beta (β) are used to generate Hybrid matching score for matching of fingerprints.

Levels of Fusion in Multimodal Biometric System
A detailed classification of various fusion techniques for multimodal biometric can be found in [Dinca and Hancke, 2017]. Fusion in multimodal biometric can occur at various levels – such as sensor, feature, matching score, rank and decision level fusion.
Each of these is explained in this section.

5.1 Sensor Level Fusion
This fusion strategy directly mixes raw data from various sensors (Figure 6) – for example, from the iris and fingerprint sensors.
Raw information is captured at several sensors to fuse at the very first level to generate raw fused information.

Sensor level fusion strategies can be put into following three groups: (i) single sensor multiple instances, (ii) intra-class multiple sensors, and (iii) inter-class multiple sensors.
In the case of single sensor multiple instances, a single sensor captures multiple instances of the same biometric trait. For example, a fingerprint sensor may capture multiple images of the same finger to reduce the effect of noise. Simple or weighted averaging, and mosaic construction are some of the common fusion methods in this case [Yang et al, 2005].

Multiple sensors are used to acquire multiple instances of the same biometric trait in the intraclass multiple sensors category [Yang et al, 2005]. For example, a 3D face image is obtained by using multiple face images taken from various cameras.

In the case of inter-class multiple sensors, two or more different biometric traits are used together. For example, images of palmprint and palm vein can be fused together for biometric recognition. Mosaicing is a nice application of sensor level fusion.

Several researchers have proposed fingerprint mosaicing. It provides a good recognition accuracy as it combines multiple images of the same fingerprint. Therefore, it can handle the difficulty in recognition due to data quality. The fingerprint mosaicing technique uses a modified Iterative Closest Point (ICP) [Jain and Ross, 2002] algorithm to generate 2D or 3D surfaces by considering the inputs from multiple instances.

In [Fatehpuria et al, 2006], a touchless fingerprint system is developed using a 3D touchless setting with multiple cameras and structured light illumination (SLI) to generate 2D fingerprint images and 3D fingerprint shape. This kind of set up is expensive due to deployment of multiple cameras. Alternatively, usage of a single camera and two mirrors are suggested in [Choi et al, 2010]. Two mirrors have been used to obtain finger side views.

Sensor level fusion is generally applied for the same trait. There are also instances of applying sensor level fusion for different traits. Few of these are mentioned here. Face and palmprint images are combined in [Jing et al, 2007]. Pixel level fusion is preceded by Gabor transform of the images. Infrared images of palmprint and palm vein are fused in [Wang et al, 2008]. At first, image, registration is carried out on these images. Subsequently, a pixel level fusion takes place.

5.2 Feature Level Fusion
Features extracted from several biometric traits are integrated into a single vector. According to this fusion strategy, biometric sensor signals (from camera or microphone) are preprocessed and feature vectors are derived from them independently. Then a composite feature vector is generated by combining these individual feature vectors .

The feature level fusion exhibits a better performance than score level and decision level fusion techniques as feature level fusion techniques directly deals with the unique biometric features. Normalization and selection of features are two important processes in feature level fusion. Min-max technique and media scheming based normalization is carried out to change the scale and location of feature values. Scale invariant feature transform is also carried out from the normalized images.

Dimensionality reduction through appropriate feature selection also enhances the accuracy of the techniques. Sequential forward selection, sequential backward selection, and partition about medoids are standard feature selection techniques. Particle Swarm Optimization (PSO) is applied on the feature vector for dimensionality reduction. The multimodal biometric techniques in [Raghavendra et al, 2009; 2011] uses this concept while combining face and palmprint features. Incompatibility of the feature sets among different biometric traits and non-linearity of the joint feature set of different biometric traits poses challenges for feature level fusion. A feature vector can be generated using weighted average of multiple feature vectors if those vectors correspond to same biometric. For example, this becomes possible if all of these vectors are obtained from fingerprint images of an individual. If these vectors correspond to different biometrics, then they are concatenated to obtain a single vector. Another example of feature level fusion can be found in [Kim et al, 2011]. Simultaneous use of time-of-flight (ToF) depth camera and near infrared (NIR) camera acquires face and hand vein images in a touchless acquisition set up.

Several multimodal system also combines face and ear, as ear is considered to be one of the most unchangeable feature of the human traits. Unlike face, human ear is not generally affected by age. PCA based feature extraction and a sparse representation method for feature level fusion is proposed in [Huang et al, 2013]. Experimental results reveal that this technique performs better than their unimodal components. Experiments also show that the performance is similar to that of the unimodal systems even if one of the modality is corrupted. Local 3D features (L3DF) are generated from ear and frontal face images in [Islam et al, 2013]. Feature level fusion is applied in these cases. A matrix interleaved concatenation based new approach is presented in [Ahmad et al, 2016] for face and palmprint biometrics. Discrete Cosine Transform (DCT) is used here to extract the features. Then, these features are concatenated in an interleaved matrix which estimates the parameters of the feature concatenation and exhibits their statistical distribution.

A fingerprint and iris based multimodal biometric recognition technique has been proposed in [Gawande et al, 2013]. Minutiae and wavelet features are extracted from fingerprint images. Haar wavelet and block sum techniques produce features from iris images. A feature level fusion of these four feature vectors exhibit better performance than a unimodal fingerprint or iris biometric. A feature level fusion of fingerprint and palm biometric traits has been proposed in [Mohi-ud-Din et al, 2011].

Another example of feature level fusion can be found for hand geometry recognition in the contactless multi-sensor system in [Svoboda et al, 2015] using an Intel RealSense 3D camera. This technique carries out foreground segmentation of the acquired hand image to determine the hand silhouette and contour. Then, the fingertips and the valleys are located alongside determination of the wrist line from the identified contour. Subsequently, two features vectors have been formed as follows: (i) comprising finger length, width, and wrist valley distances and (ii) finger widths as computed using a traversal of the overall hand surface and median axis to surface distances. A finger and finger-vein based system has been proposed in [Yang and Zhang, 2012]. Gabor features are extracted and the feature fusion strategy is based on a Supervised Local-Preserving Canonical Correlation Analysis (SLPCCAM). In [Yan et al, 2015], feature level fusion has also been used for a contactless multi-sample palm vein recognition technique. Automated access control systems in buildings and other secure premises are based on the capability of identifying an individual from a distance. Because of its importance in securing important establishments, it is emerging as an area of interest to the research community. Improper lighting condition or not-so-high resolution surveillance cameras pose a constraint for recognizing an individual based on her face. Use of multimodal biometric involving face and gait exhibits better performance [Ben et al, 2012; Huang et al, 2012].

Unlike traditional methods of face and gait multimodal biometric recognition, [Xing et al, 2015] fuses the features without normalization using coupled projections. A Robust Linear Programming (RLP) method [Miao et al, 2014] for multi-biometric recognition exhibits good results in noisy environments in spite of using less training data. It uses uncertain constraints and concatenates heterogeneous features from different biometric traits. Each biometric modality has been assigned a weight to specify its degree of contribution in the fusion. More weight is a greater relevance of the corresponding biometric trait. In the multimodal biometric recognition system by [Chang et al, 2003], features are extracted from face and ear using Principle Component Analysis (PCA). Subsequently, fusion takes place at the feature level. A method of liveliness detection to prevent spoofing attack in [Chetty and Wagner, 2005] also uses a feature level fusion. In a recent development, sparse-based feature-fusion [Huang et al, 2015] of physiological traits has drawn sufficient interest of researchers due to robust performance.
5.3 Matching Score Level Fusion
In the case of matching score level fusion, matching scores are separately obtained for each biometric trait and subsequently fused to arrive at an overall matching score. Matching score level fusion is also referred as measurement level fusion.
There exists three different approaches for matching score based fusion – density based, classification based, and transformation based.
The density based scheme is on the basis of distribution of scores and its application to popular models like Naive Bayesian and Gaussian Mixture Model [Murakami and Takahashi, 2015].
In the classification based approach, the matching scores of individual matching modules are concatenated to obtain a single feature vector. The decision to either accept or reject an individual is based on the classification of this feature vector.
According to the transformation based approach, the scores of individual matching module are, at first, transformed (normalized) into a pre-decided range. This transformation changes the position and scale parameters of the matching score distribution so that these normalized scores can be combined to obtain a single scalar score [Murakami and Takahashi, 2015].
These normalization techniques to handle the dissimilarities in matching score also draw attention of researchers. In another noted development of recent time, an order-preserving score fusion method has been proposed in [Liang et al, 2016].

5.4 Rank Level Fusion
Ranking the potential matches between the query template and the templates in the database generates an ordered list of all templates in the database. The first choice is the match. These ranked are obtained for every biometric trait. In a multimodal biometric recognition system, these rank orders are fused to generate a final ranking of each template in the database. Unlike score level fusion, normalization is not required for rank level fusion.

Rank level fusion is applied in [Kumar and Shekhar, 2011] for combining various methods for palmprint identification. A Nonlinear Weighted Ranks (NWR) method aggregates the ranks as obtained from individual matching modules. Rank level fusion may not always perform well in noisy conditions having low quality data. Though it has been applied on low quality fingerprints [Abaza and Ross, 2009]. It applies a derivation of the Borda count method, which involves image quality. This approach has a similarity with logical regression. But unlike logical regression, this approach does not need a training phase. Image quality is considered in this approach instead of weights. As ranks are assigned to only few of the stored templates with possible match, the rank level fusion may create a challenge for large databases. Ranks do not cover every template in the database. In this context, a Markov chain based method has been proposed in [Monwar and Gavrilova, 2011] for rank level fusion. Markov chain is used to represent a stochastic series of events, where the present or the preceding states determine the next state. A graph is used to formally model the Markov chain. A vertex in the graph represents a state or an event. An edge in the graph denotes the transition from one state to another state. At first, ranks are generated for each biometric trait. If the matching module creates partial ranking (for example, the first three ranking results), elements are inserted randomly to complete the list. The state transition probabilities are computed and the stationary distribution of the Markov chain is obtained. The templates in the database are ranked based on a decreasing order of the scores of the stationary distribution starting from the highest score. This fusion strategy is applied on a multimodal biometric recognition system involving iris, ear, and face. [Monwar and Gavrilova, 2009] proposes another method to solving this discussed problem of rank level fusion. The multimodal biometric recognition system has three matchers for each of signature, face and ear. However, fusion is carried out between the identities put out by at least two matchers.

5.5 Decision Level Fusion
In the case of a decision level fusion, each individual matcher, at first, takes its own decision. Subsequently, the fusion of various biometric modalities takes place by combining the decisions of these individual matchers. Hence, each biometric trait is independently pre-classified and the final classification is based on the fusion of the outputs of the various modalities (Figure 9). The simplest forms of decision level fusion uses logical operations such as ‘AND’ or ‘OR’. Some advanced fusion strategies at this level also use behavior knowledge space, the Dempster-Shafer theory of evidence, and Bayesian decision fusion.
When every individual decision module supplies positive outcome, then the ‘AND’ rule positively recognizes a query template. Otherwise, the ‘AND’ rule rejects the query. Hence, ‘AND’ rule is generally reliable with extremely low false acceptance rate (FAR). But false rejection rate (FRR) is higher than that of individual trait. On the contrary, the ‘OR’ rule provides positive output about a query template when at least one decision module gives positive response about it. As a result, FRR is extremely low and FAR is higher than individual trait. In [Tao and Veldhuis, 2009], an optimized threshold method has been proposed using the ‘AND’ and ‘OR’ rule. The thresholds of the classifiers are optimized during the training phase.
Majority voting is another common approach for decision fusion. If the majority of the individual traits decide positively about the query template, then final decision is positive. Majority voting method gives equal importance to each individual decision modules. Otherwise, a weighted majority voting can be applied. In this method, higher weights are assigned to decision modules which perform better.

A multi-algorithm decision level fusion is used in [Prabhakar and Jain, 2002] for fingerprints. This method considers four distinct fingerprint matching algorithms. These are based on Hough transform, string distance, 2D dynamic programming, and texture. This method selects appropriate classifiers prior to applying decision fusion. A threshold for each individual classifier influences the outcome of a decision level fusion. Here, threshold specifies a minimum score to decide whether the sample is genuine or an impostor. If the matching score of the sample is higher than the threshold, then the sample is considered as genuine. On the contrary, if the matching score of the sample is less than the threshold, then it belongs to an imposter. The classifiers are assumed to be independent of one another in some biometric systems. However, other works have assumed the dependency among the classifiers. A verification system has been introduced in [Veeramachaneni et al, 2008] based on two fusion strategies for correlated threshold classifiers. Between these two strategies, Likelihood Ratio Test (LRT) still depends on the threshold of each individual classifier. The Particle Swarm Optimisation (PSO) based decision strategy is considered effective in comparison. Even the PSO strategy performs better than some of the score level fusion methods. A real time sensor management using PSO is suggested in [Veeramachaneni et al, 2005] for a multimodal biometric management. This method performs a real time search of the optimal sensor configuration and optimal decision rule. A similar concept is proposed in [Kumar et al, 2010] which uses Ant Colony Optimization (ACO) technique, for a multimodal biometric system involving palmprint and hand vein. [Kumar and Kumar, 2015] extends this experiment on multiple multimodal databases involving palmprint and iris, fingerprint and face, and face and speech. Another decision level fusion for multimodal biometric recognition is proposed in [Paul et al, 2014]. In this multimodal system, signature, face and ear biometric features are combined with social network analysis. The Fisher image feature extraction, which is a combination of PCA and Linear Discriminant Analysis (LDA).

5.6 Hybrid Fusion Model
A hybrid fusion model which uses both pixel level fusion and score level fusion demonstrates good performance in [Kusuma and Chua, 2011]. This multi-sample face recognition (both in 2D and 3D) in [Kusuma and Chua, 2011] recombines images using principal component analysis (PCA). Two recombined images are fused using a pixel level fusion scheme. Additionally, score level fusion is also applied to produce good result.


No comments:

Post a Comment