acm-header
Sign In

Communications of the ACM

Contributed articles

Multimodal Biometrics For Enhanced Mobile Device Security


Multimodal Biometrics for Enhanced Mobile Device Security, illustrative photo

Credit: Andrij Borys Associates / Shutterstock

Millions of Mobile devices are stolen every year, along with associated credit card numbers, passwords, and other secure and personal information stored therein. Over the years, criminals have learned to crack passwords and fabricate biometric traits and have conquered practically every kind of user-authentication mechanism designed to stop them from accessing device data. Stronger mobile authentication mechanisms are clearly needed.

Back to Top

Key Insights

ins01.gif

Here, we show how multimodal biometrics promises untapped potential for protecting consumer mobile devices from unauthorized access, an authentication approach based on multiple physical and behavioral traits like face and voice. Although multimodal biometrics are deployed in homeland security, military, and law-enforcement applications,15,18 they are not yet widely integrated into consumer mobile devices. This can be attributed to implementation challenges and concern that consumers may find the approach inconvenient.

We also show multimodal biometrics can be integrated with mobile devices in a user-friendly manner and significantly improve their security. In 2015, we thus implemented a multimodal biometric system called Proteus at California State University, Fullerton, based on face and voice on an Samsung Galaxy S5 phone, integrating new multimodal biometric authentication algorithms optimized for consumer-level mobile devices and an interface that allows users to readily record multiple biometric traits. Our experiments confirm it achieves considerably greater authentication accuracy than systems based solely on face or voice alone. The next step is to integrate other biometrics (such as fingerprints and iris scans) into the system. We hope our experience encourages researchers and mobile-device manufacturers to pursue the same line of innovation.

Back to Top

Biometrics

Biometrics-based authentication establishes identity based on physical and behavioral characteristics (such as face and voice), relieving users from having to create and remember secure passwords. At the same time, it challenges attackers to fabricate human traits that, though possible, is difficult in practice.21 These advantages continue to spur adoption of biometrics-based authentication in smartphones and tablet computers.

Despite the arguable success of biometric authentication in mobile devices, several critical issues remain, including, for example, techniques for defeating iPhone TouchID and Samsung Galaxy S5 fingerprint recognition systems.2,26 Further, consumers continue to complain that modern mobile biometric systems lack robustness and often fail to recognize authorized users.4 To see how multimodal biometrics can help address these issues, we first examine their underlying causes.

Back to Top

The Mobile World

One major problem of biometric authentication in mobile devices is sample quality. A good-quality biometric sample—whether a photograph of a face, a voice recording, or a fingerprint scan—is critical for accurate identification; for example, a low-resolution photograph of a face or noisy voice recording can lead a biometric algorithm to incorrectly identify an impostor as a legitimate user, or "false acceptance." Likewise, it can cause the algorithm to declare a legitimate user an impostor, or "false rejection." Capturing high-quality samples in mobile devices is especially difficult for two main reasons. Mobile users capture biometric samples in a variety of environmental conditions; factors influencing these conditions include insufficient lighting, different poses, varying camera angles, and background noise. And biometric sensors in consumer mobile devices often trade sample quality for portability and lower cost; for example, the dimensions of an Apple iPhone's TouchID fingerprint scanner prohibit it from capturing the entire finger, making it easier to circumvent.4

Another challenge is training the biometric system to recognize the device user. The training process is based on extracting discriminative features from a set of user-supplied biometric samples. Increasing the number and variability of training samples increases identification accuracy. In practice, however, most consumers likely train their systems with few samples of limited variability for reasons of convenience. Multimodal biometrics is the key to addressing these challenges.

Back to Top

Promise of Multimodal Biometrics

Due to the presence of multiple pieces of highly independent identifying information (such as face and voice), multimodal systems can address the security and robustness challenges confronting today's mobile unimodal systems13,18 that identify people based on a single biometric characteristic. Moreover, deploying multimodal biometrics on existing mobile devices is practical; many of them already support face, voice, and fingerprint recognition. What is needed is a robust user-friendly approach for consolidating these technologies. Multimodal biometrics in consumer mobile devices deliver multiple benefits.

Increased mobile security. Attackers can defeat unimodal biometric systems by spoofing a single biometric modality used by the system. Establishing identity based on multiple modalities challenges attackers to simultaneously spoof multiple independent human traits—a significantly tougher challenge.21

More robust mobile authentication. When using multiple biometrics, one biometric modality can be used to compensate for variations and quality deficiencies in the others; for example, Proteus assesses face-image and voice-recording quality and lets the highest-quality sample have greater impact on the identification decision.

Likewise, multimodal biometrics can simplify the device-training process. Rather than provide many training samples from one modality (as they often must do in unimodal systems), users can provide fewer samples from multiple modalities. This identifying information can be consolidated to ensure sufficient training data for reliable identification.

A market ripe with opportunities. Despite the recent popularity of biometric authentication in consumer mobile devices, multimodal biometrics have had limited penetration in the mobile consumer market.1,15 This can be attributed to the concern users could find it inconvenient to record multiple biometrics. Multimodal systems can also be more difficult to design and implement than unimodal systems.

However, as we explain, these problems are solvable. Companies like Apple and Samsung have invested significantly in integrating biometric sensors (such as cameras and fingerprint readers) into their products. They can thus deploy multimodal biometrics without substantially increasing their production costs. In return, they profit from enhanced device sales due to increased security and robustness. In the following sections we discuss how to achieve such profitable security.

Back to Top

Fusing Face and Voice Biometrics

To illustrate the benefits of multimodal biometrics in consumer mobile devices, we implemented Proteus based on face and voice biometrics, choosing these modalities because most mobile devices have cameras and microphones needed for capturing them. Here, we provide an overview of face-and voice-recognition techniques, followed by an exploration of the approaches we used to reconcile them.

Face and voice recognition. We used the face-recognition technique known as FisherFaces3 in Proteus, as it works well in situations where images are captured under varying conditions, as expected in the case of face images obtained through mobile devices. FisherFaces uses pixel intensities in the face images as identifying features. In the future, we plan to explore other face-recognition techniques, including Gabor wavelets6 and Histogram Oriented Gradients (HOG).5

We used two approaches for voice recognition: Hidden Markov Models (HMM) based on the Mel-Frequency Cepstral Coefficients (MFCCs) as voice features,10 the basis of our score-level fusion scheme; and Linear Discriminant Analysis (LDA),14 the basis for our feature-level fusion scheme. Both approaches recognize a user's voice independent of phrases spoken.

Assessing face and voice sample quality. Assessing biometric sample quality is important for ensuring the accuracy of any biometric-based authentication system, particularly for mobile devices, as discussed earlier. Proteus thus assesses facial image quality based on luminosity, sharpness, and contrast, while voice-recording quality is based on signal-to-noise ratio (SNR). These classic quality metrics are well documented in the biometrics research literature.1,17,24 We plan to explore other promising metrics, including face orientation, in the future.

Proteus computes the average luminosity, sharpness, and contrast of a face image based on the intensity of the constituent pixels using approaches described in Nasrolli and Moeslund.17 It then normalizes each quality measure using the min-max normalization method to lie between [0, 1], finally computing their average to obtain a single quality score for a face image. One interesting problem here is determining the impact each quality metric has on the final face-quality score; for example, if the face image is too dark, then poor luminosity would have the greatest impact, as the absence of light would be the most significant impediment to recognition. Likewise, in a well-lit image distorted due to motion blur, sharpness would have the greatest impact.


To get its algorithm to scale to the constrained resources of the device, Proteus had to be able to shrink the size of face images to prevent the algorithm from exhausting the available device memory.


SNR is defined as a ratio of voice signal level to the level of background noise signals. To obtain a voice-quality score, Proteus adapts the probabilistic approach described in Vondrasek and Pollak25 to estimate the voice and noise signals, then normalizes the SNR value to the [0, 1] range using min-max normalization.

Multimodal biometric fusion. In multimodal biometric systems, information from different modalities can be consolidated, or fused, at the following levels:21

Feature. Either the data or the feature sets originating from multiple sensors and/or sources are fused;

Match score. The match scores generated from multiple trait-matching algorithms pertaining to the different biometric modalities are combined, and

Decision. The final decisions of multiple matching algorithms are consolidated into a single decision through techniques like majority voting.

Biometric researchers believe integrating information at earlier stages of processing (such as at the feature level) is more effective than having integration take place at a later stage (such as at the score level).20

Back to Top

Multimodal Mobile Biometrics Framework

Proteus fuses face and voice biometrics at either score or feature level. Since decision-level fusion typically produces only limited improvement,21 we did not pursue it when developing Proteus.

Proteus does its training and testing processes with videos of people holding a phone camera in front of their faces while speaking a certain phrase. From each video, the face is detected through the Viola-Jones algorithm24 and the system extracts the soundtrack. The system de-noises all sound frames to remove frequencies outside human voice range (85Hz– 255Hz) and drops frames without voice activity. It then uses the results as inputs into our fusion schemes.

Score-level fusion scheme. Figure 1 outlines our score-level fusion approach, integrating face and voice biometrics. The contribution of each modality's match score toward the final decision concerning a user's authenticity is determined by the respective sample quality. Proteus works as outlined in the following paragraphs.

Let t1 and t2, respectively, denote the average face- and voice-quality scores of the training samples from the user of the device. Next, from a test-video sequence, Proteus computes the quality scores Q1 and Q2 of the two biometrics, respectively. These four parameters are then passed to the system's weight-assignment module, which computes weights w1 and w2 for face and voice modalities, respectively. Each wi is calculated as cacm5904_e.gif, where p1 and p2 are percent proximities of Q1 to t1 and Q2 to t2, respectively. The system requests users train mostly through good-quality samples, as discussed later, so close proximity of the testing sample quality to that of training samples is a sign of a good-quality testing image. Greater weight is thus assigned to the modality with a higher-quality sample, ensuring effective integration of quality in the system's final authentication process.

The system then computes and normalizes matching scores S1 and S2 from the respective face- and voice-recognition algorithms applied to test images through z-score normalization. We chose this particular method because it is a commonly used normalization method, easy to implement, and highly efficient.11 However, we wish to experiment with more robust methods (such as the tanh and sigmoid functions) in the future. The system then computes the overall match score for the fusion scheme using the weighted sum rule as M = S1w1 + S2w2. If MT (T is the pre-selected threshold), the system will accept the user as authentic; otherwise, it declares the user to be an imposter.

Discussion. The scheme's effectiveness is expected to be greatest when t1 = Q1 and t2 = Q2. However, the system must exercise caution here to ensure significant representation of both modalities in the fusion process; for example, if Q2 differs greatly from t2 while Q1 is close to t1, the authentication process is dominated by the face modality, thus reducing the process to an almost unimodal scheme based on the face biometric. A mandated benchmark is thus required for each quality score to ensure the fusion-based authentication procedure does not grant access for a user if the benchmark for each score is not met. Without such benchmarks, the whole authentication procedure could be exposed to the risk of potential fraudulent activity, including deliberate attempts to alter the quality score of a specific biometric modality. The system must thus ensure the weight of each modality does not fall below a certain threshold so the multimodal scheme remains viable.


Storing and processing biometric data on the mobile device itself, rather than offloading these tasks to a remote server, eliminates the challenges of securely transmitting the biometric data and authentication decisions across potentially insecure networks.


In 2014, researchers at IBM proposed a score-level fusion scheme based on face, voice, and signature biometrics for iPhones and iPads.1 Their implementation considered only the quality of voice recordings, not face images, and is distinctly different from our approach, which incorporates the quality of both modalities. Further, because their goal was secure sign-in into a remote server, they outsourced the majority of computational tasks to the target server; Proteus performs all computations directly on the mobile device itself. To get its algorithm to scale to the constrained resources of the device, Proteus had to be able to shrink the size of face images to prevent the algorithm from exhausting the available device memory. Finally, Aronowitz et al.1 used multiple facial features (such as HOG and LBP) that, though arguably more robust than FisherFaces, can be prohibitively slow when executed locally on a mobile device; we plan to investigate using multiple facial features in the future.

Feature-level fusion scheme. Most multimodal feature-level fusion schemes assume the modalities to be fused are compatible (such as in Kisku et al.12 and in Ross and Govindarajan20); that is, the features of the modalities are computed in a similar fashion, based on, say, distance. Fusing face and voice modalities at the feature level is challenging because these two biometrics are incompatible: face features are pixel intensities and voice features are MFCCs. Another challenge for feature-level fusion is the curse of dimensionality arising when the fused feature vectors become excessively large. We addressed both challenges through the LDA approach. In addition, we observed LDA required less training data than neural networks and HMMs, with which we have experimented.

The process (see Figure 2) works like this:

Phase 1 (face feature extraction). The Proteus algorithm applies Principal Component Analysis (PCA) to the face feature set to perform feature selection;

Phase 2 (voice feature extraction). It extracts a set of MFCCs from each preprocessed audio frame and represents them in a matrix form where each row is used for each frame and each column for each MFCC index. And to reduce the dimensionality of the MFCC matrix, it uses the column means of the matrix as its voice feature vector;

Phase 3 (fusion of face and voice features). Since the algorithm measures face and voice features using different units, it standardizes them individually through the z-score normalization method, as in score-level fusion. The algorithm then concatenates these normalized features to form one big feature vector. If there are N face features and M voice features, it will have a total of N + M features in the concatenated, or fused, set. The algorithm then uses LDA to perform feature selection from the fused feature set. This helps address the curse of the dimensionality problem by removing irrelevant features from the combined set; and

Phase 4 (authentication). The algorithm uses Euclidean distance to determine the degree of similarity between the fused features sets from the training data and each test sample. If the distance value is less than or equal to a predetermined threshold, it accepts the test subject as a legitimate user. Otherwise, the subject is declared an impostor.

Back to Top

Implementation

We implemented our quality-based score-level and feature-level fusion approaches on a randomly selected Samsung Galaxy S5 phone. User friendliness and execution speed were our guiding principles.

User interface. Our first priority when designing the interface was to ensure users could seamlessly capture face and voice biometrics simultaneously. We thus adopted a solution that asks users to record a short video of their faces while speaking a simple phrase. The prototype of our graphical user interface (GUI) (see Figure 3) gives users real-time feedback on the quality metrics of their face and voice, guiding them to capture the best-quality samples possible; for example, if the luminosity in the video differs significantly from the average luminosity of images in the training database, the user may get a prompt saying, Suggestion: Increase lighting. In addition to being user friendly, the video also facilitates integration of other security features (such as liveness checking7) and correlation of lip movement with speech.8

To ensure fast authentication, the Proteus face- and voice-feature extraction algorithms are executed in parallel on different processor cores; the Galaxy S5 has four cores. Proteus also uses similar parallel programming techniques to help ensure the GUI's responsiveness.

Security of biometric data. The greatest risk from storing biometric data on a mobile device (Proteus stores data from multiple biometrics) is the possibility of attackers stealing and using it to impersonate a legitimate user. It is thus imperative that Proteus stores and processes the biometric data securely.

The current implementation stores only MFCCs and PCA coefficients in the device persistent memory, not raw biometric data, from which deriving useful biometric data is nontrivial.16 Proteus can enhance security significantly by using cancelable biometric templates19 and encrypting, storing, and processing biometric data in Trusted Execution Environment tamper-proof hardware highly isolated from the rest of the device software and hardware; the Galaxy S5 uses this approach to protect fingerprint data.22

Storing and processing biometric data on the mobile device itself, rather than offloading these tasks to a remote server, eliminates the challenge of securely transmitting the biometric data and authentication decisions across potentially insecure networks. In addition, this approach alleviates consumers' concern regarding the security, privacy, and misuse of their biometric data in transit to and on remote systems.

Back to Top

Performance Evaluation

We compared Proteus recognition accuracy to unimodal systems based on face and voice biometrics. We measured that accuracy using the standard equal error rate (EER) metric, or the value where the false acceptance rate (FAR) and the false rejection rate (FRR) are equal. Mechanisms enabling secure storage and processing of biometric data must therefore be in place.

Database. For our experiments, we created a CSUF-SG5 homegrown multimodal database of face and voice samples collected from University of California, Fullerton, students, employees, and individuals from outside the university using the Galaxy S5 (hence the name). To incorporate various types and levels of variations and distortions in the samples, we collected them in a variety of real-world settings. Given such a diverse database of multimodal biometrics is unavailable, we plan to make our own one publicly available. The database today includes video recordings of 54 people of different genders and ethnicities holding a phone camera in front of their faces while speaking a certain simple phrase.

The faces in these videos show the following types of variations:

Four expressions. Neutral, happy, sad, angry, and scared;

Three poses. Frontal and sideways (left and right); and

Two illumination conditions. Uniform and partial shadows.

The voice samples show different levels of background noise, from car traffic to music to people chatter, coupled with distortions in the voice itself (such as raspiness). We used 20 different popular phrases, including "Roses are red," "Football," and "13."

Results. In our experiments, we trained the Proteus face, voice, and fusion algorithms using videos from half of the subjects in our database (27 subjects out of a total of 54), while we considered all subjects for testing. We collected most of the training videos in controlled conditions with good lighting and low background noise levels and with the camera held directly in front of the subject's face. For these subjects, we also added a few face and voice samples from videos of less-than-ideal quality (to simulate the limited variation of training samples a typical consumer would be expected to provide) to increase the algorithm's chances of correctly identifying the user in similar conditions. Overall, we used three face frames and five voice recordings per subject (extracted from video) as training samples. We performed the testing through a randomly selected face-and-voice sample from a subject we selected randomly from among the 54 subjects in the database, leaving out the training samples. Overall, our subjects created and used 480 training and test-set combinations, and we averaged their EERs and testing times. We undertook this statistical cross-validation approach to assess and validate the effectiveness of our proposed approaches based on the available database of 54 potential subjects.

Quality-based score-level fusion. Table 1 lists the average EERs and testing times from the unimodal and multimodal schemes. We explain the high EER of our HMM voice-recognition algorithm by the complex noise signals in many of our samples, including traffic, people chatter, and music, that were difficult to detect and eliminate. Our quality-score-level fusion scheme detected low SNR levels and compensated by adjusting weights in favor of the face images that were of substantially better quality. By adjusting weights in favor of face images, the face biometric thus had a greater impact on the final decision of whether or not a user is legitimate than the voice biometric.

For the contrasting scenario, where voice samples were relatively better quality than face samples, as in Table 1, the EERs were 21.25% and 20.83% for unimodal voice and score-level fusion, respectively.

These results are promising, as they show the quality of the different modalities can vary depending on the circumstances in which mobile users might find themselves. They also show Proteus adapts to different conditions by scaling the quality weights appropriately. With further refinements (such as more robust normalization techniques), the multimodal method can yield even better accuracy.

Feature-level fusion. Table 2 outlines our performance results from the feature-level fusion scheme, showing feature-level fusion yielded significantly greater accuracy in authentication compared to unimodal schemes.

Our experiments clearly reflect the potential of multimodal biometrics to enhance the accuracy of current unimodal biometrics-based authentication on mobile devices; moreover, according to how quickly the system is able to identify a legitimate user, the Proteus approach is scalable to consumer mobile devices. This is the first attempt at implementing two types of fusion schemes on a modern consumer mobile device while tackling the practical issues of user friendliness. It is also just the beginning. We are working on improving the performance and efficiency of both fusion schemes, and the road ahead promises endless opportunity.

Back to Top

Conclusion

Multimodal biometrics is the next logical step in biometric authentication for consumer-level mobile devices. The challenge remains in making multimodal biometrics usable for consumers of mainstream mobile devices, but little work has sought to add multimodal biometrics to them. Our work is the first step in that direction.

Imagine a mobile device you can unlock through combinations of face, voice, fingerprints, ears, irises, and retinas. It reads all these biometrics in one step similar to the iPhone's TouchID fingerprint system. This user-friendly interface utilizes an underlying robust fusion logic based on biometric sample quality, maximizing the device's chance of correctly identifying its owner. Dirty fingers, poorly illuminated or loud settings, and damage to biometric sensors are no longer showstoppers; if one biometric fails, others function as backups. Hackers must now gain access to the many modalities required to unlock the device; because these are biometric modalities, they are possessed only by the legitimate owner of the device. The device also uses cancelable biometric templates, strong encryption, and the Trusted Execution Environment for securely storing and processing all biometric data.

The Proteus multimodal biometrics scheme leverages the existing capabilities of mobile device hardware (such as video recording), but mobile hardware and software are not equipped to handle more sophisticated combinations of biometrics; for example, mainstream consumer mobile devices lack sensors capable of reliably acquiring iris and retina biometrics in a consumer-friendly manner. We are thus working on designing and building a device with efficient, user-friendly, inexpensive software and hardware to support such combinations. We plan to integrate new biometrics into our current fusion schemes, develop new, more robust fusion schemes, and design user interfaces allowing the seamless, simultaneous capture of multiple biometrics. Combining a user-friendly interface with robust multimodal fusion algorithms may well mark a new era in consumer mobile device authentication.

Back to Top

References

1. Aronowitz, H., Min L., Toledo-Ronen, O., Harary, S., Geva, A., Ben-David, S., Rendel, A., Hoory, R., Ratha, N., Pankanti, S., and Nahamoo, D. Multimodal biometrics for mobile authentication. In Proceedings of the 2014 IEEE International Joint Conference on Biometrics (Clearwater, FL, Sept. 29-Oct. 2). IEEE Computer Society Press, 2014, 1–8.

2. Avila, C.S., Casanova, J.G., Ballesteros, F., Garcia, L.R.T., Gomez, M.F.A., and Sierra, D.S. State of the Art of Mobile Biometrics, Liveness and Non-Coercion Detection. Personalized Centralized Authentication System Project, Jan. 31, 2014; https://www.pcas-project.eu/images/Deliverables/PCAS-D3.1.pdf

3. Belhumeur, P.N., Hespanha, J.P., and Kriegman, D. Eigenfaces vector vs. FisherFaces: Recognition using class-specific linear projection. Pattern Analysis and Machine Intelligence, IEEE Transactions on Pattern Analysis and Machine Intelligence 19, 7 (July 1997), 711–720.

4. Bonnington, C. The trouble with Apple's Touch ID fingerprint reader. Wired (Dec. 3, 2013); http://www.wired.com/gadgetlab/2013/12/touch-id-issues-and-fixes/

5. Dalal, N. and Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (San Diego, CA, June 20–25). IEEE Computer Society Press, 2005, 886–893.

6. Daugman, J.G. Two-dimensional spectral analysis of cortical receptive field profiles. Vision Research 20, 10 (Dec. 1980), 847–856.

7. Devine, R. Face Unlock in Jelly Bean gets a 'liveness check.' AndroidCentral (June 29, 2012); http://www.androidcentral.com/face-unlock-jelly-bean-gets-liveness-check

8. Duchnowski, P., Hunke, M., Busching, D., Meier, U., and Waibel, A. Toward movement-invariant automatic lipreading and speech recognition. In Proceedings of the 1995 International Conference on Acoustics, Speech, and Signal Processing (Detroit, MI, May 9–12). IEEE Computer Society Press, 1995, 109–112.

9. Hansen, J.H.L. Analysis and compensation of speech under stress and noise for environmental robustness in speech recognition. Speech Communication 20, 1 (Nov. 1996), 151–173.

10. Hsu, D., Kakade, S.M., and Zhang, T. A spectral algorithm for learning hidden Markov models. Journal of Computer and System Sciences 78, 5 (Sept. 2012), 1460–1480.

11. Jain, A.K., Nandakumar, K., and Ross, A. Score normalization in multimodal biometric systems. Pattern Recognition 38, 12 (Dec. 2005), 2270–2285.

12. Kisku, D.R., Gupta, P., and Sing, J.K. Feature-level fusion of biometrics cues: Human identification with Doddingtons Caricature. Security Technology (2009), 157–164.

13. Kuncheva, L.I., Whitaker, C.J., Shipp, C.A., and Duin, R.P.W. Is independence good for combining classifiers? In Proceedings of the 15th International Conference on Pattern Recognition (Barcelona, Spain, Sept. 3-7). IEEE Computer Society Press, 2000, 168–171.

14. Lee, C. Automatic recognition of animal vocalizations using averaged MFCC and linear discriminant analysis. Pattern Recognition Letters 27, 2 (Jan. 2006), 93–101.

15. M2SYS Technology. SecuredPass AFIS/ABIS Immigration and Border Control System; http://www.m2sys.com/automated-fingerprint-identification-system-afis-border-control-and-border-protection/

16. Milner, B. and Xu, S. Speech reconstruction from mel-frequency cepstral coefficients using a source-filter model. In Proceedings of the INTERSPEECH Conference (Denver, CO, Sept. 16–20). International Speech Communication Association, Baixas, France, 2002.

17. Nasrollahi, K. and Moeslund, T.B. Face-quality assessment system in video sequences. In Proceedings of the Workshop on Biometrics and Identity Management (Roskilde, Denmark, May 7–9). Springer, 2008, 10–18.

18. Parala, A. UAE Airports get multimodal security. FindBiometrics Global Identity Management (Mar. 13, 2015); http://findbiometrics.com/uae-airports-get-multimodal-security-23132/

19. Rathgeb, C. and Andreas U. A survey on biometric cryptosystems and cancelable biometrics. EURASIP Journal on Information Security (Dec. 2011), 1–25.

20. Ross, A. and Govindarajan, R. Feature-level fusion of hand and face biometrics. In Proceedings of the Conference on Biometric Technology for Human Identification (Orlando, FL). International Society for Optics and Photonics, Bellingham, WA, 2005, 196–204.

21. Ross, A. and Jain, A. Multimodal biometrics: An overview. In Proceedings of the 12th European Signal Processing Conference (Sept. 6–10). IEEE Computer Society Press, 2004, 1221–1224.

22. Sacco, A. Fingerprint faceoff: Apple TouchID vs. Samsung Finger Scanner. Chief Information Officer (July 16, 2014); http://www.cio.com/article/2454883/consumer-technology/fingerprint-faceoffapple-touch-id-vs-samsung-finger-scanner.html

23. Tapellini, D.S. Phone thefts rose to 3.1 million last year. Consumer Reports finds industry solution falls short, while legislative efforts to curb theft continue. Consumer Reports (May 28, 2014); http://www.consumerreports.org/cro/news/2014/04/smartphone-thefts-rose-to-3-1-million-last-year/index.htm

24. Viola, P. and Jones, M. Rapid object detection using a boosted cascade of simple features. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Kauai, HI, Dec. 8–14). IEEE Computer Society Press, 2001.

25. Vondrasek, M. and Pollak, P. Methods for speech SNR estimation: Evaluation tool and analysis of VAD dependency. Radioengineering 14, 1 (Apr. 2005), 6–11.

26. Zorabedian, J. Samsung Galaxy S5 fingerprint reader hacked—It's the iPhone 5S all over again! Naked Security (Apr. 17, 2014); https://nakedsecurity.sophos.com/2014/04/17/samsung-galaxy-s5-fingerprint-hacked-iphone-5s-all-over-again/

Back to Top

Authors

Mikhail I. Gofman ([email protected]) is an assistant professor in the Department of Computer Science at California State University, Fullerton, and director of its Center for Cybersecurity.

Sinjini Mitra ([email protected]) is an assistant professor of information systems and decision sciences at California State University, Fullerton.

Tsu-Hsiang Kevin Cheng ([email protected]) is a Ph.D. student at Binghamton University, Binghamton, NY, and was at California State University, Fullerton, while doing the research reported in this article.

Nicholas T. Smith ([email protected]) is a software engineer in the advanced information technology department of the Boeing Company, Huntington Beach, CA.

Back to Top

Figures

F1Figure 1. Schematic diagram illustrating the Proteus quality-based score-level fusion scheme.

F2Figure 2. Linear discriminant analysis-based feature-level fusion.

F3Figure 3. The GUI used to interact with Proteus.

Back to Top

Tables

T1Table 1. EER results from score-level fusion.

T2Table 2. EER results from feature-level fusion.

Back to top


Copyright held by authors. Publication rights licensed to ACM.

The Digital Library is published by the Association for Computing Machinery. Copyright © 2016 ACM, Inc.


 

No entries found