Skip navigation links

November

08

3575 Engineering Building and Zoom

Doctoral Defense - Shivangi Yadav

Register
the famous Belmont tower facing a sunset

About the Event

The Department of Computer Science & Engineering

Michigan State University

Ph.D. Dissertation Defense

 

November 8th, 2024 at 10:30am EST

3575 Engineering Building and Zoom

Zoom Information: Upon Request from Vincent Mattison or Advisor

 

Synthesizing Iris and Ocular Images using Adversarial Networks and Diffusion Models

By: Shivangi Yadav

Advisor: Dr. Arun Ross

 

Synthetic biometric data – such as fingerprints, face, iris and speech – can overcome some of the limitations associated with the use of real data in biometric systems. The focus of this work is on the iris biometric. Current methods for generating synthetic irides and ocular images have limitations in terms of quality, realism, intra-class diversity and uniqueness. Different methods are proposed in this thesis to overcome these issues while evaluating the utility of synthetic data for two biometric tasks: iris matching and presentation attack (PA) detection.

Two types of synthetic iris images are generated: (1) partially synthetic and (2) fully synthetic. The goal of “partial synthesis” is to introduce controlled variations in real data. This can be particularly useful in scenarios where real data are limited, imbalanced, or lack specific variations. We present three different techniques to generate partially synthetic iris data: one that leverages the classical Relativistic Average Standard Generative Adversarial Network (RaSGAN), a novel Cyclic Image Translation Generative Adversarial Network (CIT-GAN) and a novel Multi-domain Image Translative Diffusion StyleGAN (MID-StyleGAN). While RaSGAN can generate realistic looking iris images, this method is not scalable to multiple domains (such as generating different types of PAs). To overcome this limitation, we propose CIT-GAN that generates iris images using multi-domain style transfer. To further address the issue of quality imbalance across different domains, we develop MID-StyleGAN that exploits the stable and superior generative power of diffusion based StyleGAN. The goal of “full synthesis” is to generate iris images with both inter and intra-class variations. In this regard, we propose two novel architectures, viz., iWarpGAN and IT-diffGAN. The proposed iWarpGAN focuses on generating iris images that are different from the identities in the training data using two transformation pathways: (1) Identity Transformation and (2) Style Transformation. On the other hand, IT-diffGAN projects input images onto the latent space of a diffusion GAN, identifying and manipulating the features most relevant to identity and style. By adjusting these features in the latent space, IT-diffGAN generates new identities while preserving image realism.

A number of experiments are conducted using multiple iris and ocular datasets in order to evaluate the quality, realism, uniqueness, and utility of the synthetic images generated using the aforementioned techniques. An extensive analysis conveys the benefits and the limitations of each technique. In summary, this thesis advances the state of the art in iris and ocular synthesis by leveraging the prowess of GANs and Diffusion Models.

 

Tags

Doctoral Defenses

Date

Friday, November 08, 2024

Time

10:30 AM

Location

3575 Engineering Building and Zoom

Organizer

Shivangi Yadav