I recently completed my PhD at UNSW Sydney, where I was advised by Dr. Lina Yao and Dr. Dong Gong.

Prior to that, I completed my Erasmus Mundus Joint Master's Degree in Advanced Systems Dependability from the University of St Andrews, UK and l'Université de Lorraine, France. During my master's, I interned with the MULTISPEECH group at Inria Nancy where I worked with Dr. Emmanuel Vincent.

My current research interests span multimodal generative models, agentic frameworks, and continual learning.

updates

  • Jun 2025: Presented my paper "Mining Your Own Secrets" at the Sydney AI meetup 🇦🇺
  • May 2025: Mentoring session + Talk on continual learning in Dr. Bishesh Khanal's group, NAAMII, Nepal 🇳🇵
  • Apr 2025: Gave a virtual talk on continual learning in Dr. Marinka Zitnik's group, Harvard University 🇺🇸
  • Apr 2025: Attending ICLR'25 in Singapore to present “Mining your Own Secrets” 🇸🇬
  • Dec 2024: Presented my paper "CLAP4CLIP" at NeurIPS 2024 in Vancouver 🇨🇦
  • Nov 2024: Won the Tiktok-sponsored Best Student Presentation award for "CLAP4CLIP" at 2024 Sydney AI meetup 🇦🇺
  • Jul 2024: Presented my paper "CLAP4CLIP" at the 2nd Bayes duality workshop, RIKEN AIP, Tokyo 🇯🇵
  • Apr 2024: Won a travel grant for presenting my paper "NPCL" at the EEML summer school, Novi Sad, Serbia 🇷🇸
  • Feb 2022: Started my PhD at UNSW Sydney 🇦🇺

Experience

AI Research Intern (Sep 2024 - Mar 2025)

LightSpeed Studios, Tencent Sydney, Australia 🇦🇺

Mentor: Dr. Shengju Qian.

Worked on controllable image generation and preference optimization for multi-modal LLMs.

Research Scientist Intern (May 2024 - Aug 2024)

Creative AI Lab, Sony Group Corporation Tokyo, Japan 🇯🇵

Mentor: Dr. Shiqi Yang

Worked on continual personalization of pre-trained text-to-image diffusion models.

Research Assistant (Sep 2021 - Jan 2022)

Computer Vision Centre, Universitat Autònoma de Barcelona Barcelona, Spain 🇪🇸

Supervisor: Dr. Joost van de Weijer

Worked on rehearsal-free continual learning (CL) for Vision Transformers (ViTs).

Research Intern (Mar 2021 - Jul 2021)

Multispeech group, Inria Nancy Nancy, France 🇫🇷

Supervisor: Dr. Emmanuel Vincent

Worked on learning domain-specific language models for speech recognition.

Machine Learning Engineer (Jun 2018 - Jul 2019)

FactSet Research Systems Inc. Hyderabad, India 🇮🇳

Mentor: Keval Dave

Worked on improving FactSet's named entity recognition service.

Awards & Recognition

Academic Services

tutoring at UNSW

Selected Publications

Mining Your Own Secrets

Mining Your Own Secrets: Diffusion Classifier Scores for Continual Personalization of Text-to-Image Diffusion Models

Saurav Jha, Shiqi Yang, Masaki Ishii, Meng Zhao, Christian Simon, Jehanzeb Mirza, Dong Gong, Lina Yao, Shusuke Takahashi, Yuki Mitsufuji

ICLR 2025

We propose using diffusion classifier scores for regularizing the parameter-space and function-space of text-to-image diffusion models, to achieve continual personalization.

CLAP4CLIP

CLAP4CLIP: Continual LeArning with Probabilistic finetuning for Vision-Language Models

Saurav Jha, Dong Gong, Lina Yao

NeurIPS 2024

Our work proposes Continual LeArning with Probabilistic finetuning (CLAP) - a probabilistic modeling frame- work over visual-guided text features per task, thus providing more calibrated CL finetuning.

NPCL

NPCL: Neural Processes for Uncertainty-Aware Continual Learning

Saurav Jha, Dong Gong, He Zhao, Lina Yao

NeurIPS 2023

We propose a neural process-based continual learning approach with task-specific modules arranged in a hierarchical latent variable model. We tailor regularizers on the learned latent distributions to alleviate forgetting.

Distilled Reverse Attention Network

Distilled Reverse Attention Network for Open-world Compositional Zero-Shot Learning

Yun Li, Zhe Liu, Saurav Jha, Lina Yao

ICCV 2023

To model attribute-object entanglement, we design a reverse-and-distill strategy that learns disentangled representations of elementary components in training data supervised by reverse attention and knowledge distillation.

Towards Exemplar-Free Continual Learning

Towards Exemplar-Free Continual Learning in Vision Transformers: an Account of Attention, Functional and Weight Regularization

Francesco Pelosin*, Saurav Jha*, Andrea Torsello, Bogdan Raducanu, Joost van de Weijer

CVPR 2022 Workshop on Continual Learning (CLVision)

We investigate the continual learning of Vision Transformers (ViT) for the challenging exemplar-free scenario, with special focus on how to efficiently distill the knowledge of its crucial self-attention mechanism.