2025.12.02 Tuesday
Approximate Domain Unlearning: Enabling Safer and More Controllable Vision-Language Models
The proposed approximate unlearning technique differentiates between domains while preserving the generalization capability
Vision-language model (VLM) is a core technology of modern artificial intelligence (AI), and it can be used to represent different forms of expression or learning, such as photographs, illustrations, and sketches. It has high generalization ability, which allows it to accurately recognize objects in images within a domain. However, this generalization ability is at risk. For example, VLM recognizes both real cars and illustrated cars as cars. If this is installed in a system, there is a risk that a car illustrated in a roadside advertisement may be mistaken for a real vehicle, leading to a serious accident. To put safe and reliable AI into practical use, it is essential to establish technology that appropriately controls learned knowledge according to the application.
Addressing this issue, a team of researchers, led by Associate Professor Go Irie from Tokyo University of Science, has proposed the approximate domain unlearning (ADU) algorithm, which allows VLM to forget specific domains so that it cannot recognize them. For example, according to these results, real-life vehicles can be recognized with high accuracy.
This innovative study, co-authored by Mr. Kodai Kawamura and Mr. Yuta Goto from Tokyo University of Science, Japan, Dr. Rintaro Yanagi from National Institute of Advanced Industrial Science and Technology, and Dr. Hirokatsu Kataoka from AIST and University of Oxford, has been presented at the 39th Conference on Neural Information Processing Systems (NeurIPS 2025), held from November 30 to December 7, 2025.
AI technologies have long pursued the ability to recognize objects accurately across all domains, as seen in decades of research on domain adaptation and domain generalization. While such versatility remains important, the emergence of VLMs with remarkable domain generalization ability made us realize that this assumption itself deserves to be re-examined. This idea led us to conceive ADU--a new approach that deliberately enables models to forget specified domains, when necessary, explains Prof. Irie.
Notably, the technical difficulty is that the domains cannot be distinguished within the VLM. As different domains overlap in the feature space, it is difficult to select and forget only specific domains. Therefore, in this research, the team introduced a method called Domain Disentangling Loss that promotes separation between domains in the feature space and captures the different domain appearances in each image. Furthermore, by introducing the instance-wise prompt generator, the proposed algorithm reduces the recognition accuracy for unnecessary domains while minimizing the need for them. This allows for flexible AI configuration suited to individual practical scenarios and enables flexible knowledge control that was previously impossible, such as making it impossible to recognize cars in illustrations while maintaining functionality.
Interestingly, ADU introduces a new perspective on risk management. While the concept of managing AI risks has long existed, the very generalization ability of modern AI models can sometimes generate new risks. This research presents a framework for building AI that can be flexibly configured according to individual usage scenarios, ensuring both safety and adaptability.
While AI performance is becoming more sophisticated, in order to promote sustainable industrial applications, it is necessary to adapt it to practical scenarios. We feel that our system, which allows us to freely control functions, will enable us to provide the world with safe and reliable AI technology, concludes Prof. Irie.
Image title: High-Performance Approximate Domain Unlearning
Image caption: First proposal and realization of approximate domain unlearning--selective forgetting at the domain level for vision-language models.
Image credit: Associate Professor Go Irie from Tokyo University of Science
License type: Original content
Usage restrictions: Cannot be reused without permission.
Reference
| Title of original paper | : | Approximate Domain Unlearning for Vision-Language Models |
| Conference: | : | Neural Information Processing Systems (NeurIPS 2025) |
About The Tokyo University of Science
Tokyo University of Science (TUS) is a well-known and respected university, and the largest science-specialized private research university in Japan, with four campuses in central Tokyo and its suburbs and in Hokkaido. Established in 1881, the university has continually contributed to Japan's development in science through inculcating the love for science in researchers, technicians, and educators.
With a mission of "Creating science and technology for the harmonious development of nature, human beings, and society," TUS has undertaken a wide range of research from basic to applied science. TUS has embraced a multidisciplinary approach to research and undertaken intensive study in some of today's most vital fields. TUS is a meritocracy where the best in science is recognized and nurtured. It is the only private university in Japan that has produced a Nobel Prize winner and the only private university in Asia to produce Nobel Prize winners within the natural sciences field.
■
Tokyo University of Science(About TUS)
About Associate Professor Go Irie
from Tokyo University of Science
Dr. Go Irie graduated in Engineering from Keio University in 2004 and obtained his PhD from The University of Tokyo in 2011. He specializes in pattern recognition, machine learning, and media understanding, with a strong focus on perceptual information processing. He has participated in numerous international conferences on these topics, including prestigious ones such as NeurIPS, CVPR, and ACM Multimedia.
Laboratory website 
Official TUS website
Funding information
This project is partially supported by JPNP25006, commissioned by the New Energy and Industrial Technology Development Organization (NEDO).

