Highlights

In brief

Integrating Double Oracle Algorithm and Neural Architecture Search into deep learning frameworks enables diverse outputs during training, leading to models that can produce more realistic images.

Photo by Freepik

When AI plays to learn

27 Mar 2026

A new framework to train artificial intelligence models draws on concepts from game theory to improve model stability and performance.

If you were shown two photos side by side, could you tell which one was generated by artificial intelligence (AI) and which was real? A few years ago, that would have been easy. Today, much less so. Deep learning models—a subtype of AI—have become remarkably adept at creating realistic images by repeatedly playing, and learning from, this game of spotting the fake.

This training method is called Generative Adversarial Network (GAN). It involves two neural networks locked in competition: a generator and a discriminator. “The generator tries to create realistic images, while the discriminator tries to spot the fakes,” explained Aye Phyu Phyu Aung, a Scientist at the A*STAR Institute for Infocomm Research (A*STAR I2R). “They keep improving their tactics until neither can easily win.”

However, GAN may run into issues of mode collapse, where the generator and discriminator become trapped in a narrow set of strategies. The result is subpar, repetitive outputs that make for an ineffective training regime for deep learning models.

To mitigate this, Aung and A*STAR I2R colleagues, including Senior Principal Scientist Xiaoli Li and Senior Scientist J. Senthilnath, teamed up with collaborators from Nanyang Technological University, Singapore; Singapore Management University; KTH Royal Institute of Technology, Sweden; and University of Nebraska–Lincoln, US. As generator/discriminator pairs act like opponents in a game, the researchers believed that adopting game theory principles could be the key to improving GAN.

One such concept, the Double Oracle (DO) algorithm, starts with a smaller version of the fake-spotting game instead of determining the best strategies for the whole game from the get-go. “DO solves a small restricted game, asks each player’s best response to find a better strategy, adds those strategies, and repeats until no improvement is possible,” said Aung.

The team further complemented DO with Neural Architecture Search (NAS), melding them into a framework dubbed DONAS. Scouring through a variety of player architectures, NAS identifies those that best match the optimal strategies determined by DO—much like selecting athletes with skillsets and playstyles that align with a coach’s tactical vision.

Testing revealed that DONAS effectively enhanced GAN’s performance, making it more robust against mode collapse. “We are able to get vastly different models, generating samples of diverse features and patterns,” said Aung. “Moreover, the trained models could create realistic images resembling those in a given dataset, outperforming other GAN approaches across several benchmarks.”

The researchers also observed similar improvements when they applied DONAS to another framework, which uses classifier/attacker pairs rather than generator/discriminator to analyse imaging datasets. Aung and the team have since continued to develop more robust and effective AI training frameworks, including a recently patented GAN-based module.

The A*STAR-affiliated researchers contributing to this research are from the A*STAR Institute for Infocomm Research (A*STAR I2R).

Want to stay up to date with breakthroughs from A*STAR? Follow us on Twitter and LinkedIn!

References

Aung, A.P.P., Wang, X., Wang, R., Chan, H., An, B., et al. Double oracle neural architecture search for game theoretic deep learning models. IEEE Transactions on Image Processing 34, 2463–2472 (2025). | article

About the Researchers

Aye Phyu Phyu Aung earned her PhD degree in Computer Science from Nanyang Technological University, Singapore in 2023, with seminal works published in CVPR, ICAPS and ICIP. Her research interests include generative models, computational game theory, reinforcement learning and optimisation. Affiliated with the A*STAR Institute of Infocomm Research (A*STAR I2R), her collaborative research aims to bridge academia and real-world tech solutions.
Xiaoli Li was formerly a Senior Principal Scientist and the Head of Machine Intellection at the A*STAR Institute for Infocomm Research (A*STAR I2R). He was Co-director of the KPMG-A*STAR joint lab and previously, an Adjunct Professor at Nanyang Technological University, Singapore. Li has published more than 200 papers with more than 10,000 citations, including several award-winning publications in the fields of data analytics and artificial intelligence. Li has led over ten research projects in collaboration with industry partners across a range of sectors. He actively serves as an organiser and contributor to top global AI and data analytics conferences, including AAAI, IJCAI, KDD and ICDM.
Senthilnath J. is a Senior Scientist at the A*STAR Institute for Infocomm Research (A*STAR I2R). He obtained his PhD degree in Aerospace Engineering from the Indian Institute of Science (IISc), India. His current research focuses on artificial intelligence (AI), multi-agent systems, generative models, reinforcement learning and optimisation. He has authored over 100 high-quality papers and won five best paper awards. He has also been recognised among the world's top 2% of scientists in AI by Stanford University. Senthilnath is a Senior Member of IEEE and has served in various leadership roles for leading AI and data analytics conferences, including Organising Chair, Co-Chair, Senior Program Committee Member and Program Committee Member.

This article was made for A*STAR Research by Wildtype Media Group