Machine Intelligence

Machine Intelligence is a cutting-edge field that encompasses next-generation algorithms, enterprise applications, and hardware advancements in artificial intelligence (AI) research. As the forefront of AI continues to evolve rapidly, researchers in A*STAR are dedicated to pushing the boundaries of machine intelligence. This multidisciplinary field combines expertise in areas such as machine learning, deep learning, natural language processing, computer vision, and knowledge representation. By harnessing the power of these advancements, scientists strive to develop intelligent systems that can perceive, reason, learn, and make informed decisions, ultimately transforming industries and revolutionizing the way we interact with technology.

MI

Key Researchers

Dr Li Xiaoli, Institute for Infocomm Research (I2R)
Dr Rick Goh, Computing and Intelligence (IHPC)

Key Projects

Real World 3D Deep Learning

Recent advancements in 3D deep learning have led to significant progress in improving accuracy and reducing processing time, with applications spanning various domains such as medical imaging, robotics, and autonomous vehicle navigation for identifying and segmenting different structures.

 

To allow deep learning on 3D data to be adopted by companies, we addressg the difficulties in data annotation, representation learning and deployment challenges in real-world scenarios.

Self-Aware Continuous Learning Models

Learning continually for new knowledge accumulation is one of the prominent features of human intelligence. On the other hand, deep learning-based AI models are mostly static with an  assumption that all the distributions are known apriori and are pre-defined in the training data.

 

The project aims to bridge the gap by developing Self-Aware Continuous Learning Models that are continuously:

-       Aware of the historical distributions,

-       Detect unknown novel distributions,

-       Estimate the complexity of the unknown distributions,

-       Choose relevant data to represent the unknown distributions, and

-       Adapt to represent new distribution without (catastrophically) forgetting any historical distribution.

 

It also improves AI models’ generalisation capability through representational adaptations with new data. The project will build AI models that learn throughout life, to enable critical applications that warrant dynamic and up-to-date decision making, in diverse application domains (more information).


Memory-Efficient Online Continuous Object Recognition on Video Streams

The world is not stationary. To thrive in evolving environments, humans are capable of continual acquisition and transfer of new knowledge, from a continuous stream of highly correlated visual stimuli over multiple tasks, while still retaining previously-learned experiences.

 

In contrast to human learning, most AI systems focus on recognizing a pre-defined number of object classes. During training, these systems are often presented with a stationary pool of carefully shuffled, balanced, and homogenised images. In most cases, these systems struggle to recognise new objects when presented with a video stream of new object classes that are incrementally available over time. In addition, they also fail to retain good performance on earlier learned classes without repeated passes over the data, known as catastrophic forgetting.

 

Dr Zhang Mengmi and Dr Sun Ying aim to design a continual learning framework that recognises incrementally available object classes on video streams without repeated visits of the old data in a memory-efficient manner.

 

With a multi-disciplinary approach spanning neuroscience, cognitive science, and AI, we hope to gain a better understanding of the following scientific questions and provide insights about replay mechanisms in biological and artificial intelligence:

 

  1. How AIs can make use of compositional structure and representations in the memory to learn to recognize new objects faster;
  2. What levels of abstract compositional representations from the old object classes are good for replay, as well as how many times of replay required and in what sequence to replay;
  3. How far our current AI continual learning systems are from human continual learning;
  4. How future generations of continual learning algorithms can be inspired from human continual learning behaviors; and
  5. How the strategies in AI systems assist human continual learning.

More information can be found here.

Learn with Less

The amount of compute and data points used to train AI systems has increased dramatically. GPT-4 training costs over $100 million. LlaMa, a large language model developed by Meta, was trained on around one billion data points—a more than 160-million fold increase from Perceptron Mark 1, the first artificial neural network.

This has led to challenges in terms of:
    - High costs to train models due to higher volumes of data labelling required.
    - Edge devices with limited resources struggle to execute AI workloads.
    - Anomaly detection use cases may struggle with the lack of anomalies data given that those are rare events.

To improve data and computation efficiency in machine learning, we work on the following research topics:

Learn with Less Data:
    - Achieve efficient learning from a small amount of labelled data or even no labelled data. 
    - We devote to explore on semi-supervised learning/weekly supervised learning, self-supervised learning, transfer learning and unsupervised learning to reduce the data annotation requirement.

Computation efficiency: 
    - Reduce the computation cost of deep AI models, especially on edge devices. 
    - We focus on constraint-aware model compression and energy-efficient learning to compress heavy deep model into a lighter one to meet the requirement with constraints on memory and power consumption.

Physics-based AI

To describe the real world, we build models. Most modelling techniques are broken down into these 2 groups: 
    - AI-driven 
    - Physics-driven.
 
Traditional AI models are data-driven, where the models are amenable to incomplete knowledge available or complex systems. However these models are “black-boxes”, as they are unable to justify how they come to an outcome. Also their accuracy is dependent on data quality/quantity.

On the other hand, physics-driven models have no data requirements as well as have good interpretability and generalisability. This is since they are formulated based on well-defined first order principles (e.g force acting on an object = mass and the acceleration it creates). However these models are compute-intensive and their accuracy is as good as the known model.

We aim to couple traditional data-driven AI with physics-driven models, such as to leverage their complementary strengths and improve performance of target tasks.

Our research areas include: 
    - Physics-based Data Augmentation: to augment & enrich the data for AI model training in both quantity and quality.
    - Physics-based Representation Learning: to learn features with physical meanings from original data.
    - Physics-based Modelling: to build AI models with physics guidance/incorporation.