Experiences at school

A good system of evaluation is not just to criticize what teachers are doing wrong, I mean, is better to see areas where teachers and administrative department need to improve. It is very similar as…

Smartphone

独家优惠奖金 100% 高达 1 BTC + 180 免费旋转




Continuous Learning

The standard machine learning approach is to learn to accomplish a specific task with an associated dataset. A model is trained using the dataset and is only able to perform that one task. This is in stark contrast to animals which continue to learn throughout life and accumulate and re-purpose knowledge and skills. The limitation has been widely acknowledged and addressed in different ways, and with a variety of terminology, which can be confusing. I wanted to take a brief look at those approaches and to create a precise definition of the Continuous Learning that we want to implement in our pursuit of AGI.

One type of Transfer Learning is Domain Adaptation. It refers to the idea of learning on one domain, or data distribution, and then applying the model to and optimising it for a related data distribution. Training a model on different data distributions is often referred to as Multi Domain Learning. In some cases the distributions are similar, but other times they are deliberately unrelated.

Online Learning algorithms learn iteratively with new data, in contrast to learning from a pass of a whole dataset, as is commonly done in conventional supervised and unsupervised learning, referred to as Batch Learning. Batches can also refer to portions of the dataset.

Online Learning is useful when the whole dataset does not fit into memory at once. Or more relevant for AGI, in scenarios where new data is observed over time. For example, with new samples being generated by users of a system, by an agent exploring its environment or for cases where the phenomena being modelled changes. Another way to describe it is that the underlying input data distribution is not static i.e. a non-stationary distribution, hence these are referred to as Non-stationary Problems.

Online learning systems can be susceptible to ‘forgetting’. That is, becoming less effective at modelling older data. The worst case is failing completely and suddenly, known as Catastrophic Forgetting or Catastrophic Interference.

Incremental Learning, as the name suggests, is about learning bit by bit, extending the model and improving performance over time. Incremental Learning explicitly handles the level of forgetting of past data. In this way, it is a type of online learning that avoids catastrophic forgetting.

In One-shot Learning, the algorithm is able to learn from one or very few examples. Instance Learning is one way of achieving that, constructing hypotheses from the training instances directly.

A related concept is Multi-Modal Learning, where a model is trained on different types of data for the same task. An example is learning to classify letters from the way they look with visual data, and the way they sound, with audio.

Now that we have some greater clarity around these terms, we recognise that they are all important features of what we consider to be Continuous Learning for a successful AGI agent. I think it’s instructive to express it in terms of traits in the context of an autonomous agent. I’ve mapped these traits to the associated Machine Learning algorithm concepts, in brackets.

Uses learnt information to help with subsequent tasks. Builds on its knowledge. Enables more complex behaviour and faster learning. (Transfer Learning, Curriculum Learning)

As features of the task change gradually, it will adapt. This will not cause catastrophic forgetting. (Domain Adaption, Non-stationary input distributions, Iterative Learning)

Can learn entirely new tasks. This will not cause catastrophic forgetting of old tasks. Also, it can learn these new tasks as well as it would have, if it was the first task learnt i.e. learning a task does not impede the ability to learn subsequent tasks. (Iterative Learning)

Learns important aspects of the task from very few examples. It has the ability to learn fast when necessary. (One-shot Learning)

Continues to learn as it collects more data. (Online Learning)

Combines sensory modalities to learn a task. (Multi-modal Learning)

Note that in continuous learning, if there are fixed resources, and you are operating at your limit, then there has to be some forgetting, but as mentioned in the table, it should not be ‘catastrophic forgetting’.

Add a comment

Related posts:

Inheritance Right In Islam Family Law

Today out of five people, at least one belongs to the Islamic faith. Besides, people in common belong to other religions have obviously heard about the religion but they may not have more idea about…

Showing up against hate

Today our family went to a rally against hate to show our support against the white supremacists who threatened to march in Berkeley, California. The event was advertised as a peaceful gathering but…

Inside the Mind of a Bon Viveur

A fiction short story about a girl's journey into the world of toxicity, drugs and partying and society's stereotypes about women and girls who are open about their sexuality or preferences.