In a paper, published last week, a member from the US Air Force talks about a model for artificial general intelligence (AGI). The author of the paper, A Model for General Intelligence, is Paul Yaworsky, Information Directorate of the US Air Force Research Laboratory. There have been many efforts in the past to model intelligence in machines, but with little progress in terms of real cognitive intelligence like those of humans.
Currently, the way AI systems work is not understood completely. Also, AI systems are good at performing narrow tasks but not complex cognitive problems. Artificial general intelligence aims to covers the gap between lower level and higher level work in AI—to try and make sense of the abstract general nature of intelligence. Three basic aspects of artificial intelligence need to be understood to bridge this gap.
The paper proposes a hierarchical model to help capture and exploit the order within intelligence. The underlying order contains patterns of signals that become organized, stored and then activated in space and time.
The paper portrays intelligence as an orderly, organized process using a simple hierarchy as shown:
Source: A Model for General Intelligence
The real world has order and organization. The human brain understands this and forms an internal model based on that understanding. This model enables learning, which further gives way to decision making, movement, and communication. The flow of input signals and learning within the shown model is bottom-up which is in contrast to the top-down of learned signal representations.
The paper says that external order and organization can be modeled internally in the brain in the form of various hierarchies. The hierarchies discussed are temporal, spatial, and general.
When computers continue to improve and cooperation increases between humans and computers, people themselves will become more productive in information processing. A point to remember as the paper states is that computers work for humans.
Yaworsky also talks about concerns with AI and it taking over the world. Problems like those are heard today due to sketchy predictions involving intelligence he says. It is difficult to make good scientific predictions in itself but when the predictions have to be done in intelligence it is almost impossible to get them right.
This is because a proper understanding of intelligence itself is not good enough to be able to make accurate predictions.
Do you buy this explanation or fear the US Air Force working on killer drones that may one day go rampant like in Terminator 2?!
Either way, the conclusion is that intelligence involves multiple levels of abstraction. Human intelligence has high processing levels—abstract, general, etc. Majority of the work done in AI currently is at lower levels of abstraction. There is a lot needed for the current AI to become real AI. A high-level hierarchical model for artificial intelligence as explored in the paper addresses this.
For more details, you can read the research paper.
The ethical dilemmas developers working on Artificial Intelligence products must consider
Technical and hidden debts in machine learning – Google engineers’ give their perspective
Sex robots, artificial intelligence, and ethics: How desire shapes and is shaped by algorithms