“I believe it’s just the beginning of a different style of brain-inspired computation,” Bengio said, adding, “I think we have a lot of the tools to get started.Montréal Institute of Learning Algorithms Though the ability of neural networks to parallel-process on a large scale has given us breakthroughs in computer vision, translation, and memory, research is now shifting to developing novel deep architectures and training frameworks for addressing tasks like reasoning, planning, capturing causality, and obtaining systematic generalization. The next challenge, they say, is to actually use the tools available in CausalWorld to build more generalizable systems.ĭespite how dazzled we are by AI’s ability to perform certain tasks, Yoshua Bengio, in 2019, estimated that present-day deep learning is less intelligent than a two-year-old child. In other words, users are free to intervene on a large number of variables in the environment, and thus draw systemic conclusions about what the agent generalizes to-or doesn’t. “Having a model architecture or method that can learn these underlying rules or causal mechanisms, and utilize them could overcome these challenges,” Träuble says.ĬausalWorld’s evaluation protocols, say Ahmed and Träuble, are more versatile than those in previous studies because of the possibility of “disentangling” generalization abilities. Träuble adds that “What’s actually interesting is that we humans can generalize much, much quicker we don’t need such a vast amount of experience… We can learn from the underlying shared rules of environments… use this to generalize better to yet other environments that we haven’t seen.”Ī standard neural network, on the other hand, would require insane amounts of experience with myriad environments in order to do the same. “If we continue scaling up training and network architectures beyond the experiments we report, current methods could potentially solve more of the block stacking environments we propose with CausalWorld,” points out Frederik Träuble, one of the contributors to the study. They observed that as the curricula got more complex, the agents showed less ability to transfer their skills to the new conditions. The first involved no environment changes the second had changes to a single variable and the third allowed full randomization of all variables in the environment. In their study, the researchers gave the robots a number of tasks ranging from simple to extremely challenging, based on three different curricula. There is a large set of parameters, such as weight, shape, and appearance of the blocks and the robot itself, on which the user can intervene at any point to evaluate the robot’s generalization capabilities. Robotic agents can be given tasks that comprise pushing, stacking, placing, and so on, informed by how children have been observed to play with blocks and learn to build complex structures. The main purpose of CausalWorld is to accelerate research in causal structure and transfer learning using this simulated environment, where learned skills could potentially be transferred to the real world. The paper on CausalWorld, available as a preprint, describes benchmarks in a simulated robotics manipulation environment using the open-source TriFinger robotics platform. The group’s tool, called CausalWorld, demonstrates that with some of the methods currently available, the generalization capabilities of robots aren’t good enough-at least not to the extent that “we can deploy safely in any arbitrary situation in the real world,” says Ahmed. One of the reasons is that the physical properties of the simulation are quite different from the real world,” says Ahmed. “Robots are trained in simulation, and then when you try to deploy in the real world…they usually fail to transfer their learned skills. The lack of generalization is a big problem, says Ossama Ahmed, a master’s student at ETH Zurich who has worked with Bengio’s team to develop a robotic benchmarking tool for causality and transfer learning. This, in turn, limits AI from being able to generalize their learning and transfer their skills to another related environment. Nor are they particularly good at tasks that involve imagination, reasoning, and planning. However, neural nets do not interpret cause-and effect, or why these associations and correlations exist. So far, deep learning has comprised learning from static datasets, which makes AI really good at tasks related to correlations and associations. “Causality is very important for the next steps of progress of machine learning,” said Yoshua Bengio, a Turing Award-wining scientist known for his work in deep learning, in an interview with IEEE Spectrum in 2019.
0 Comments
Leave a Reply. |
Details
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |