Researchers at Oxford University’s Department of Materials, working in collaboration with colleagues from Exeter and Munster, have developed an on-chip optical processor capable of detecting similarities in datasets up to 1,000 times faster than conventional machine learning algorithms running on electronic processors.
The new research published in Optica took its inspiration from Nobel Prize laureate Ivan Pavlov’s discovery of classical conditioning. In his experiments, Pavlov found that by providing another stimulus during feeding, such as the sound of a bell or metronome, his dogs began to link the two experiences and would salivate at the sound alone. The repeated associations of two unrelated events paired together could produce a learned response—a conditional reflex.
Co-first author Dr. James Tan You Sian, who did this work as part of his DPhil in the Department of Materials, University of Oxford, said, “Pavlovian associative learning is regarded as a basic form of learning that shapes the behavior of humans and animals—but adoption in AI systems is largely unheard of. Our research on Pavlovian learning in tandem with optical parallel processing demonstrates the exciting potential for a variety of AI tasks.”
The neural networks used in most AI systems often require a substantial number of data examples during a learning process—training a model to reliably recognize a cat could use up to 10,000 cat/non-cat images—at a computational and processing cost.
Rather than relying on backpropagation favored by neural networks to “fine-tune” results, the Associative Monadic Learning Element (AMLE) uses a memory material that learns patterns to associate together similar features in datasets—mimicking the conditional reflex observed by Pavlov in the case of a “match.”
The AMLE inputs are paired with the correct outputs to supervise the learning process, and the memory material can be reset using light signals. In testing, the AMLE could correctly identify cat/non-cat images after being trained with just five pairs of images.
The considerable performance capabilities of the new optical chip over a conventional electronic chip are down to two key differences in design:
A unique network architecture incorporating associative learning as a building block rather than using neurons and a neural network
The use of “wavelength-division multiplexing” to send multiple optical signals on different wavelengths on a single channel to increase computational speed.
The chip hardware uses light to send and retrieve data to maximize information density—several signals on different wavelengths are sent simultaneously for parallel processing which increases the detection speed of recognition tasks. Each wavelength increases the computational speed.
Professor Wolfram Pernice, co-author from Münster University, explained, “The device naturally captures similarities in datasets while doing so in parallel using light to increase the overall computation speed—which can far exceed the capabilities of conventional electronic chips.”
An associative learning approach could complement neuralnetworks rather than replace them, clarified co-first author Professor Zengguang Cheng, now at Fudan University.
“It is more efficient for problems that don’t need substantial analysis of highly complex features in the datasets,” said Professor Cheng. “Many learning tasks are volume-based and don’t have that level of complexity—in these cases, associative learning can complete the tasks more quickly and at a lower computational cost.”
“It is increasingly evident that AI will be at the center of many innovations we will witness in the coming phase of human history. This work paves the way towards realizing fast optical processors that capture data associations for particular types of AI computations, although there are still many exciting challenges ahead,” said Professor Harish Bhaskaran, who led the study. https://techxplore.com/news/2022-07-ai-associatively.html
Recent Comments