Thank you for reading the news about This AI-based system can teach human intentions to robots and now with the details
Aden - Yasmine El Tohamy - In a bid to give machines the ability to predict intent when interacting with humans, a team at the University of New South Wales (UNSW) Sydney is developing artificial intelligence-driven prototype human-machine interface system that will assist humans to be seen not merely as tools, but as partners.
Dr Lina Yao, a senior lecturer of engineering at UNSW and principal investigator, is busy getting AI systems and human-machine interfaces up to speed with the finer nuances of human behaviour.
The ultimate goal is for her research to be used in autonomous AI systems, robots and even cyborgs, but the first step is focused on the interface between humans and intelligent machines.
"What we're doing in these early phases is to help machines learn to act like humans based on our daily interactions and the actions that are influenced by our own judgment and expectations - so that they can be better placed to predict our intentions," Yao said in a university statement.
At the moment, AI may do a plausible job at detecting the intent of another person (in other words, after the fact).
It may even have a list of predefined, possible responses that a human will respond within a given situation. But when an AI system or machine only has a few clues or partial observations to go on, its responses can sometimes be a little robotic.
Dr Yao is working on less obvious examples of human behaviour integrated into AI systems to improve intent prediction.
Things like gestures, eye movement, posture, facial expression and even micro-expressions - the tell-tale physical signs when someone reacts emotionally to a stimulus but tries to keep it hidden.
"We can learn and predict what a human would like to do when they're wearing an EEG [electroencephalogram] device," said Dr Yao.
While wearing one of these devices, whenever the person makes a movement, their brainwaves are collected which researchers can then analyse.
"Later we can ask people to think about moving with a particular action - such as raising their right arm. So not actually raising the arm, but thinking about it, and we can then collect the associated brain waves".
Recording this data has the potential to help people unable to move or communicate freely due to disability or illness.
Brain waves recorded with an EEG device could be analysed and used to move machinery such as a wheelchair, or even to communicate a request for assistance.
According to Yao, autonomous AI systems and machines may one day look at us as belonging to one of three categories after observing our behaviour - peer, bystander or competitor.
"While this may seem cold and aloof, these categories may dynamically change from one to another according to their evolving contexts.
At any rate, she said, this sort of cognitive categorisation is actually very human.
These were the details of the news This AI-based system can teach human intentions to robots for this day. We hope that we have succeeded by giving you the full details and information. To follow all our news, you can subscribe to the alerts system or to one of our different systems to provide you with all that is new.
It is also worth noting that the original news has been published and is available at Khaleej Times and the editorial team at AlKhaleej Today has confirmed it and it has been modified, and it may have been completely transferred or quoted from it and you can read and follow this news from its main source.