• A
  • A
  • A
  • ABC
  • ABC
  • ABC
  • А
  • А
  • А
  • А
  • А
Regular version of the site

How Modern Robots Are Developed

How Modern Robots Are Developed

© iStock

Today, neuroscience and robotics are developing hand in hand. Mikhail Lebedev, Academic Supervisor at HSE University’s Centre for Bioelectric Interfaces, spoke about how studying the brain inspires the development of robots.

Robots are interesting to neuroscience and neuroscience is interesting to robots —this is what our article ‘Neuroengineering challenges of fusing robotics and neuroscience’ was about in the journal Science Robotics. Such collaborative development contributes to progress in both fields, bringing us closer to developing more advanced android robots and a deeper understanding of the structure of the human brain. And, to some extent, to combining biological organisms with machines, to create cybernetic organisms (cyborgs).

Neuroscience for Robots

Robots often resemble humans in their make-up. This is true for robots that are meant to mimic human actions and behaviour — neuroscience is less important to industrial machines.

The most obvious thing to use in the design of a robot is to make it look human. Robots often have two arms, two legs and a head, even if it is not necessary from an engineering point of view. This is especially important when the robot will interact with people — a machine that looks like us is easier to trust.

It is possible to ensure that not only the appearance, but also the ‘brain’ of the robot resembles that of a human. In developing the mechanisms for perception, information processing, and control, engineers are inspired by the structure of the human nervous system.

For example, a robot's eyes — TV cameras that can move on different axes — mimic the human visual system. Based on the knowledge of how human vision is structured and how the visual signal is processed, engineers design the robot's sensors according to the same principles. In this way the robot can be endowed with the human ability to see the world in three dimensions, for example.

Humans have a vestibulo-ocular reflex: the eyes apply stabilisation using vestibular information when we move,  allowing us to maintain the stability of the picture we see. There may also be acceleration and orientation sensors on a robot's body. These help the robot to take body movements into account to stabilise the visual perception of the outside world and improve agility.

In addition, a robot can experience the sense of touch just like a human — a robot can have skin, it can feel touched. And then it doesn't just move randomly in space: if it touches an obstacle, it senses it and reacts to it just like a human does. It can also use this artificial tactile information to grip objects.

© iStock

Robots can even simulate sensations of pain: some forms of physical contact feel normal and some cause pain, which drastically changes the robot's behaviour. It starts to avoid pain and develop new behaviour patterns, i.e. it learns — like a child who has been burned by something hot for the first time.

Not only sensory systems, but also a robot's body control can be designed analogously to that of humans. In humans, walking is controlled by so-called central rhythm generators — specialized nerve cells designed to control autonomous motor activity. There are robots in which the same idea is used to control walking.

In addition, robots can learn from humans. A robot can perform actions in an infinite number of ways, but if it wants to mimic a human, it must observe the human and try to repeat their movements. When it makes mistakes, it compares itself with how a human performs the same action.

Robots for Neuroscience

How can neuroscience use robots? When we build a model of a biological system, we begin to better understand the principles by which it works. Therefore, developing mechanical and computer models of human nervous system movement control brings us closer to understanding neurological functions and biomechanics.

And the most promising area of using robots in modern neuroscience is in designing neurointerfaces — systems for controlling external devices using brain signals. Neurointerfaces are necessary for the development of neuroprostheses (for example, an artificial arm for people who have lost a limb) and exoskeletons — external frames or skeletons for a human body to increase its strength or restore lost motor ability.

A robot can interact with the nervous system through a bi-directional interface: the nervous system can send a command signal to the robot, and the robot from its sensors can return sensory information to the human, causing real sensations by stimulating nerves, nerve endings in the skin, or the sensory cortex itself. Such feedback mechanisms make it possible to restore the sensation of a limb if it has been lost. They are also necessary for more precise movements of the robotic limb, since it is on the basis of sensory information received from the arms and legs that we correct our movements.

There is an interesting question that arises here: Should we control all degrees of freedom of the robot through a neural interface? In other words, how should we send specific commands to it? For example, we can ‘order’ the robotic arm to pick up a bottle of water, and it will perform specific operations: it will lower its arm, turn it, and unclench and clench the fingers on its hand — all on its own. This approach is called combined control — we give simple commands through a neural interface, and a special controller inside the robot selects the best strategy for implementation. Or we can create a mechanism that will not understand the ‘take the bottle’ command: it needs to be sent information about specific, detailed movements.

Current Studies

Neuroscientists and robotic scientists study various aspects of brain operation and robotic devices. For example, at Duke University I conducted experiments with neural interfaces on monkeys, since interfaces need to be directly connected to brain areas for them to work accurately and such experimental interventions are not always possible to be performed on humans.

In one of my studies, a monkey walked along a path and the activity of its motor cortex, which is responsible for leg movement, was read and triggered a robot to start walking. At the same time, the monkey observed this walking robot on a screen placed in front of it.

The monkey used feedback, so it corrected its movements based on what it saw on the screen. This is how the most effective neural interfaces for implementing walking are developed.

The Cybernetic Future

Such research leads us to innovative developments in the future. For example, creating an exoskeleton to restore the movements of completely paralysed people no longer seems like an unattainable fantasy — it just takes time. Progress may be held back by the lack of computer power, but development over the past ten years has also been enormous here. It is likely that we will soon see people around us using light, comfortable exoskeletons rather than wheelchairs or strollers to get around. Human cyborgs will become commonplace.

Commercial development of such systems is taking place all over the world, including in Russia. For example, the famous ExoAtlet project is developing exoskeletons for the rehabilitation of people with motor disabilities. The HSE Centre for Bioelectric Interfaces participated in the development of algorithms for these machines: Centre Head, Professor Alexey Ossadtchi, and his doctoral students developed a neurointerface that triggers the walking movements of the exoskeleton.

Nikolay Smetanin, Junior Research Fellow of HSE Centre for Bioelectrical Interfaces, debugs the ideomotor neurointerface module for controlling the exoskeleton of the lower limbs of ExoAtlet
© HSE University

The rapid development of humanoid robots is also becoming a reality. It is likely that we will soon have robots walking around imitating us in many respects —moving like us and thinking like us. They will be able to do some of the work previously available only to humans.

Obviously, we will see the development of both robotics and neuroscience, and these fields will converge. This not only opens up new opportunities, but also creates new ethical questions, such as how we should treat android robots or human cyborgs.

And yet, so far, humans are better than robots in many ways. Our muscles are the most economical: eat a sandwich and you will have enough energy for the whole day. The robot will have a flat battery in half an hour. And although it may be much more powerful than a human, it is often too heavy. When it comes to elegance and energy capacity optimisation — so far, a human is still superior to a robot.

It is not far into the future when this will change — there are tens of thousands of talented scientists and engineers working towards this goal.

Interviewed by Anastasia Lobanova

See also:

Cognitive Reappraisal of Negative Emotions Can Help Manage Stress

Researchers at the HSE International Laboratory of Social Neurobiology assessed the effectiveness of two strategies for regulating emotions: reappraisal and suppression. Having analysed data on the electrical activity of 60 individuals’ brains, the scientists discovered that both approaches put additional strain on the nervous system. It was also found that individuals who are prone to emotional contagion tend to be more effective in using reappraisal and managing negative emotions. The paper has been published in Experimental Brain Research.

Russian Researchers Unveil Mechanism Underlying Language Processing Disruptions in Epilepsy Patients

Researchers at HSE University and the Pirogov National Medical and Surgical Centre have examined alterations induced by epilepsy in the language-related neural network within the brain. Using graph-based analysis, the researchers studied fMRI data from 28 patients and found that in epilepsy, both hemispheres of the brain become activated during language processing and short connections form between the hemispheres, while long connections within one hemisphere are disrupted. The study has been published in Epilepsy&Behavior.

HSE Creates ‘Transfer of Neurocognitive Technologies’ Consortium

HSE, the Pirogov National Medical and Surgical Centre, and the Centre for Speech Pathology and Neurorehabilitation of the Moscow Healthcare Department have signed an agreement on cooperation and the creation of a ‘neuro-consortium’ under the name ‘Transfer of Neurocognitive Technologies’. The new body will boost the development and implementation of advanced solutions in neurotechnology aimed at maintaining and improving people's health. The agreement was signed for five years, and the consortium is open to new participants.

'While it May Sound Futuristic, It Holds Great Promise': Olga Dragoy Shares Her Thoughts on Language Function Restoration and the Future of Neurotechnology

In the spring of 2023, the fifth strategic project of the Priority 2030 programme, 'Human Brain Resilience: Neurocognitive Technologies for Adaptation, Learning, Development and Rehabilitation in a Changing Environment,' was launched at HSE University. The strategic project brings together researchers from all campuses of HSE University. In her interview with the HSE News Service, Olga Dragoy, head of the strategic project and Director of the HSE Centre for Language and Brain, shares an overview of the advanced technologies neuroscientists are creating today, the underlying inspiration driving these efforts, and the operational dynamics of interdisciplinary applied projects.

‘It Was Great to Look at Scientific Achievements through the Eyes of a Journalist, not a Scientist’

HSE University in Nizhny recently hosted the 2nd Autumn Neuro-linguistic School ‘NeuroSciCom: Popularising Language and Brain Studies’ for scientists and students at the HSE Centre for Language and Brain Studies in Nizhny Novgorod. The school was held as part of the 'Human Brain Resilience: Neurocognitive Technologies for Adaptation, Learning, Development and Rehabilitation in a Changing Environment' Strategic Project of the Priority 2030 programme.

The Brain Is a Network of Networks. Scientists Have Found a Way to Unravel Them

A team of researchers from HSE University and the Artificial Intelligence Research Institute (AIRI) have demonstrated the effectiveness of the PSIICOS method they had previously developed for non-invasive mapping the neural networks in the brain based on its electrical activity. Unlike other methods, it does not search for individual neuronal sources to be then combined into networks but instead looks directly for the functional networks of interconnected neuronal populations—and does so swiftly and accurately. The study findings have been published in NeuroImage.

HSE Neuroscientists Use Neural Network to Enhance Neurofeedback Technology

Researchers from HSE University and the Artificial Intelligence Research Institute (AIRI) have successfully lowered the latency between a change in brain activity and the presentation of the corresponding neurofeedback signal by a factor of 50. The results were obtained by employing a neural network trained in low-latency filtering of brain activity signals from diverse individuals. This approach opens up new prospects for the treatment of attention deficit disorder and epilepsy. A paper with the study findings has been published in Journal of Neural Engineering.

Online Rhythmic Transcranial Magnetic Stimulation Can Reveal the Precise Moment When Preferences Shift in the Human Brain

Cognitive dissonance is a complex and multifaceted psychological phenomenon that arises in challenging decision-making scenarios. Multiple regions of the brain participate in its occurrence, yet the neurodynamics of underlying cognitive mechanisms remain a subject of debate. Researchers from the HSE Institute for Cognitive Neuroscience have proposed the use of online transcranial magnetic stimulation applied to participants as they were actively engaged in tasks, to pinpoint the moment of cognitive dissonance resolution. Their findings have been published in a review paper in Frontiers in Human Neuroscience.

HSE University Holds 10th Summer School ‘Eye-tracking in the Lab and Beyond’

This year, more than 100 students from Russia and abroad took part in the 10th summer neurolinguistic school, ‘Eye-tracking in the Lab and Beyond’. The school is held annually by the HSE Center for Language and Brain. Leading experts spoke about advanced developments and research in the field of video-oculography.

New Technologies for Preserving Brain Functions: ‘Not Magic, but Normal Engineering’

New methods of brain mapping will make it easier to identify the cortex areas responsible for speech functions and to perform operations on the brain, as well as reduce the likelihood of damage to important areas. In addition, this will allow for more frequent use of non-invasive methods for restoring speech and other functions lost due to injuries and illnesses.