News / Tech News

    Making AI smarter with an artificial, multisensory integrated neuron

    The feel of a cat's fur can reveal some information, but seeing the feline provides critical details: Is it a housecat or a lion? While the sound of fire crackling may be ambiguous, its scent confirms the burning wood. Our senses synergize to give a comprehensive understanding, particularly when individual signals are subtle.



    Researchers have developed a bio-inspired artificial neuron that can process visual and tactile sensory inputs together. Photo: Tyler Henderson/Penn State


    The collective sum of biological inputs can be greater than their individual contributions. Robots tend to follow more straightforward addition, but Penn State researchers have now harnessed the biological concept for application in artificial intelligence to develop the first artificial, multisensory integrated neuron.

    Led by Saptarshi Das, the team published its work in Nature Communications.

    "Robots make decisions based on the environment they are in, but their sensors do not generally talk to each other," said Das.

    "A collective decision can be made through a sensor processing unit, but is that the most efficient or effective method? In the human brain, one sense can influence another and allow the person to better judge a situation."

    A car might have one sensor scanning for obstacles, while another senses darkness to modulate the intensity of the headlights. Individually, these sensors relay information to a central unit that then instructs the car to brake or adjust the headlights.

    According to Das, this process consumes more energy. Allowing sensors to communicate directly with each other can be more efficient in terms of energy and speed — particularly when the inputs from both are faint.

    "Biology enables small organisms to thrive in environments with limited resources, minimizing energy consumption in the process," said Das.

    "The requirements for different sensors are based on the context — in a dark forest, you'd rely more on listening than seeing, but we don't make decisions based on just one sense. We have a complete sense of our surroundings, and our decision making is based on the integration of what we're seeing, hearing, touching, smelling, et cetera. The senses evolved together in biology, but separately in AI. In this work, we're looking to combine sensors and mimic how our brains work."

    Das said that an artificial multisensory neuron system could enhance sensor technology's efficiency, paving the way for more eco-friendly AI uses.

    As a result, robots, drones and self-driving vehicles could navigate their environment more effectively while using less energy. (U.S. National Science Foundation)

    NOVEMBER 18, 2023



    YOU MAY ALSO LIKE

    Researchers develop novel bubble-based technique to watch immune cells at work.
    Sensitive technique could revolutionize treatment during pandemics.
    An international group of scientists based in Russia’s National Research University of Electronic Technology (MIET) have managed to promote glycine crystals' piezoelectric effect - where an electric charge is produced by applying pressure.
    Researchers tracked how current language models such as ChatGPT mistake nonsense sentences as meaningful.
    As current courses through a battery, its materials erode over time. Mechanical influences such as stress and strain affect this trajectory, although their impacts on battery efficacy and longevity are not fully understood.
    High energy density batteries could potentially outperform lithium-ion batteries.

    © 1991-2023 The Titi Tudorancea Bulletin | Titi Tudorancea® is a Registered Trademark | Terms of use and privacy policy
    Contact