주요메뉴 바로가기 본문 바로가기

주메뉴

IBS Conferences

Brain-Inspired AI Breakthrough: Making Computers See More Like Humans

- IBS-Yonsei Team Unveils Novel Lp-Convolution at ICLR 2025 -

A team of researchers from the Institute for Basic Science (IBS), Yonsei University, and the Max Planck Institute have developed a new artificial intelligence (AI) technique that brings machine vision closer to how the human brain processes images. Called Lp-Convolution, this method improves the accuracy and efficiency of image recognition systems while reducing the computational burden of existing AI models.

Bridging the Gap Between CNNs and the Human Brain

The human brain is remarkably efficient at identifying key details in complex scenes, an ability that traditional AI systems have struggled to replicate. Convolutional Neural Networks (CNNs)—the most widely used AI model for image recognition—process images using small, square-shaped filters. While effective, this rigid approach limits their ability to capture broader patterns in fragmented data.

More recently, Vision Transformers (ViTs) have shown superior performance by analyzing entire images at once, but they require massive computational power and large datasets, making them impractical for many real-world applications.

Inspired by how the brain’s visual cortex processes information selectively through circular, sparse connections, the research team sought a middle ground: Could a brain-like approach make CNNs both efficient and powerful?

Introducing Lp-Convolution: A Smarter Way to See

To answer this, the team developed Lp-Convolution, a novel method that uses a multivariate p-generalized normal distribution (MPND) to reshape CNN filters dynamically. Unlike traditional CNNs, which use fixed square filters, Lp-Convolution allows AI models to adapt their filter shapes—stretching horizontally or vertically based on the task, much like how the human brain selectively focuses on relevant details.

This breakthrough solves a long-standing challenge in AI research, known as the large kernel problem. Simply increasing filter sizes in CNNs (e.g., using 7×7 or larger kernels) usually does not improve performance, despite adding more parameters. Lp-Convolution overcomes this limitation by introducing flexible, biologically inspired connectivity patterns.

Real-World Performance: Stronger, Smarter, and More Robust AI

In tests on standard image classification datasets (CIFAR-100, TinyImageNet), Lp-Convolution significantly improved accuracy on both classic models like AlexNet and modern architectures like RepLKNet. The method also proved to be highly robust against corrupted data, a major challenge in real-world AI applications.

Moreover, the researchers found that when the Lp-masks used in their method resembled a Gaussian distribution, the AI’s internal processing patterns closely matched biological neural activity, as confirmed through comparisons with mouse brain data.

“We humans quickly spot what matters in a crowded scene,” said Dr. C. Justin LEE, Director of the Center for Cognition and Sociality within the Institute for Basic Science. “Our Lp-Convolution mimics this ability, allowing AI to flexibly focus on the most relevant parts of an image—just like the brain does.”

Impact and Future Applications

Unlike previous efforts that either relied on small, rigid filters or required resource-heavy transformers, Lp-Convolution offers a practical, efficient alternative. This innovation could revolutionize fields such as:

  • - Autonomous driving, where AI must quickly detect obstacles in real time
  • - Medical imaging, improving AI-based diagnoses by highlighting subtle details
  • - Robotics, enabling smarter and more adaptable machine vision under changing conditions

“This work is a powerful contribution to both AI and neuroscience,” said Director C. Justin Lee. “By aligning AI more closely with the brain, we’ve unlocked new potential for CNNs, making them smarter, more adaptable, and more biologically realistic.”

Looking ahead, the team plans to refine this technology further, exploring its applications in complex reasoning tasks such as puzzle-solving (e.g., Sudoku) and real-time image processing.

The study will be presented at the International Conference on Learning Representations (ICLR) 2025, and the research team has made their code and models publicly available: https://github.com/jeakwon/lpconv/

Figure 1. Information Processing Structures of the Brain’s Visual Cortex and Artificial Neural Networks
        In the actual brain’s visual cortex, neurons are connected broadly and smoothly around a central point, with connection strength varying gradually with distance (a, b). This spatial connectivity follows a bell-shaped curve known as a ‘Gaussian distribution,’ enabling the brain to integrate visual information not only from the center but also from the surrounding areas.
        In contrast, traditional Convolutional Neural Networks (CNNs) process information by having neurons focus on a fixed rectangular region (e.g., 3×3, 5×5, etc.) (c, d). CNN filters move across an image at regular intervals, extracting information in a uniform manner, which limits their ability to capture relationships between distant visual elements or respond selectively based on importance.
        This study addresses the differences between these biological structures and CNNs, proposing a new filter structure called 'Lp-Convolution' that mimics the brain’s connectivity patterns. In this structure, the range and sensitivity of a neuron’s input are designed to naturally spread in a Gaussian-like form, allowing the system to self-adjust during training—emphasizing important information more strongly while downplaying less relevant details. This enables image processing that is more flexible and biologically aligned compared to traditional CNNs.
Figure 1. Information Processing Structures of the Brain’s Visual Cortex and Artificial Neural Networks
In the actual brain’s visual cortex, neurons are connected broadly and smoothly around a central point, with connection strength varying gradually with distance (a, b). This spatial connectivity follows a bell-shaped curve known as a ‘Gaussian distribution,’ enabling the brain to integrate visual information not only from the center but also from the surrounding areas.
In contrast, traditional Convolutional Neural Networks (CNNs) process information by having neurons focus on a fixed rectangular region (e.g., 3×3, 5×5, etc.) (c, d). CNN filters move across an image at regular intervals, extracting information in a uniform manner, which limits their ability to capture relationships between distant visual elements or respond selectively based on importance.
This study addresses the differences between these biological structures and CNNs, proposing a new filter structure called 'Lp-Convolution' that mimics the brain’s connectivity patterns. In this structure, the range and sensitivity of a neuron’s input are designed to naturally spread in a Gaussian-like form, allowing the system to self-adjust during training—emphasizing important information more strongly while downplaying less relevant details. This enables image processing that is more flexible and biologically aligned compared to traditional CNNs.

Figure 2. Brain-Inspired Design of Lp-Convolution
        The brain processes visual information using a Gaussian-shaped connectivity structure that gradually spreads from the center outward, flexibly integrating a wide range of information. In contrast, traditional CNNs face issues where expanding the filter size dilutes information or reduces accuracy (d, e).
        To overcome these structural limitations, the research team developed Lp-Convolution, inspired by the brain’s connectivity (a–c). This design spatially distributes weights to preserve key information even over large receptive fields, effectively addressing the shortcomings of conventional CNNs.
Figure 2. Brain-Inspired Design of Lp-Convolution
The brain processes visual information using a Gaussian-shaped connectivity structure that gradually spreads from the center outward, flexibly integrating a wide range of information. In contrast, traditional CNNs face issues where expanding the filter size dilutes information or reduces accuracy (d, e).
To overcome these structural limitations, the research team developed Lp-Convolution, inspired by the brain’s connectivity (a–c). This design spatially distributes weights to preserve key information even over large receptive fields, effectively addressing the shortcomings of conventional CNNs.

Notes for editors

- References
Jea Kwon, Sungjun Lim, Kyungwoo Song, C. Justin Lee. Brain-Inspired Lp-Convolution Benefits Large Kernels and Aligns Better with Visual Cortex. ICLR 2025.
https://openreview.net/forum?id=0LSAmFCc4p


- Media Contact
For further information or to request media assistance, please contact C. Justin Lee at the Center for Cognition and Sociality, Institute for Basic Science (IBS) (cjl@ibs.re.kr) or William I. Suh at the IBS Public Relations Team (willisuh@ibs.re.kr).


- About the Institute for Basic Science (IBS)
IBS was founded in 2011 by the government of the Republic of Korea with the sole purpose of driving forward the development of basic science in South Korea. IBS has 7 research institutes and 32 research centers as of January 2025. There are eight physics, three mathematics, five chemistry, seven life science, two earth science, and seven interdisciplinary research centers.


Research

Are you satisfied with the information on this page?

Content Manager
Public Relations Team : Yim Ji Yeob   042-878-8173
Last Update 2023-11-28 14:20