• HOME
  • NEWS
  • EXPLORE
    • CAREER
      • Companies
      • Jobs
    • EVENTS
    • iGEM
      • News
      • Team
    • PHOTOS
    • VIDEO
    • WIKI
  • BLOG
  • COMMUNITY
    • FACEBOOK
    • INSTAGRAM
    • TWITTER
Saturday, September 13, 2025
BIOENGINEER.ORG
No Result
View All Result
  • Login
  • HOME
  • NEWS
  • EXPLORE
    • CAREER
      • Companies
      • Jobs
        • Lecturer
        • PhD Studentship
        • Postdoc
        • Research Assistant
    • EVENTS
    • iGEM
      • News
      • Team
    • PHOTOS
    • VIDEO
    • WIKI
  • BLOG
  • COMMUNITY
    • FACEBOOK
    • INSTAGRAM
    • TWITTER
  • HOME
  • NEWS
  • EXPLORE
    • CAREER
      • Companies
      • Jobs
        • Lecturer
        • PhD Studentship
        • Postdoc
        • Research Assistant
    • EVENTS
    • iGEM
      • News
      • Team
    • PHOTOS
    • VIDEO
    • WIKI
  • BLOG
  • COMMUNITY
    • FACEBOOK
    • INSTAGRAM
    • TWITTER
No Result
View All Result
Bioengineer.org
No Result
View All Result
Home NEWS Science News

Take a look, and you’ll see, into your imagination

Bioengineer by Bioengineer
May 31, 2017
in Science News
Reading Time: 2 mins read
0
Share on FacebookShare on TwitterShare on LinkedinShare on RedditShare on Telegram
IMAGE

Credit: Kyoto University

Kyoto, Japan — Scanning your brain to decode the contents of your mind has been a subject of intense research interest for some time. As studies have progressed, scientists have gradually been able to interpret what test subjects see, remember, imagine, and even dream.

There have been significant limitations, however, beginning with a necessity to extensively catalog each subject's unique brain patterns, which are then matched with a small number of pre-programmed images. These procedures require that subjects undergo lengthy and expensive fMRI testing.

Now a team of researchers in Kyoto has used neural network-based artificial intelligence to decode and predict what a person is seeing or imagining, referring to a significantly larger catalog of images. Their results are reported in Nature Communications.

"When we gaze at an object, our brains process these patterns hierarchically, starting with the simplest and progressing to more complex features," explains team leader Yukiyasu Kamitani of Kyoto University.

"The AI we used works on the same principle. Named 'Deep Neural Network', or DNN, it was trained by a group now at Google."

The team from Kyoto University and ATR (Advanced Telecommunications Research) Computational Neuroscience Laboratories discovered that brain activity patters can be decoded, or translated, into signal patterns of simulated neurons in the DNN when both are shown the same image.

Additionally, the researchers found that lower and higher visual areas in the brain were better at decoding respective layers of the DNN, revealing a homology between the human brain and the neural network.

"We tested whether a DNN signal pattern decoded from brain activity can be used to identify seen or imagined objects from arbitrary categories," explains Kamitani. "The decoder takes neural network patterns and compares these with image data from a large database. Sure enough, the decoder could identify target objects with high probability."

As brain decoding and AI development advance, Kamitani hopes to improve the image identification accuracy of their technique. He concludes, "Bringing AI research and brain science closer together could open the door to new brain-machine interfaces, perhaps even bringing us closer to understanding consciousness itself."

###

The paper "Generic decoding of seen and imagined objects using hierarchical visual features" appeared 22 May 2017 in Nature Communications, with doi: 10.1038/ncomms15037

Kyoto University is one of Japan and Asia's premier research institutions, founded in 1897 and responsible for producing numerous Nobel laureates and winners of other prestigious international prizes. A broad curriculum across the arts and sciences at both undergraduate and graduate levels is complemented by numerous research centers, as well as facilities and offices around Japan and the world. For more information please see: http://www.kyoto-u.ac.jp/en

Media Contact

Raymond Kunikane Terhune
[email protected]
81-757-535-728
@KyotoU_News

http://www.kyoto-u.ac.jp/en

############

Story Source: Materials provided by Scienmag

Share12Tweet7Share2ShareShareShare1

Related Posts

Polyacrylic Acid-Copper System Detects Gaseous Hydrogen Peroxide

Polyacrylic Acid-Copper System Detects Gaseous Hydrogen Peroxide

September 13, 2025
blank

Unveiling Arabidopsis Aminotransferases’ Multi-Substrate Specificity

September 13, 2025

Insights on Menstrual Health in Eating Disorder Units

September 12, 2025

Nicotine Dependence Linked to Health Behaviors in Korean Smokers

September 12, 2025
Please login to join discussion

POPULAR NEWS

  • blank

    Breakthrough in Computer Hardware Advances Solves Complex Optimization Challenges

    152 shares
    Share 61 Tweet 38
  • New Drug Formulation Transforms Intravenous Treatments into Rapid Injections

    116 shares
    Share 46 Tweet 29
  • Physicists Develop Visible Time Crystal for the First Time

    65 shares
    Share 26 Tweet 16
  • A Laser-Free Alternative to LASIK: Exploring New Vision Correction Methods

    49 shares
    Share 20 Tweet 12

About

We bring you the latest biotechnology news from best research centers and universities around the world. Check our website.

Follow us

Recent News

Polyacrylic Acid-Copper System Detects Gaseous Hydrogen Peroxide

Unveiling Arabidopsis Aminotransferases’ Multi-Substrate Specificity

Insights on Menstrual Health in Eating Disorder Units

  • Contact Us

Bioengineer.org © Copyright 2023 All Rights Reserved.

Welcome Back!

Login to your account below

Forgotten Password?

Retrieve your password

Please enter your username or email address to reset your password.

Log In
No Result
View All Result
  • Homepages
    • Home Page 1
    • Home Page 2
  • News
  • National
  • Business
  • Health
  • Lifestyle
  • Science

Bioengineer.org © Copyright 2023 All Rights Reserved.