skip to main content
10.1145/3404555.3404593acmotherconferencesArticle/Chapter ViewAbstractPublication PagesiccaiConference Proceedingsconference-collections
research-article

End-to-End Image Reconstruction of Image from Human Functional Magnetic Resonance Imaging Based on the "Language" of Visual Cortex

Authors Info & Claims
Published:20 August 2020Publication History

ABSTRACT

In recent years, with the development of deep learning, the integration between neuroscience and computer vision has been deepened. In computer vision, it has been possible to generate images from text as well as semantic understanding from images based on deep learning. Here, text refers to human language, and the language that a computer can understand typically requires text to be encoded. In human brain visual expression, it also produces "descriptions" of visual stimuli, that is, the "language" that generates from the brain itself. Reconstruction of visual information is the process of reconstructing visual stimuli from the understanding of human brain, which is the most difficult to achieve in visual decoding. And based on the existing research of visual mechanisms, it is still difficult to understand the "language" of human brain. Inspired by generating images from text, we regarded voxel responses as the "language" of brain in order to reconstruct visual stimuli and built an end-to-end visual decoding model under the condition of small number of samples. We simply retrained a generative adversarial network (GAN) used to generate images from text on 1200 training data (including natural image stimuli and corresponding voxel responses). We regarded voxel responses as semantic information of brain, and sent them to GAN as prior information. The results showed that the decoding model we trained can reconstruct the natural images successfully. It also suggested the feasibility of reconstructing visual stimuli from "brain language", and the end-to-end model was more likely to learn the direct mapping between brain activity and visual perception. Moreover, it further indicated the great potential of combining neuroscience and computer vision.

References

  1. COX, D.D. and SAVOY, R.L., 2003. Functional magnetic resonance imaging (fMRI)"brain reading": detecting and classifying distributed patterns of fMRI activity in human visual cortex. Neuroimage 19, 2, 261--270.Google ScholarGoogle ScholarCross RefCross Ref
  2. HAYNES, J.D. and REES, G. Decoding mental states from brain activity in humans. Nature reviews neuroscience 7, 7, 523--534.Google ScholarGoogle Scholar
  3. MITCHELL, T.M., SHINKAREVA, S.V., CARLSON, A., CHANG, K.-M., MALAVE, V.L., MASON, R.A., and JUST, M.A., 2008. Predicting human brain activity associated with the meanings of nouns. Science 320, 5880, 1191--1195.Google ScholarGoogle Scholar
  4. HUTH, A.G., NISHIMOTO, S., VU, A.T., and GALLANT, J.L., 2012. A continuous semantic space describes the representation of thousands of object and action categories across the human brain. Neuron 76, 6, 1210--1224.Google ScholarGoogle ScholarCross RefCross Ref
  5. KAY, K.N., NASELARIS, T., PRENGER, R.J., and GALLANT, J.L., 2008. Identifying natural images from human brain activity. Nature 452, 7185, 352--355.Google ScholarGoogle Scholar
  6. NASELARIS, T., PRENGER, R.J., KAY, K.N., OLIVER, M., and GALLANT, J.L., 2009. Bayesian reconstruction of natural images from human brain activity. Neuron 63, 6, 902--915.Google ScholarGoogle ScholarCross RefCross Ref
  7. THIRION, B., DUCHESNAY, E., HUBBARD, E., DUBOIS, J., POLINE, J.-B., LEBIHAN, D., and DEHAENE, S., 2006. Inverse retinotopy: inferring the visual content of images from brain activation patterns. Neuroimage 33, 4, 1104--1116.Google ScholarGoogle ScholarCross RefCross Ref
  8. MIYAWAKI, Y., UCHIDA, H., YAMASHITA, O., SATO, M.-A., MORITO, Y., TANABE, H.C., SADATO, N., and KAMITANI, Y., 2008. Visual image reconstruction from human brain activity using a combination of multiscale local image decoders. Neuron 60, 5, 915--929.Google ScholarGoogle ScholarCross RefCross Ref
  9. YARGHOLI, E. and HOSSEIN-ZADEH, G.-A., 2015. Reconstruction of digit images from human brain fMRI activity through connectivity informed Bayesian networks. Journal of Neuroscience Methods.Google ScholarGoogle Scholar
  10. YARGHOLI, E. and HOSSEIN-ZADEH, G.-A., 2016. Brain Decoding-Classification of Hand Written Digits from fMRI Data Employing Bayesian Networks. Frontiers in Human Neuroscience 10(07/1302/17/received06/28/accepted), 351. DOI= http://dx.doi.org/10.3389/fnhum.2016.00351.Google ScholarGoogle Scholar
  11. YAMINS, D.L. and DICARLO, J.J., 2016. Using goal-driven deep learning models to understand sensory cortex. Nature Neuroscience 19, 3, 356--365.Google ScholarGoogle ScholarCross RefCross Ref
  12. VAN GERVEN, M.A., DE LANGE, F.P., and HESKES, T., 2010. Neural decoding with hierarchical generative models. Neural computation 22, 12, 3127--3142. Google ScholarGoogle Scholar
  13. HINTON and E., G. Reducing the Dimensionality of Data with Neural Networks. Science 313, 5786, 504--507.Google ScholarGoogle Scholar
  14. HAN, K., WEN, H., SHI, J., LU, K.-H., ZHANG, Y., and LIU, Z., 2018. Variational autoencoder: An unsupervised model for modeling and decoding fMRI activity in visual cortex. bioRxiv, 214247.Google ScholarGoogle Scholar
  15. SEELIGER, K., G L, U., AMBROGIONI, L., G L T RK, Y., and VAN GERVEN, M., 2018. Generative adversarial networks for reconstructing natural images from brain activity. Neuroimage 181, 775--785.Google ScholarGoogle ScholarCross RefCross Ref
  16. SHEN, G., DWIVEDI, K., MAJIMA, K., HORIKAWA, T., and KAMITANI, Y., 2018. End-to-end deep image reconstruction from human brain activity. bioRxiv, 272518. DOI= http: //dx. doi. org/10.1101/272518.Google ScholarGoogle Scholar
  17. REED, S., AKATA, Z., YAN, X., LOGESWARAN, L., SCHIELE, B., and LEE, H. Generative Adversarial Text to Image Synthesis.Google ScholarGoogle Scholar
  18. HORIKAWA, T. and KAMITANI, Y. Generic decoding of seen and imagined objects using hierarchical visual features. Nature Communications 8, 15037.Google ScholarGoogle Scholar
  19. GOODFELLOW, I., POUGET-ABADIE, J., MIRZA, M., XU, B., WARDE-FARLEY, D., OZAIR, S., COURVILLE, A., and BENGIO, Y., 2014. Generative adversarial nets. In Advances in neural information processing systems, 2672--2680. Google ScholarGoogle Scholar
  20. IOFFE, S. and SZEGEDY, C., 2015. Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv preprint arXiv:1502.03167.Google ScholarGoogle Scholar
  21. WANG, Z., BOVIK, A.C., SHEIKH, H.R., and SIMONCELLI, E.P., 2004. Image quality assessment: from error visibility to structural similarity. IEEE Transactions on Image Processing 13, 4, 600--612. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. End-to-End Image Reconstruction of Image from Human Functional Magnetic Resonance Imaging Based on the "Language" of Visual Cortex

        Recommendations

        Comments

        Login options

        Check if you have access through your login credentials or your institution to get full access on this article.

        Sign in
        • Published in

          cover image ACM Other conferences
          ICCAI '20: Proceedings of the 2020 6th International Conference on Computing and Artificial Intelligence
          April 2020
          563 pages
          ISBN:9781450377089
          DOI:10.1145/3404555

          Copyright © 2020 ACM

          Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          • Published: 20 August 2020

          Permissions

          Request permissions about this article.

          Request Permissions

          Check for updates

          Qualifiers

          • research-article
          • Research
          • Refereed limited
        • Article Metrics

          • Downloads (Last 12 months)8
          • Downloads (Last 6 weeks)1

          Other Metrics

        PDF Format

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader