Institut
    für Grundlagen der Informationsverarbeitung (7080)
Lecturer:
    O.Univ.-Prof. Dr. Wolfgang Maass
Office hours: by appointment (via e-mail)
E-mail: maass@igi.tugraz.at
    Homepage: https://igi-web.tugraz.at/people/maass/
Assoc. Prof. Dr. Robert Legenstein
Office hours: by appointment (via e-mail)
  E-mail: robert.legenstein@igi.tugraz.at
    Homepage: www.igi.tugraz.at/legi/
    
Marblestone, Adam, Greg Wayne, and Konrad Kording (2016). "Toward an Integration of Deep Learning and Neuroscience." Frontiers in Computational Neuroscience, 10. [PDF@Frontiers].
hypothesizes that biological neuronal systems may utilize learning processes that share similarities with deep learning techniques. We will discuss in this semester this paper and selected references therein.Abstract: Neuroscience has focused on the detailed implementation of computation, studying neural codes, dynamics and circuits. In machine learning, however, artificial neural networks tend to eschew precisely designed codes, dynamics or circuits in favor of brute force optimization of a cost function, often using simple and relatively uniform initial architectures. Two recent developments have emerged within machine learning that create an opportunity to connect these seemingly divergent perspectives. First, structured architectures are used, including dedicated systems for attention, recursion and various forms of short- and long-term memory storage. Second, cost functions and training procedures have become more complex and are varied across layers and over time. Here we think about the brain in terms of these ideas. We hypothesize that (1) the brain optimizes cost functions, (2) these cost functions are diverse and differ across brain locations and over development, and (3) optimization operates within a pre-structured architecture matched to the computational problems posed by behavior. Such a heterogeneously optimized system, enabled by a series of interacting cost functions, serves to make learning data-efficient and precisely targeted to the needs of the organism. We suggest directions by which neuroscience could seek to refine and test these hypotheses.
(1) R. C. O’Reilly, D. Wyatte, and J. Rohrlich
    (2014). Learning through time in the thalamocortical loops.
    arXiv:1407.3432v1. https://arxiv.org/abs/1407.3432
    
(2) W. Lotter, G. Kreiman, and D. Cox (2016).
    Unsupervised learning of visual structure using predictive
    generative networks. arXiv:1511. http://arxiv.org/abs/1511.06380
    
(3) R. C. O’Reilly, T. E. Hazy, J. Mollick, P.
    Mackie, and S. Herd (2014). Goal-driven cognition in the brain:
    A computational framework. arXiv:1404.7591v1. https://arxiv.org/abs/1404.7591
    
(4) J. O. Rombouts, S. M. Bohte, and P. R.
    Roelfsema (2015). How attention can create synaptic tags for
    the learning of working memories in sequential tasks. PLOS
    Computational Biology | DOI:10.1371/journal.pcbi.1004060.
    
    http://journals.plos.org/ploscompbiol/article?id=10.1371/journal.pcbi.1004060
    
(5) G. Wayne, and L. F. Abbott (2014).
    Hierarchical control using networks trained with higher-level
    forward models. Neural Comput 26(10):2163-2193. doi:
    10.1162/NECO_a_00639. http://www.ncbi.nlm.nih.gov/pubmed/25058706
    
(6) T. J. Sejnowski, H. Poizner, G. Lynch, S.
    Gepshtein, and R. J. Greenspan (2014). Prospective
    Optimization. Proc IEEE Inst Electr Electron Eng. 2014
    May;102(5):799-811. DOI: 10.1109/JPROC.2014.2314297. https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4201124/
    
(7) E. Jonas, and K. Kording (2016). Could a
    neuroscientist understand a microprocessor?. bioRxiv, 055624.
    
    http://www.biorxiv.org/content/early/2016/05/26/055624.abstract
    
(8) I. Sutskever, J. Martens, G. E. Dahl, and
    G. E. Hinton (2013). On the importance of initialization and
    momentum in deep learning. Proceedings of the 30 th
    International Conference on Machine Learning, Atlanta, Georgia,
    USA, 2013. JMLR: W&CP volume 28. http://www.cs.toronto.edu/~fritz/absps/momentum.pdf
    
(9) J. Yosinski, J. Clune, and Y. Bengio
    (2014). How transferable are features in deep neural networks?
    In Advances in Neural Information Processing Systems 27:
    3320-3328. 
    http://papers.nips.cc/paper/5347-how-transferable-are-features-in-deep-neural-networks
    
(10) C. Gülcehre, and Y. Bengio (2016).
    Knowledge matters: Importance of prior information for
    optimization. Journal of Machine Learning Research, 17(8),
    1-32. 
    www.jmlr.org/papers/volume17/gulchere16a/gulchere16a.pdf
    
| 
          Date | 
          
          Speaker | 
          
          Talks | 
          |
| 24.10.2016  
                
          16:15-18:00 | 
          Maass,
          Legenstein | 
          
          Introduction | 
          
          PDF | 
        
| 
          21.11.2016         
          15:45-18:00 | 
          Absenger,
          Mulle | 
          Goal-driven cognition in
          the brain: A computational framework. R. C. O’Reilly, T.
          E. Hazy, J. Mollick, P. Mackie, and S. Herd
          (2014) | 
          
          PDF | 
        
| 28.11.2016 15:45-18:00 | Marchetto,
          Raggam | 
          Learning through time in
          the thalamocortical loops. R. C. O’Reilly, D. Wyatte, and
          J. Rohrlich (2014) | 
          |
| Harb, Micorek | Prospective Optimization. T. J.
          Sejnowski, H. Poizner, G. Lynch, S. Gepshtein, and R. J.
          Greenspan (2014) | 
          
          PDF | 
        |
| 12.12.2016 15:45-18:00 | Steger,
          Zöhrer | 
          Unsupervised learning of visual structure using predictive generative networks. W. Lotter, G. Kreiman, and D. Cox (2016) | 
          PDF | 
        
| Wohlhart, Müller | 
          Could a neuroscientist understand a microprocessor?. E. Jonas, and K. Kording (2016) | 
          PDF | 
        |
| 09.01.2017 15:45-18:00 | 
          Legenstein | 
          A brief introduction
          into deep learning | 
          |
| Lindner, Narnhofer | 
          Knowledge matters: Importance of prior information for optimization. C. Gülcehre, and Y. Bengio (2016) | 
          PDF | 
        |
| 
          23.01.2017        
            15:45-18:00 | 
          Topic,
          Eibl | 
          On the importance of initialization and momentum in deep learning. I. Sutskever, J. Martens, G. E. Dahl, and G. E. Hinton (2013) | 
          PDF | 
        
| Fuchs,
          Ainetter | 
          How transferable are features in deep neural networks? J. Yosinski, J. Clune, and Y. Bengio (2014) | 
          PDF |