YOU ARE DOWNLOADING DOCUMENT

Please tick the box to continue:

Transcript
Page 1: Semester presentation

SEMESTER REPORT

TOPIC “NEURAL CIRCUITS AS

COMPUTATIONAL DYNAMICAL SYSTEMS”

DAVID SUSSILLO

PRESENTED BY KHUSH BAKHAT

Page 2: Semester presentation

NEURAL NETWORK

• In computer science , artificial neural network ANNS are “Computational Models” inspired by the animals’ central nervous system

• These models are capable of machine learning and pattern recognition.

Page 3: Semester presentation

WHY USE ANIMAL NERVOUS SYSTEM

• The objective of learning in biological organism is to achieve a closer optimalstate

Page 4: Semester presentation

NEURAL NETWORK TASK

• Control

• Classification

• Predication

• Approximation

Page 5: Semester presentation

NEURAL CIRCUIT

• Neurons never function in isolation they are organized into circuits that process specific kind of data

• Neural Circuit is a functional entity of interconnected neurons that are able to regulate its own activity

Page 6: Semester presentation

NEURAL CIRCUITS AS COMPUTATIONAL

DYNAMICAL SYSTEMS

• Many recent Studies of neuron recorded from “Cortex” reveal complex temporal

Dynamics

• How such dynamics embody the computations that ultimately lead to behavior

remains a mystery

• Approaching this issue requires developing plausible hypotheses couched in terms

of “Neural Dynamics”

• A tool ideally suited to aid this question is “Recurrent Neural Network” (RNN)

• RNN straddle the fields of non linear dynamical systems and machine learning

• Recently RNN have seen great advances in both theory and application

• In this paper David summarize recent theoretical and technological advances &

highlight an examples of how RNNs helped to explain perplexing high

dimensional neurophysiological data in the “Prefrontal Cortex”

Page 7: Semester presentation
Page 8: Semester presentation
Page 9: Semester presentation

SPECIAL TOPIC IN TOC“RECURRENT NEURAL NETWORK”

• RNN is a class of neural network where connections between units form“Direct circles (cycle graphs) “

• This creates an internal state of the network, which allow it to exhibit dynamictemporal behavior. RNN can use their internal memory to process an arbitrarysequence of inputs

Page 10: Semester presentation

Relationship Between TOC and Neural Network

Page 11: Semester presentation
Page 12: Semester presentation

OPTIMIZING RNNS

• A network model is designed by hand to reproduce and thus explain a setexperimental findings

• Modeling using RNNs that have optimized , or trained

• Optimized means desired inputs and outputs are first defined before training

• Optimizing a network tells the network “WHAT” it should accomplish , with avery few explicit instructions on “HOW "to do it

• RNNs becomes a method of “hypothesis generation” for futureexperimentation and data analysis

Page 13: Semester presentation

REVERSE ENGINEERING AN RNN AFTER OPTIMIZATION

• Revealing the dynamical mechanism employed by an RNN to solve a particular task involves a final step after optimization: one must reverse engineer the solution found by the RNN

• Solution was not constructed with reverse engineer step

• RNNs could understand the employing techniques from non linear dynamical systems theory

Page 14: Semester presentation

APPLICATIONS

• IN this paper reverse engineer variety of RNNs that were optimized to perform simple tasks

A memory Device

An input dependent pattern generator

• The key step in RE involves

Finding the fixed points of the network

Performing linearization of the network dynamics around those fixed points

• The Fixed Points provide a “Dynamical Skeleton” for understanding the global structure of dynamics in the state space

Page 15: Semester presentation

A 3-BIT MEMORY

• Understanding how memories can be represented in biological neural networks haslong been studied in neuroscience.

• In this toy example and he trained an RNN to generate the dynamics necessary toimplement a 3-bit memory.

• Three inputs enter the RNN and specify the states of the three bits individually.

• This 3- bit memory resistant to cross talk.

• After training, the RNN successfully implemented the 3-bit memory.

• RE RNNs for finding all the fixed points and linear system around these fixed points

• A saddle point is a fixed point with both stable and unstable dimensions

• The Saddle nodes were responsible for implementing the input- dependenttransitions between the stable attractors

Page 16: Semester presentation

Trained RNN to generate the dynamics necessary to implement a 3-bit memory

Page 17: Semester presentation

CONTEXT DEPENDENT DECISION MAKING IN PREFRONTAL CORTEX

• Animals are not limited to simple stimulus and response reflexes.

• They can rapidly and flexibly accommodate to context: as the context changes, the same stimuli can elicit dramatically different behaviors.

• To study this type of contextually dependent decision making, monkeys were trained to flexibly select and accumulate evidence from noisy visual stimuli in order to make discrimination.

• On the basis of a contextual cue, the monkeys either differentiated the direction of motion or color of a random-dot display (Figure 3a). While the monkeys engaged in the task, neural responses in prefrontal cortex (PFC) were recorded.

• These neurons showed mixed selectivity to both motion and color sensory evidence, regardless of which stimulus was relevant.

Page 18: Semester presentation

CONTINUE…..

• To discover how a single circuit could selectivity integrate one stimulus while ignoring another, despite the presence of both the RNN approach was applied

• The output of the RNN was in future to be analogous to the decision important to the saccade of the monkey

Page 19: Semester presentation
Page 20: Semester presentation
Page 21: Semester presentation

CONCLUSION:

• The study of neural dynamics at the circuit and systems level is an area of extremely active research. RNNs are a near ideal modeling framework for studying neural circuit dynamics because they share fundamental features with biological tissue, for example, feedback, nonlinearity, and parallel and distributed computing.

• By training RNNs on what to compute, but not how to compute it, researchers can generate novel ideas and testable hypotheses regarding the biological circuit mechanism.

• Further, RNNs provide a rigorous test bed in which to test ideas related to neural computation at the network level.

• The combined approaches of animal behavior and neurophysiology, alongside RNN modeling, may prove a powerful combination of handling the onslaught of high dimensional neural data that is to come


Related Documents