Skip to main content

第一部分 考试说明

 一、考试性质

本《神经计算科学》考试大纲适用于北京师范大学脑与认知科学学院各专业的硕士研究生入学考试。要求考生全面系统地掌握计算神经科学的基本概念和研究方法,能熟练运用计算神经科学知识分析神经生物学基本问题。  考试对象为报考我校硕士研究生入学考试的准考考生。

 二、考试形式与试卷

(一)答卷方式:闭卷、笔试

(二)答题时间:180分钟

(三)各题型分值(总分300

1 名词解释30 

2.简答题  90

3.论述题  180 

 三、考试要求

1.掌握计算神经科学的基本概念和基础理论

2.了解计算神经科学的一些经典进展 

3.具有运用基本概念和基础理论分析问题与解决问题的能力  

 

第二部分 考察要点

PART I - ANALYZING AND MODELING NEURAL RESPONSES

Chapter 1 - Neural Encoding I: Firing Rates and Spike Statistics

1. Properties of NeuronsRecording Neuronal ResponsesFrom Stimulus to Response

2. Spike Trains and Firing Rates; Tuning CurvesSpike-Count Variability

3. Describing the StimulusThe Spike-Triggered AverageWhite-Noise StimuliMultiple-Spike-Triggered Averages and Spike-Triggered Correlations

4. Spike Train Statistics; The Homogeneous Poisson ProcessThe Spike-Train Autocorrelation FunctionThe Inhomogeneous Poisson ProcessThe Poisson Spike GeneratorComparison with Data

5. The Neural Code; Independent-Spike, Independent Neuron and Correlation CodesTemporal Codes

Chapter 2 - Neural Encoding II: Reverse Correlation and Receptive Fields

1. Estimating Firing Rates; The Most Effective Stimulus; Static Nonlinearities

2. Early Visual System; The Retinotopic Map; Visual Stimuli; The Nyquist Frequency

3. Reverse Correlation Methods - Simple Cells;

Spatial Receptive Fields; Temporal Receptive Fields; Response of a Simple Cell to a Counterphase Grating;  Space-Time Receptive Fields; Nonseparable Receptive Fields; Static Nonlinearities - Simple Cells

4. Static Nonlinearities - Complex Cells

5. Receptive Fields in the Retina and LGN

6. Constructing V1 Receptive Fields

Chapter 3 - Neural Decoding

1. Encoding and Decoding

2. Discrimination; ROC Curves; ROC Analysis of Motion Discrimination; The Likelihood Ratio Test

3. Population Decoding; Encoding and Decoding Direction; Optimal Decoding Methods; Fisher Information; Optimal Discrimination

4. Spike Train Decoding

 

PART II - MODELING NEURONS AND NETWORKS

Chapter 4 - Model Neurons I: Neuroelectronics

1. Levels of Neuron Modeling

2. Electrical Properties of Neurons

Intracellular Resistance; Membrane Capacitance and Resistance; Equilibrium and Reversal Potentials; The Membrane Current

3. Single-Compartment Models

Integrate-and-Fire Models; Spike-Rate Adaptation and Refractoriness

4. Voltage-Dependent Conductances

Persistent Conductances; Transient Conductances; Hyperpolarization-Activated Conductances

5. The Hodgkin-Huxley Model

6. Modeling Channels

7. Synaptic Conductances

The Postsynaptic Conductance; Release Probability and Short-Term Plasticity

8. Synapses on Integrate-and-Fire Neurons

Regular and Irregular Firing Modes

Chapter 5 - Model Neurons II: Conductances and Morphology

1. Levels of Neuron Modeling

2. Conductance-Based Models

The Connor-Stevens Model; Postinhibitory Rebound and Bursting

3. The Cable Equation

Linear Cable Theory (An Infinite Cable; An Isolated Branching Node); The Rall Model; The Morphoelectrotonic Transform

4. Multi-Compartment Models

Action Potential Propagation Along an Unmyelinated Axon; Propagation Along a Myelinated Axon

Chapter 6 - Network Models

1. Firing-Rate Models

Feedforward and Recurrent Networks; Continuously Labelled Networks

2. Feedforward Networks

Neural Coordinate Transformations

3. Recurrent Networks

Linear Recurrent Networks (Selective Amplification; Input Integration; Continuous Linear Recurrent Networks);

Nonlinear Recurrent Networks (Nonlinear Amplification; A Recurrent Model of Simple Cells in Primary Visual Cortex; A Recurrent Model of Complex Cells in Primary Visual Cortex; Winner-Take-All Input Selection; Gain Modulation; Sustained Activity; Maximum Likelihood and Network Recoding )

4. Network Stability

Associative Memory

5. Excitatory-Inhibitory Networks

Homogeneous Excitatory and Inhibitory Populations (Phase-Plane Methods and Stability Analysis); The Olfactory Bulb; Oscillatory Amplification

6. Stochastic Networks


PART III - PLASTICITY AND LEARNING

Chapter 7 - Plasticity and Learning

1. Stability and Competition

2. Synaptic Plasticity Rules

The Basic Hebb Rule; The Covariance Rule; The BCM Rule; Synaptic Normalization (Subtractive Normalization; Multiplicative Normalization and the Oja Rule); Timing-Based Rules

3. Unsupervised Learning

Single Postsynaptic Neuron (Principal Component Projection; Hebbian Development and Ocular Dominance; Hebbian Development of Orientation Selectivity; Temproal Hebbian Rules and Trace Learning)

Multiple Postsynaptic Neurons (Fixed Linear Recurrent Connections; Competitive Hebbian Learning; Feature-Based Models; Anti-Hebbian Modification; Timing-Based Plasticity and Prediction)

4. Supervised Learning

Supervised Hebbian Learning (Classification and the Perceptron; Function Approximation)

5. Supervised Error-Correcting Rules

The Perceptron Learning Rule

6. The Delta Rule

Contrastive Hebbian Learning