Learning in neural networks
CS-479
- Announcements (Forum)
- Ed Discussion Forum (External tool)
- Mini Project 1 : Visual Receptive Field Development (File)
- Mini Project 2 : Third Factors in Reinforcement Learning (File)
- Registration List for the Miniprojects (URL)
- Exam Booking Papers Sheet (URL)
- List of papers for Oral Exam (File)
- Mini-projects: First deadline (Assignment)
- Fraud Detection 1st Deadline (Scheduler)
- Fraud Detection 2nd Deadline (Scheduler)
- Oral Exam Schedule (Scheduler)
Introduction to class: No Backprop please; Background in neuroscience and hardware.
First Lecture: Hebbian learning for PCA
- Exercises Week 1 (File)
- Lecture01LearningInNN (File)
- Week1Blackboard (File)
- Solutions Exercises Week 1 (File)
Hebbian learning for ICA
The problem of independent component analysis (ICA) is introduced and its relation to PCA (wee last week) is discussed. We will show that nonlinear Hebbian learning gives rise to ICA. To implement ICA, data normally needs to be centered and prewhitened. All concepts are discussed step by step.
- Exercises Week 2 (File)
- Lecture02Blackboard-ICA (File)
- Lecture02ICA-LearningInNN (File)
- Exercises Week 2 solutions (File)
This week we apply Hebbian learning rules in a network with lateral interactions. We will see that neurons can then extract several principal components or independent components. We will also see that such a network can perform k-means clustering. With strong inhibitory interactions one also talks about 'competitive dynamics' or competitive learning.
This lecture ends the overview of 2-factor rules. Next week we continue with Reinforcement learning and 3-factor rules.
- Exercises Week 3 (File)
- Lecture03CompetitiveLearningInNN (File)
- BlackboardLecture3 (File)
- Week 3 Solutions (File)
A preliminary introduction to Reinforcement Learning (RL0):
We look at bandit problems which I call 1-step-horizon problems: you take an action and immediately the trial ends with a reward of variable magnitude.
- Exercises Week 4 (File)
- Lecture04RL0 (File)
- Scan Blackboard RL0 (File)
- Exercises Week 4 solutions (File)
RL 1 (continued). The discussion on Reinforcement Learning in the multistep horizon is continued.
Bellman equation, SARSA, and variations of SARSA. Backup diagram. Eligibility traces.
- Exercises Week 5 (File)
- Computer Exercise Week 5 (Folder)
- Lecture05RL1 (File)
- Exercises Week 5 solutions (File)
RL2 - The full framework of TD learning, including deep reinforcement learning.
- Exercises Week 6 (File)
- Computer exercises Week 6 (File)
- Lecture06RL2 (File)
- Blackboard RL2 (File)
- Exercises Week 6 solutions (File)
- Computer exercises solutions (File)
Policy gradient methods in Reinforcement Learning.
- Lecture07RL3 (File)
- Exercises Week 7 (File)
- Computer exercises (File)
- week7BlackboardScan1 (File)
- Solutions Week 7 (File)
- Computer exercises solutions (File)
Final session on the Foundation of Reinforcement Learning.
We discuss the Actor-Critic architecture and the Actor-Critic-Algorithm in the narrow sense, sometimes also called advantage actor critic.
- Exercises Week 8 (File)
- Lecture08RL5Actor-CriticFinal (File)
- Week08Blackboard (File)
- Exercises week 8 solutions (File)
- Lecture09RLandBrain1-threefactor (File)
- Exercises Week 9 (File)
- Computer exercises (related to last week's content) (File)
- Exercises Week 9 solutions (File)
Easter Break
Reinforcement Learning and the Brain: Surprise and Novelty are alternatives to Reward as a third factor.
Detailed Models of Plasticity with Spiking Neurons.
Representation learning in deep networks without Backprop.