Learning Flashcards

(61 cards)

1
Q

Representation

A

How we represent world so can make decisions about it
Sensory representations
Uncertainty estimation
Bayesian integration

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

valuation

A

value representations in pfc
Availability/desirability
Assign value, encode diff aspects of value

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

action selection

A

Striatum - bias process
Action selection in striatum

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

outcome evaluation

A

Confidence/metacognition

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

learning

A

Key part, feedback changes preferences
If incorrect = something wrong = estimate how much you should update Your beliefs or represention, or valuation or action selection to get better

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

mice adapt to changes in reward contingencies

A

Adaptation in learning
Use reward history to assess whether urewarded trials is bc its unrewarded or bc block change
Invert contingencies of task
Lesions - like gambling task, impairment in ability, in leanring to adapt to environment

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

what are the ingredients to build an intelligent agent

A

Computational approaches = control everything coming in and out fo system and lean rain g process
Need all various competent to learn/be flexible in envir
What part of leanring process is important to learn a certain behaviour

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

computational approach to cognition

A

agent interact with world
Goal = Maximzie reward
Aspects of environment are represented in diff brain areas
How do observations, rewards, actions change internal model of an agent
What are diff types of leanring in situations where fed observations vs when really interacting with world and receiving reward

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

what is evidence that a behaviour is learned

A

Action, conditioning, habit
Learning of new Jamir
Improving accuracy = signature fo leanring
Is behaviour innate or learned - motor control

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

nature vs nurture

A

Giraffe baby can walk immediately
Human baby can’t - learns over time
Complexity of a behaviour is not an indication of leanring vs innate behaviour - lots of structure exists inn the works that can enable complex innate behaviours

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

implicit vs explicit

A

Learning can be =
Explicit = taught, like read book about physics
Implicit = Jenga, acquire understanding at some point, instinctive - intuitive understanding of physics - gravity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

motor skill learning and memory

A

Recall = HM
Acquired specific motor skill but doesn’t remember = improved over days, based on implicit leanring and doesn’t transfer to other hand

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

learning to see and to navigate

A

Across many domains = many brain processes involved but we cna think about some key concepts that are needed for learning
Is leanring in a given cortex domain general or domain specific (sensory/motor leanring)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

learning to locate a place

A

First trial = rat in pool swimming
If swimming = doesn’t see hidden platform so has to explore space and encounter platform
After 10 trials =built map of space and can travel directly to platform = learn where it is
= learn cognitive map fo space

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

learning to discriminate

A

Have to compare - butterflies
If do task over amd over = will get better

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

what is an essential competent of learning

A

Memeory
Repetition- test leanring
feed back = internal or external
Adaptation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

name diff types of tasks

A

2 main goals = classification or regression
Do them everyday

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

describe classification

A

Each data point has label
Goal of algorithm = infer label of new data points
Is it a or b
Will it snow tmr

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

describe regression

A

Each data point has a value
Goal of algorithm = infer value of new data points
More continuous prediction of value
What will temperature be at 12pm tmr

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

types of learning algorithms

A

Diff types of leanring styles depending on feedback
Distinguished by how data is handled and what signal is for leanring - how is info you have about world processed
How cna they help us understand leanring in the brain - what is type of feedback
What type of learning in diff brain areas at diff stages

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

name the types of learning algorithms

A

Supervised learning
Unsupervised learning
Reinforcement learning

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

describe learning in artificial networks

A

Control data, learning rules
Can observe activity of all the units in the network
Can define task
= build everything, compete control of leanring process and capabilities of system
- can help us think about brain

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

describe learning in biological networks

A

Have some control over data
Don’t control leanring rules
Can observe activity of a very small fraction of units in network - some of neural activity only
Have some control over task
Way harder

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

architecture of convolutional neural network

A

Simple feature map —> object representations
Hierarchical representation
Train on data and learn

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
What Is supervised learning
Each data point associated with label Goal = learn features fo data that predict the label
26
Supervised learning = training
Each image in training set associated with category = cat or dog Algorithm learns to associate each image in dataset with correct label Have to figure our features
27
Supervised learning = test
If algorithm generalizes well = will also associate new unseen images of dogs and cats to correct label Tell if right or wrong so can update model = build up from sensory pathway
28
credit assignment problem
How do we assign credit to each neuron for its contribution to output of network How do we change connectivity in network to improve the accuracy output = complicated How do we assign credit to some change or part of system to output Train = tell it’s wrong - change neural activity to output the correct identity
29
what is algorithmic rule to update weights in network
Back propagation Used in machine learning Governs how weights are changed given the output of the network and the activity of the units Formation to nudge weights towards giving correct output Compute = relative to how strong error is and how influential given neuron is to network Only possible in artificial systems
30
describe back propagation generally
forwards pass of activity = see something and do this, leads to output Backwards pass of errors = if errors then look back into system and nudge it towards correct answer
31
forward pass
Compute output of net work given the input
32
backwards pass
Update weights in network o reduce the error Structure of network
33
is it easy to measure synaptic changes in biological networks ?
NOO HARD Can only measure pairwaise changes of a few neurons Across a few neurons - have potentiation and depression (depreciation) Weights change in simailr way to artificial network More supervised approach Output - update weights to reach output
34
Measure connectivity between neurons
Measure the connectivity change between 2 neurons = ahve to measure activity of both neurons simultaneously and usually intracellularly (membrane potential rather than spikes)
35
what can cause synaptic changes
Can be due to numerous factors = distribution fo receptors, number of synapses All of these are difficult to measure without access to the membrane potential
36
learning a convolutional neural network
Define architecture of the network - number of layers, feature maps, their sizes etc Use training data and back propagation to learn the weights Test accuracy on a test set of data not used during the training Thinking about tehse supervised approaches = cna help understand how we cold build these representations In humans = not super simailr to how you build your own visual representation of objects
37
Does supervised learning apply to classification or regression
Applies to both Independent of type of task Apply same principle to both cases
38
describe ex = dataset of botanist
Contains random samples of 3 species of flowers Each of species = data set contains 50 obsertaosn for septal length, septal width, petal length and width = how do you classify things when data doesn’t have a label - only have where data lays in space Once you look in more dimensions = see a gradient
39
describe unsupervised learning
Data points unlabelled Goal= to find structure in data Usually done by finding statistical regularity that is indicative of an underlying structure Point are clsoe together inn space of parameters should belong to same category Difficulty= man definitions of clsoe together As the data points are not labelled = this definition of distance is central to the results of the algorithm = clustering challenge = mesure of distance
40
describe und supervised learning = specifically
Not given identity of points Classify and identify the diff types in data Data unlabelled- algorithm learns without supervision Goal = find structure in data What is structure tho - or distance within space of parameters No teaching signal = only internal measure of distance of things
41
what type fo learning happens in brain
Unsupervised=experience world and amke associations Only based on features in data Reinformcent leanring = in between, partial feedback
42
predict grades and train system on labelled things = what type fo learning
Supervised
43
Have everyone in class and all courses they ahve taken= what type fo learning
Unsupervised
44
learning so far
Supervised and unsupervised = presented data and agent had to learn either from labelled or unlabelled exs How does agent learn to act in world Some leanring = extremely fast, one shot, like fear/threat conditioning Info = from actions and learn which actions good
45
what is reinforcement leanring
Learn a policy to take actions that maximize rewards given Rewards sparse, most actions don’t lead directly to a reward How to assign credit to actin in the past that led to a reward
46
define reinforcement leaning
Agent receives occasional rewards and must learn how to act to maximize them
47
maze Ex of reinforcement learning
Only info that right decisions as made = at time of exit Once agent find s exit = can reinforce actions that led to finding the right solution After experiencing maze several times = agent can learn best exit strategy
48
animal behaviour exp that inspired irl
a
49
name the 2 diff kinds of learning
Classical/pavlovian Operant conditoning
50
define classical/pavlovian conditioning
Stimulus given No action necessary Value and associated behavioural response is learned No action of animal
51
define operant condition
Learn to associate response with a cue/state of envir to obtain rewards Have to make voluntary action = lie press lever, need to learn action What is being strengthened = different slightly
52
describe Pavlovian condition = graphs
Acquisition = learn, reaches max Extinction = decouple association = have some sort of forgetting First spontaneous recovery = persistence in memory - ltm happens tho Dynamics in leanring
53
describe Pavlovian condition = specifically
Associate intrinsically rewarding stimulus like food (us) with a simtulus that otherwise has no intrinsic value like sound (cs) Measure behavioural response to cs with leanring across diff conditions Bell has no value but when associated with us (food)= condition stimulus acquires value
54
learning from trial and error = rescorla-Wagner rule
Formalize process Makes some predictions in simple classifications conditioning scenarios= predictions in terms of quantitative changes = how much given reward predicts how much you associate Cs to value - more food =stronger association Why saturates at some point = once predicting value correct, if have correct prediction of value = do not need to increase value
55
limits of r-w rule
Based on trial structure and not easily transferred to sequences of actions and states world is more continuous Ex = thorndikes cat = how do you assign reward credit to any of the specific actions that the cat took in whole sequence Hard to assign credit bc many actions - is it the pressing lever, how do we assign credit to previous actions
56
how do solve a simple navigation task
Reinforcement leanring How can mouse in maze learn which actions are more likely to lead to reward = have to work backwards from reward state Driving to restaurant or putting fork in mouth = both have value bc leads to food, in a state that leads to a state that gives rewards Framework = assign value to state as a function of reward obtained in that state bit also the value of states that can be reached from that state
57
agency in world
Reinforcement learning How do you link everything you did before to the feedback and how do you maximize reward
58
Perception
visual representations in ventral stream
59
value functions
value representations in pfc
60
reactive policies
biasing action selection in striatum From perception to decision in dorsal stream - Mt/lip
61
transition model
bayes rule to update world model