Transformers Flashcards

(5 cards)

1
Q

CBOW steps

A

From input vector take embeddings of cont4xt words

Average the context words v’

Multiplying this v’ vectory with output matrix

Get the score and apply softmax

Learn from result and tune the input matrix

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Transposed convolution feature out dimension calculation

A

(i-1)*s+k-2p

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Limitations of one hot encoding

A

Produces sparse, high dimensional vectors and captures no semantic relationship

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

When is TF - IDF vector outperform bag of words

A

When the distinguished power of a word is crucial and common words are noisy.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Trade off between CBoW and Skip gram

A

CBOW fasteand better for frequent words
Skip gram is slower but better for rare words

How well did you know this?
1
Not at all
2
3
4
5
Perfectly