Give some text pre-processing methods.
What are sparse and dense feature vectors?
One hot-vectors (sparse):
Each feature is its own dimension, dim of vector = no of features, features are independent.
Dense vectors:
Each feature is a d-dim vector, similar features have similar vectors, provide better generalization
What is a word embedding?
Word embedding is the collective name for a set of language modeling and feature learning techniques in natural language processing (NLP) where words or phrases from the vocabulary are mapped to vectors of real numbers. Conceptually it involves a mathematical embedding from a space with one dimension per word to a continuous vector space with a much lower dimension.
Describe three types of ambiguities that make NLP complex.