Provides an overview of Natural Language Processing, focusing on transformers, tokenization, and self-attention mechanisms for effective language analysis and synthesis.
Introduces Natural Language Processing (NLP) and its applications, covering tokenization, machine learning, sentiment analysis, and Swiss NLP applications.
Explores word embeddings, models like CBOW and Skipgram, Fasttext, Glove, subword embeddings, and their applications in document search and classification.
Explores methods for information extraction, including traditional and embedding-based approaches, supervised learning, distant supervision, and taxonomy induction.