Receive in-depth knowledge from industry professionals, test your skills with hands-on assignments & demos, and get access to valuable resources and tools.
This course is a deep dive into Natural Language Processing (NLP). The lessons that are presented here focus on NLP applications such as sentiment analysis, feature extraction, and several models for NLP, including the latest state-of-the-art algorithms such as BERT. After this course, you will gain hands-on experience in semi-supervised and unsupervised machine learning methods for NLP as well as theoretical understanding of the concepts and models. As requirements, experience with python and machine learning are needed.
Are you interested? Contact us and we will get in touch with you.
The NLP training is split in 3 days. Click below to see a detailed description of each class:
In the first NLP class of the series, we will teach you the foundations needed to analyze linguistic data and understand basic Natural Language Processing concepts.
The training starts with a discussion about the challenges of linguistic data, followed by techniques of handling, cleaning, and normalizing text data. The lesson concludes with two language models: *N-grams* and *word embeddings*. The latter model will be discussed further in NLP 2, as it requires a better understanding of RNNs and Deep Learning.
The theoretical lesson is followed by a few lab exercises where participants get familiarized with main NLP toolkits used in the industry (NLTK, Spacy, Gensim) and train a Bayesian model to predict the author’s gender using word frequency features from Twitter data.
The training includes theory, demos, and hands-on-exercises.
By the end of the training participants will have gained knowledge about:
The second class in the NLP series will provide you with an overview of the most widely-used NLP models currently and how to implement them as part of a Machine Learning pipeline with text data.
The training starts with a survey of a few relevant NLP models, from simple ones like *Bag of Words*, through RNN-based ones like *word2vec, fastText, and ELMo*, through the more recent transformers, such as *BERT*. The theoretical NLP model survey is followed by two labs where participants utilize various NLP models to retrain on new tasks: author profiling and sentiment analysis.
The training includes theory, demos, and hands-on-exercises.
By the end of the training participants will have gained gained knowledge about the characteristics of the following language models:
In the last NLP class of the series we will teach you how to use semi-supervised and unsupervised Machine Learning methods by working on NLP tasks and text data as a case study.
The class starts with a discussion about the problem of scarce data in NLP, especially for languages other than English. We discuss solutions for this problem — and indeed the problem of shortage of good data in general — namely *pre-training, self-training* and *consistency regulation*.
The lesson ends with the introduction of *LDA (Latent Dirichlet Allocation)*, an unsupervised learning model useful in topic modeling. The theoretical lesson of semi-supervised and unsupervised learning is followed by a lab exercise where participants utilize an *LDA* model to extract features from the text that are then used to perform author profiling (the same task used throughout the NLP sequence).
The training includes theory, demos, and hands-on-exercises.
By the end of the training participants will have gained knowledge about the characteristics of the following semi-supervised and unsupervised methods:
info.nl@eraneos.com
KvK nr. 02070702
BTWnr. NL813364103B01
Terms and Conditions
Privacy Statement (English)
Privacy Verklaring (Dutch)