Natural Language Processing and LLMs

Type product

Natural Language Processing and LLMs

OEM ICT Trainingen & Advies
Logo van OEM ICT Trainingen & Advies
Opleiderscore: starstarstarstarstar_half 9,0 OEM ICT Trainingen & Advies heeft een gemiddelde beoordeling van 9,0 (uit 223 ervaringen)

Tip: meer info over het programma, prijs, en inschrijven? Download de brochure!

Beschrijving

Natural Language Processing and LLMs.

Embark on an enriching journey into the world of Natural Language Processing (NLP) and Large Language Models (LLMs) with our comprehensive track series. Beginning with the "Fundamentals of Natural Language Processing," participants will build a solid foundation in NLP techniques, mastering text preprocessing, representation, and classification. Moving forward, the "Natural Language Processing with Deep Learning" track delves into advanced deep learning methodologies for NLP tasks, while the "Natural Language Processing with LLMs" track pushes the boundaries with cutting-edge LLMs, attention mechanisms, and transformer architectures. From understanding a…

Lees de volledige beschrijving

Veelgestelde vragen

Er zijn nog geen veelgestelde vragen over dit product. Als je een vraag hebt, neem dan contact op met onze klantenservice.

Nog niet gevonden wat je zocht? Bekijk deze onderwerpen: ChatGPT, Microsoft Azure, Business intelligence, Unified Modeling Language (UML) en Blockchain.

Natural Language Processing and LLMs.

Embark on an enriching journey into the world of Natural Language Processing (NLP) and Large Language Models (LLMs) with our comprehensive track series. Beginning with the "Fundamentals of Natural Language Processing," participants will build a solid foundation in NLP techniques, mastering text preprocessing, representation, and classification. Moving forward, the "Natural Language Processing with Deep Learning" track delves into advanced deep learning methodologies for NLP tasks, while the "Natural Language Processing with LLMs" track pushes the boundaries with cutting-edge LLMs, attention mechanisms, and transformer architectures. From understanding attention mechanisms to implementing state-of-the-art LLMs for tasks like language translation and text summarization, this journey equips participants with the knowledge and skills to navigate the forefront of NLP innovation.

This Learning Kit with more than 21 hours of learning is divided into three tracks:

Course content

Track 1: Natural Language Processing

This track provides a comprehensive introduction to the core concepts and techniques in NLP. Beginning with an overview of NLP components, including natural language understanding (NLU) and natural language generation (NLG), the track explores common NLP tasks such as speech recognition and sentiment analysis. Participants will then delve into preprocessing text data using NLTK, covering essential techniques such as text cleaning, sentence segmentation, and parts-of-speech tagging. Additionally, the track explores methods for representing text in numeric format, including one-hot encoding and TF-IDF encoding, before introducing classification models for text data. Through hands-on exercises and practical examples, participants will learn how to build classification models using rule-based approaches, Naive Bayes classification, and other techniques, leveraging tools like Scikit-learn pipelines and grid search for optimal performance. Participants will then harness the power of TensorFlow for building deep learning models, followed by an in-depth exploration of text preprocessing techniques such as normalization, tokenization, and text vectorization. Through hands-on exercises, learners will delve into the intricacies of modeling building, training, and evaluation for text classification tasks, encompassing binary classification and multi-class classification using dense neural networks, recurrent neural networks (RNNs), and RNNs with LSTM cells. The track will also cover hyperparameter tuning using the Keras tuner to optimize model performance. Participants will gain proficiency in leveraging word embeddings, including training embedding layers in models, exploring and visualizing embeddings, and utilizing embeddings for tasks like word and semantic similarity. Moreover, the track will explore text translation using RNNs and demonstrate the utilization of pre-trained models for semantic textual similarity, providing participants with a comprehensive understanding of cutting-edge NLP techniques in the context of deep learning.

Courses:

Fundamentals of NLP: Introducing Natural Language Processing

Course: 48 Minutes

  • Course Overview
  • Introducing Natural Language Processing (NLP) with NLTK and spaCy
  • Text Preprocessing for Natural Language Processing
  • Setting up the Environment and Installing NLP Libraries
  • Exploring the Gutenberg and Brown NLTK Corpora
  • Course Summary

Fundamentals of NLP: Preprocessing Text Using NLTK and SpaCy

Course: 1 Hour, 57 Minutes

  • Course Overview
  • Implementing Word and Sentence Tokenization with NLTK
  • Implementing Word and Sentence Tokenization Using SpaCy
  • Performing Stop Word Removal Using NLTK
  • Performing Stopword Removal Using SpaCy
  • Understanding WordNet Synsets
  • Computing Word Similarity Using WordNet
  • Understanding Hypernyms, Hyponyms, Antonyms, Meronyms, and Holonyms
  • Performing Stemming Using NLTK
  • Performing Lemmatization Using NLTK
  • Performing Lemmatization Using SpaCy
  • Performing Parts of Speech Tagging and Named Entity Recognition
  • Course Summary

Fundamentals of NLP: Rule-based Models for Sentiment Analysis

Course: 46 Minutes

  • Course Overview
  • Sentiment Analysis Introduction
  • Loading and Understanding Review Data
  • Cleaning and Visualizing Review Data
  • Performing Sentiment Analysis Using VADER
  • Performing Sentiment Analysis Using TextBlob
  • Course Summary

Fundamentals of NLP: Representing Text as Numeric Features

Course: 2 Hours

  • Course Overview
  • One-hot Encoding to Represent Text in Numeric Form
  • Utilizing One-hot Encoding to Represent Text Data
  • Performing One-hot Encoding Using the Count Vectorizer
  • Frequency-based Encodings to Represent Text in Numeric Form
  • Perform Count Vector Encoding Using the Count Vectorizer
  • Working with Bag-of-Words and Bag-of-N-grams Representation
  • Perform TF-IDF Encoding to Represent Text Data
  • Exploring the Product Reviews Dataset
  • Building a Classification Model Using Count Vector Encoding
  • Comparing Models Trained with Stemmed Words and Stopword Removed
  • Classifying Text Using Frequency Filtering and TF-IDF Encodings
  • Training Classification Models Using Bag of N-grams
  • Training Classification Models with N-grams and TF-IDF Representation
  • Course Summary

Fundamentals of NLP: Word Embeddings to Capture Relationships in Text

Course: 1 Hour, 1 Minutes

  • Course Overview
  • Word Embeddings to Represent Text in Numeric Form
  • Generating Word2Vec Embeddings
  • Training a Classification Model Using Word2Vec Embeddings
  • Working with Pre-trained GloVe Embeddings
  • Training a Classification Model Using GloVe Embeddings
  • Training Different Classification Models for Sentiment Analysis
  • Course Summary

Natural Language Processing Using Deep Learning

Course: 1 Hour, 55 Minutes

  • Course Overview
  • Deep Learning with TensorFlow and Keras
  • Loading and Exploring a Text Dataset
  • Cleaning and Visualizing Data
  • Generating Count Vector Representations
  • Training a Deep Neural Network (DNN) Classification Model
  • TF-IDF Representations Using the TextVectorization Layer
  • Training a DNN Using TF-IDF Vectors
  • Visualizing the Results of TensorFlow Callbacks
  • Loading and Preprocessing Data for Sentiment Analysis
  • Training a DNN Using Word Embeddings
  • Training a DNN Using Pretrained GloVe Word Embeddings
  • Using a Convolutional Neural Network (CNN) for Sentiment Analysis
  • Course Summary

Using Recurrent Networks For Natural Language Processing

Course: 1 Hour, 15 Minutes

  • Course Overview
  • Recurrent Neural Networks (RNNs) for Sequential Data
  • Visualizing Word Embeddings Using the Embedding Projector Plug-in
  • Setting up Word Vector Representations for Training
  • Training an RNN for Sentiment Analysis
  • Training an RNN with LSTM and Bidirectional LSTM Layers
  • Performing Hyperparameter Tuning
  • Course Summary

Using Out-of-the-Box Transformer Models for Natural Language Processing

Course: 1 Hour, 30 Minutes

  • Course Overview
  • Transfer Learning
  • Using Pre-trained Embeddings from the TensorFlow Hub
  • Attention-based Models and Transformers
  • Performing Subword Tokenization with WordPiece
  • Using the FNet Encoder for Sentiment Analysis
  • Using the Universal Sentence Encoder (USE) for Semantic Textual Similarity
  • Structuring Data for Sentence Similarity Prediction Using BERT
  • Using a Fine-tuned BERT Model for Sentence Classification
  • Course Summary

Attention-based Models and Transformers for Natural Language Processing

Course: 2 Hours, 20 Minutes

  • Course Overview
  • Language Translation Models and Attention
  • Preparing Data for Language Translation
  • Configuring the Encoder-Decoder Architecture
  • Defining the Loss and Accuracy for the Translation Model
  • Training Validation and Prediction Using Encoder and Decoder
  • Setting up the Decoder Architecture with Attention Layer
  • Generating Translations Using the Attention Model
  • The Transformer Architecture: Part I
  • The Transformer Architecture: Part II
  • Using Query, Key, and Value in the Attention Mechanism
  • Structuring Translations for Input to a Transformer Model
  • Setting up the Encoder and Decoder in the Transformer Architecture
  • Training the Transformer Model and Using It for Predictions
  • Course Summary

Track 2: Architecting LLM for your Technical solutions

This track is designed to immerse participants in the transformative world of Large Language Models (LLMs), leveraging state-of-the-art techniques powered by deep learning and attention mechanisms. Participants will gain a deep understanding of attention mechanisms and the revolutionary transformer architecture, including self-attention and multi-head attention mechanisms. Through hands-on exercises and practical demonstrations, learners will explore the foundational concepts of LLMs and delve into implementing translation models using transformers. Moreover, participants will be introduced to the Hugging Face platform, learning to leverage pre-trained models from the Hugging Face library and fine-tune them for specific use cases. From text classification to language translation, question answering, text summarization, and natural language generation, participants will acquire the skills needed to harness the full potential of LLMs for a wide range of NLP tasks.

Courses:

NLP with LLMs: Working with Tokenizers in Hugging Face

Course: 2 Hours, 18 Minutes

  • Course Overview
  • Hugging Face Introduction
  • Hugging Face Tokenizers
  • Exploring the Hugging Face Platform
  • Setting up the Colab Environment
  • Normalizers and Pre-tokenizers
  • Byte Pair Encoding (BPE), Wordpiece, and Unigram Tokenization
  • Implementing Byte Pair Encoding Tokenization - I1
  • Implementing Byte Pair Encoding Tokenization - II1
  • Implementing Wordpiece Tokenization - I1
  • Implementing Wordpiece Tokenization - II
  • Building and Training a BPE Tokenizer
  • Configuring the Normalizer and Pre-tokenizer for Wordpiece Tokenization
  • Building and Training a Wordpiece Tokenizer
  • Course Summary

NLP with LLMs: Hugging Face Classification, QnA, & Text Generation Pipelines

Course: 1 Hour, 50 Minutes

  • Course Overview
  • Hugging Face Pipelines
  • Performing Zero-shot Classification
  • Performing Sentiment Analysis Using DistilBERT
  • Detecting Emotion and Sentiment Analysis on Financial Data
  • Performing Named Entity Recognition (NER) with a Fine-tuned BERT Model
  • Performing Named Entity Recognition Using Tokenizer and Model
  • Performing Question Answering Using Pipelines
  • Performing Question Answering Using Tokenizer and Model
  • Performing Greedy Search and Beam Search for Text Generation Using GPT
  • Generating Text Using Sampling
  • Performing Mask Filling Using Variations of the BERT Model
  • Course Summary

NLP with LLMs: Language Translation, Summarization, & Semantic Similarity

Course: 1 Hour, 29 Minutes

  • Course Overview
  • Performing Language Translation Using Two Variants of the T5 Model
  • Performing Language Translation Using the M2M 100 and Opus Models
  • Summarizing Text Using a BART Model and a T5 Model
  • Loading Data and Cleaning Text for Summarization
  • Evaluating Summaries Using ROUGE Scores
  • Computing Semantic Textual Similarity Using Sentence Transformers
  • Performing Clustering Using Sentence Embeddings
  • Computing Embeddings and Similarity Using the Tokenier and Model
  • Course Summary

NLP with LLMs: Fine-tuning Models for Classification & Question Answering

Course: 1 Hour, 34 Minutes

  • Course Overview
  • Loading Data and Creating a Dataset for Fine-tuning
  • Setting up for Fine-tuning a BERT Classifier
  • Fine-tuning a BERT Model and Pushing to Hugging Face Hub
  • Getting Predictions from a Fine-tuned Model
  • Structuring Text for Named Entity Recognition
  • Aligning NER Tags to Match Subword Tokenization
  • Fine-tuning a BERT Model for Named Entity Recognition
  • Dealing with Long Contexts for Question Answering
  • Structuring QnA Data in the Right Format for Fine Tuning
  • Fine-tuning a DistilBERT Model for Question Answering
  • Course Summary

NLP with LLMs: Fine-tuning Models for Language Translation, & Summarization

Course: 1 Hour, 38 Minutes

  • Course Overview
  • Processing and Structuring Data for Causal Language Modeling (CLM)
  • Fine-tuning a DistilGPT-2 Model for Causal Language Modeling
  • Fine-tuning a DistilRoBERTa Model for Masked Language Modeling (MLM)
  • Preparing the Translation Data for Fine-tuning
  • Preprocessing Text and Computing Metrics for Translation
  • Fine-tuning the T5-small Model for English to Spanish Translation
  • Loading and Visualizing Summarization Data
  • Evaluating the Baseline Performance of the Pretrained T5-small Model
  • Fine-tuning the T5-small Model for Summarization
  • Comparing the Fine-tuned Model's Performance with the Baseline Model
  • Course Summary

Assessment:

  • Final Exam: Architecting LLMs for Your Technical Solutions

Specificaties

Taal: Engels
Kwalificaties van de Instructeur: Gecertificeerd
Cursusformaat en Lengte: Lesvideo's met ondertiteling, interactieve elementen en opdrachten en testen
Lesduur: 21:35 uur
Assesments: De assessment test uw kennis en toepassingsvaardigheden van de onderwerpen uit het leertraject. Deze is 365 dagen beschikbaar na activering.
Online Virtuele labs: Ontvang 12 maanden toegang tot virtuele labs die overeenkomen met de traditionele cursusconfiguratie. Actief voor 365 dagen na activering, beschikbaarheid varieert per Training.
Online mentor: U heeft 24/7 toegang tot een online mentor voor al uw specifieke technische vragen over het studieonderwerp. De online mentor is 365 dagen beschikbaar na activering, afhankelijk van de gekozen Learning Kit.
Voortgangsbewaking: Ja
Toegang tot Materiaal: 365 dagen
Technische Vereisten: Computer of mobiel apparaat, Stabiele internetverbindingen Webbrowserzoals Chrome, Firefox, Safari of Edge.
Support of Ondersteuning: Helpdesk en online kennisbank 24/7
Certificering: Certificaat van deelname in PDF formaat
Prijs en Kosten: Cursusprijs zonder extra kosten
Annuleringsbeleid en Geld-Terug-Garantie: Wij beoordelen dit per situatie
Award Winning E-learning: Ja


Tip! Zorg voor een rustige leeromgeving, tijd en motivatie, audioapparatuur zoals een koptelefoon of luidsprekers voor audio, accountinformatie zoals inloggegevens voor toegang tot het e-learning platform.

Verrijk Uw Carrière met OEM's ICT Trainingen

Waarom kiezen voor OEM?
Ervaring: Meer dan 20 jaar expertise in ICT-trainingen.
Uitgebreide Selectie: Meer dan 1000 cursussen van 200 topmerken.
Hoge Tevredenheid: Beoordeeld met een 9.0 op Springest.
Kwaliteitsgarantie: Gecertificeerde docenten en award-winning E-learning.
Partnerschappen: Microsoft Partner, EC-Council Partner, Certiport en Pearson VUE.

Blijf op de hoogte van nieuwe ervaringen

Er zijn nog geen ervaringen.

Deel je ervaring

Heb je ervaring met deze cursus? Deel je ervaring en help anderen kiezen. Als dank voor de moeite doneert Springest € 1,- aan Stichting Edukans.

Er zijn nog geen veelgestelde vragen over dit product. Als je een vraag hebt, neem dan contact op met onze klantenservice.

Download gratis en vrijblijvend de informatiebrochure

(optioneel)
(optioneel)
(optioneel)
(optioneel)
We slaan je gegevens op, en delen ze met OEM ICT Trainingen & Advies, om je via e-mail en evt. telefoon verder te helpen. Meer info vind je in ons privacybeleid.