This course provides you with an overview of how to use transformer-based models for natural language processing (NLP).
This course provides you with an overview of how to use transformer-based models for natural language processing (NLP).
In this course, you will learn to apply transformer-based models for text classification, focusing on the encoder component.
You’ll learn about positional encoding, word embedding, and attention mechanisms in language transformers and their role in capturing contextual information and dependencies.
Additionally, you will be introduced to multi-head attention and gain insights on decoder-based language modeling with generative pre-trained transformers (GPT) for language translation, training the models, and implementing them in PyTorch.
Further, you’ll explore encoder-based models with bidirectional encoder representations from transformers (BERT) and train using masked language modeling (MLM) and next sentence prediction (NSP).
Finally, you will apply transformers for translation by gaining insight into the transformer architecture and performing its PyTorch implementation.
The course offers practical exposure with hands-on activities that enables you to apply your knowledge in real-world scenarios.
This course is part of a specialized program tailored for individuals interested in Generative AI engineering.
This course requires a working knowledge of Python, PyTorch, and machine learning.
OpenCourser helps millions of learners each year. People visit us to learn workspace skills, ace their exams, and nurture their curiosity.
Our extensive catalog contains over 50,000 courses and twice as many books. Browse by search, by topic, or even by career interests. We'll match you to the right resources quickly.
Find this site helpful? Tell a friend about us.
We're supported by our community of learners. When you purchase or subscribe to courses and programs or purchase books, we may earn a commission from our partners.
Your purchases help us maintain our catalog and keep our servers humming without ads.
Thank you for supporting OpenCourser.