We may earn an affiliate commission when you visit our partners.
Course image
Course image
Coursera logo

Generative AI and LLMs

Architecture and Data Preparation

Joseph Santarcangelo and Roodra Pratap Kanwar

This IBM short course, a part of Generative AI Engineering Essentials with LLMs Professional Certificate, will teach you the basics of using generative AI and Large Language Models (LLMs). This course is suitable for existing and aspiring data scientists, machine learning engineers, deep-learning engineers, and AI engineers.

Read more

This IBM short course, a part of Generative AI Engineering Essentials with LLMs Professional Certificate, will teach you the basics of using generative AI and Large Language Models (LLMs). This course is suitable for existing and aspiring data scientists, machine learning engineers, deep-learning engineers, and AI engineers.

You will learn about the types of generative AI and its real-world applications. You will gain the knowledge to differentiate between various generative AI architectures and models, such as Recurrent Neural Networks (RNNs), Transformers, Generative Adversarial Networks (GANs), Variational AutoEncoders (VAEs), and Diffusion Models. You will learn the differences in the training approaches used for each model. You will be able to explain the use of LLMs, such as Generative Pre-Trained Transformers (GPT) and Bidirectional Encoder Representations from Transformers (BERT).

You will also learn about the tokenization process, tokenization methods, and the use of tokenizers for word-based, character-based, and subword-based tokenization. You will be able to explain how you can use data loaders for training generative AI models and list the PyTorch libraries for preparing and handling data within data loaders. The knowledge acquired will help you use the generative AI libraries in Hugging Face. It will also prepare you to implement tokenization and create an NLP data loader.

For this course, a basic knowledge of Python and PyTorch and an awareness of machine learning and neural networks would be an advantage, though not strictly required.

Enroll now

What's inside

Syllabus

Generative AI Architecture
In this module, you will learn about the significance of generative AI models and how they are used across a wide range of fields for generating various types of content. You will learn about the architectures and models commonly used in generative AI and the differences in the training approaches of these models. You will learn how large language models (LLMs) are used to build NLP-based applications. You will build a simple chatbot using the transformers library from Hugging Face.
Read more
Data Preparation for LLMs
In this module, you will learn to prepare data for training large language models (LLMs) by implementing tokenization. You will learn about the tokenization methods and the use of tokenizers. You will also learn about the purpose of data loaders and how you can use the DataLoader class in PyTorch. You will implement tokenization using various libraries such as nltk, spaCy, BertTokenizer, and XLNetTokenizer. You will also create a data loader with a collate function that processes batches of text.

Good to know

Know what's good
, what to watch for
, and possible dealbreakers
Suitable for existing and aspiring data scientists, machine learning engineers, deep learning engineers, and AI engineers
Taught by Joseph Santarcangelo and Roodra Pratap Kanwar, who are recognized for their work in NLP
Part of Generative AI Engineering Essentials with LLMs Professional Certificate
Provides hands-on experience with Generative AI libraries in Hugging Face

Save this course

Save Generative AI and LLMs: Architecture and Data Preparation to your list so you can find it easily later:
Save

Activities

Be better prepared before your course. Deepen your understanding during and after it. Supplement your coursework and achieve mastery of the topics covered in Generative AI and LLMs: Architecture and Data Preparation with these activities:
Review Python and PyTorch basics
Reviewing these topics will help you refresh your understanding of the core concepts used in this course.
Browse courses on Python
Show steps
  • Go through online tutorials or documentation on Python and PyTorch.
  • Complete practice exercises or coding challenges to test your understanding.
Mentor a junior student in generative AI
Mentoring others will reinforce your understanding of generative AI and help you develop your communication and leadership skills.
Browse courses on Generative AI
Show steps
  • Reach out to junior students or peers who are interested in generative AI.
  • Provide guidance on topics such as model selection, training techniques, and evaluation metrics.
  • Review their code, provide feedback, and help them troubleshoot any issues.
Attend a workshop on generative AI applications
Attending a workshop will expose you to real-world use cases and industry best practices in generative AI.
Browse courses on Generative AI
Show steps
  • Research and identify upcoming workshops on generative AI.
  • Register for a workshop that aligns with your interests and learning goals.
  • Attend the workshop, actively participate in discussions, and take notes.
Four other activities
Expand to see all activities and additional details
Show all seven activities
Follow tutorials on generative AI models
Following tutorials will provide you with hands-on experience in implementing and training generative AI models.
Browse courses on Generative AI
Show steps
  • Identify relevant tutorials from reputable sources such as Coursera, edX, or YouTube.
  • Go through the tutorials step-by-step, implementing the code and experimenting with different parameters.
  • Document your learnings and any challenges you faced during the tutorials.
Practice tokenization and data loading techniques
Practicing these techniques will improve your proficiency in preparing data for generative AI models.
Browse courses on Tokenization
Show steps
  • Find datasets suitable for generative AI tasks, such as text generation or image synthesis.
  • Implement tokenization methods using libraries like spaCy or Hugging Face's Tokenizers.
  • Create data loaders using PyTorch's DataLoader class to handle data efficiently.
  • Experiment with different tokenization and data loading parameters to optimize your models.
Develop a research proposal on a generative AI topic
Writing a research proposal will challenge you to think critically about generative AI's potential and applications.
Browse courses on Generative AI
Show steps
  • Identify a specific research question or problem you want to address.
  • Review existing literature and research on generative AI.
  • Develop a research methodology and plan for your proposed study.
  • Write a research proposal outlining your research question, methods, and expected outcomes.
Build a simple NLP application using a pre-trained language model
Building an NLP application will allow you to apply your understanding of generative AI in a practical context.
Browse courses on NLP
Show steps
  • Choose a specific NLP task, such as text classification or question answering.
  • Select a pre-trained language model from Hugging Face or other providers.
  • Fine-tune the model on your dataset using transfer learning techniques.
  • Deploy your NLP application and evaluate its performance.

Career center

Learners who complete Generative AI and LLMs: Architecture and Data Preparation will develop knowledge and skills that may be useful to these careers:

Reading list

We haven't picked any books for this reading list yet.

Share

Help others find this course page by sharing it with your friends and followers:

Similar courses

Here are nine courses similar to Generative AI and LLMs: Architecture and Data Preparation.
LLMs Mastery: Complete Guide to Transformers & Generative...
Most relevant
NVIDIA-Certified Associate - Generative AI LLMs (NCA-GENL)
Most relevant
Exploring Generative AI Models and Architecture
Most relevant
Generative AI: Foundation Models and Platforms
Most relevant
Models and Platforms for Generative AI
Most relevant
Mathematics Behind Large Language Models and Transformers
Most relevant
Understanding Large Language Models in Business
Most relevant
Introduction to Generative AI
Most relevant
Quantization Fundamentals with Hugging Face
Most relevant
Our mission

OpenCourser helps millions of learners each year. People visit us to learn workspace skills, ace their exams, and nurture their curiosity.

Our extensive catalog contains over 50,000 courses and twice as many books. Browse by search, by topic, or even by career interests. We'll match you to the right resources quickly.

Find this site helpful? Tell a friend about us.

Affiliate disclosure

We're supported by our community of learners. When you purchase or subscribe to courses and programs or purchase books, we may earn a commission from our partners.

Your purchases help us maintain our catalog and keep our servers humming without ads.

Thank you for supporting OpenCourser.

© 2016 - 2024 OpenCourser