Course Outline

Introduction

  • What are Large Language Models (LLMs)?
  • LLMs vs traditional NLP models
  • Overview of LLMs features and architecture
  • Challenges and limitations of LLMs

Understanding LLMs

  • The lifecycle of an LLM
  • How LLMs work
  • The main components of an LLM: encoder, decoder, attention, embeddings, etc.

Getting Started

  • Setting up the Development Environment
  • Installing an LLM as a development tool, e.g. Google Colab, Hugging Face

Working with LLMs

  • Exploring available LLM options
  • Creating and using an LLM
  • Fine-tuning an LLM on a custom dataset

Text Summarization

  • Understanding the task of text summarization and its applications
  • Using an LLM for extractive and abstractive text summarization
  • Evaluating the quality of the generated summaries using metrics such as ROUGE, BLEU, etc.

Question Answering

  • Understanding the task of question answering and its applications
  • Using an LLM for open-domain and closed-domain question answering
  • Evaluating the accuracy of the generated answers using metrics such as F1, EM, etc.

Text Generation

  • Understanding the task of text generation and its applications
  • Using an LLM for conditional and unconditional text generation
  • Controlling the style, tone, and content of the generated texts using parameters such as temperature, top-k, top-p, etc.

Integrating LLMs with Other Frameworks and Platforms

  • Using LLMs with PyTorch or TensorFlow
  • Using LLMs with Flask or Streamlit
  • Using LLMs with Google Cloud or AWS

Troubleshooting

  • Understanding the common errors and bugs in LLMs
  • Using TensorBoard to monitor and visualize the training process
  • Using PyTorch Lightning to simplify the training code and improve the performance
  • Using Hugging Face Datasets to load and preprocess the data

Summary and Next Steps

Requirements

  • An understanding of natural language processing and deep learning
  • Experience with Python and PyTorch or TensorFlow
  • Basic programming experience

Audience

  • Developers
  • NLP enthusiasts
  • Data scientists
 14 Hours

Number of participants



Price per participant

Related Courses

LangChain: Building AI-Powered Applications

14 Hours

LangChain Fundamentals

14 Hours

Advanced Stable Diffusion: Deep Learning for Text-to-Image Generation

21 Hours

Introduction to Stable Diffusion for Text-to-Image Generation

21 Hours

AlphaFold

7 Hours

TensorFlow Lite for Embedded Linux

21 Hours

TensorFlow Lite for Android

21 Hours

TensorFlow Lite for iOS

21 Hours

Tensorflow Lite for Microcontrollers

21 Hours

Deep Learning Neural Networks with Chainer

14 Hours

Distributed Deep Learning with Horovod

7 Hours

Accelerating Deep Learning with FPGA and OpenVINO

35 Hours

Building Deep Learning Models with Apache MXNet

21 Hours

Deep Learning with Keras

21 Hours

Advanced Deep Learning with Keras and Python

14 Hours

Related Categories

1