• Stars
    star
    211
  • Rank 186,867 (Top 4 %)
  • Language
    Jupyter Notebook
  • Created almost 4 years ago
  • Updated over 3 years ago

Reviews

There are no reviews yet. Be the first to send feedback to the community and the maintainers!

Repository Details

Build and train state-of-the-art natural language processing models using BERT

Getting started with Google BERT

Build and train state-of-the-art natural language processing models using BERT

About the book

Book Cover

BERT (bidirectional encoder representations from transformer) has revolutionized the world of natural language processing (NLP) with promising results. This book is an introductory guide that will help you get to grips with Google's BERT architecture. With a detailed explanation of the transformer architecture, this book will help you understand how the transformer's encoder and decoder work.

You'll explore the BERT architecture by learning how the BERT model is pre-trained and how to use pre-trained BERT for downstream tasks by fine-tuning it for NLP tasks such as sentiment analysis and text summarization with the Hugging Face transformers library. As you advance, you'll learn about different variants of BERT such as ALBERT, RoBERTa, and ELECTRA, and look at SpanBERT, which is used for NLP tasks like question answering. You'll also cover simpler and faster BERT variants based on knowledge distillation such as DistilBERT and TinyBERT.

The book takes you through MBERT, XLM, and XLM-R in detail and then introduces you to sentence-BERT, which is used for obtaining sentence representation. Finally, you'll discover domain-specific BERT models such as BioBERT and ClinicalBERT, and discover an interesting variant called VideoBERT.

Get the book


Clone the repo and run in Google Colab

1. A Primer on Transformer

2. Understanding the BERT model

  • 2.1. Basic idea of BERT
  • 2.2. Working of BERT
  • 2.3. Configuration of BERT
  • 2.4. Pre-training the BERT
  • 2.5. Pre-training strategies
  • 2.6. Pre-training procedure
  • 2.7. Subword tokenization algorithms
  • 2.8. Byte pair encoding
  • 2.9. Byte-level byte pair encoding
  • 2.10. WordPiece

3. Getting hands-on with BERT

4. BERT variants I - ALBERT, RoBERTa, ELECTRA, SpanBERT

5. BERT variants II - Based on knowledge distillation

  • 5.1. Knowledge distillation
  • 5.2. DistilBERT - distilled version of BERT
  • 5.3. Training the DistilBERT
  • 5.4. TinyBERT
  • 5.5. Teacher-student architecture
  • 5.6. Training the TinyBERT
  • 5.7. Transferring knowledge from BERT to neural network
  • 5.8. Teacher-student architecture
  • 5.9. Training the student network
  • 5.10. Data augmentation method

6. Exploring BERTSUM for text summarization

  • 6.1. Text summarization
  • 6.2. Fine-tuning BERT for text summarization
  • 6.3. Extractive summarization using BERT
  • 6.4. Abstractive summarization using BERT
  • 6.5. Understanding ROUGE evaluation metric
  • 6.6. Performance of BERTSUM model
  • 6.7. Training the BERTSUM model

7. Applying BERT for other languages

8. Exploring Sentence and Domain Specific BERT

9. Understanding VideoBERT, BART, and more

More Repositories

1

Awesome-Meta-Learning

A curated list of Meta Learning papers, code, books, blogs, videos, datasets and other resources.
1,479
star
2

Hands-On-Meta-Learning-With-Python

Learning to Learn using One-Shot Learning, MAML, Reptile, Meta-SGD and more with Tensorflow
Jupyter Notebook
1,157
star
3

Hands-On-Reinforcement-Learning-With-Python

Master Reinforcement and Deep Reinforcement Learning using OpenAI Gym and TensorFlow
Jupyter Notebook
834
star
4

Hands-On-Deep-Learning-Algorithms-with-Python

Master Deep Learning Algorithms with Extensive Math by Implementing them using TensorFlow
Jupyter Notebook
327
star
5

Deep-Reinforcement-Learning-With-Python

Master classic RL, deep RL, distributional RL, inverse RL, and more using OpenAI Gym and TensorFlow with extensive Math
Jupyter Notebook
322
star
6

Bitcoin-price-Prediction-using-LSTM

Bitcoin price Prediction ( Time Series ) using LSTM Recurrent neural network
Jupyter Notebook
124
star
7

Word2vec-from-scratch

simple Word2vec from scratch using tensorflow for understanding
Jupyter Notebook
31
star
8

Whatsapp-analytics

performing sentiment analysis on the whatsapp chats.
R
21
star
9

Document-Classification-using-LSA

Document classification using Latent semantic analysis in python
Jupyter Notebook
16
star
10

Recommendation-System

Simple movie recommendation system using lighfm in python.
Python
2
star
11

Face-Detection-webcam

Face Detection in the webcam using opencv
Python
2
star
12

Tagging-Stackoverflow-posts-using-BOW

Tagging Stackoverflow posts using BOW in keras
Jupyter Notebook
2
star
13

SVM-hyperparameters-in-R-shiny

R shiny app to understand various hyperparameter in SVM
R
1
star
14

Cryptography-Algorithms

Implementation of Various Cryptographic algorithms
Java
1
star