Natural Language Processing (NLP) for Semantic Search

By James Briggs

Learn how to make machines understand language as people do. This free course covers everything you need to build state-of-the-art language models, from machine translation to question-answering, and more.

Natural Language Processing (NLP) for Semantic Search

Introduction

Semantic search has long been a critical component in the technology stacks of giants such as Google, Amazon, and Netflix. The recent democratization of these technologies has ignited a search renaissance, and these once guarded technologies are being discovered and quickly adopted by organizations across every imaginable industry.

Why the explosion of interest in semantic search? It unlocks an essential recipe to many products and applications, the scope of which is unknown but already broad. Search engines, autocorrect, translation, recommendation engines, error logging, and much more are already heavy users of semantic search. Many tools that can benefit from a meaningful language search or clustering function are supercharged by semantic search.

Two pillars support semantic search; vector search and NLP. In this course, we focus on the pillar of NLP and how it brings ‘semantic’ to semantic search. We introduce concepts and theory throughout the course before backing them up with real, industry-standard code and libraries.

You will learn what dense vectors are and why they’re fundamental to NLP and semantic search. We cover how to build state-of-the-art language models covering semantic similarity, multilingual embeddings, question-answering, and more. Learn how to apply these in the real world, where we often lack suitable datasets or masses of computing power.

In short, you will learn everything you need to know to begin applying NLP in your semantic search use-cases.

Let’s begin!

Chapter 1

Dense Vectors

An overview of dense vector embeddings with NLP.

Chapter 2

Sentence Transformers and Embeddings

How sentence transformers and embeddings can be used for a range of semantic similarity applications.

Chapter 3

Training Sentence Transformers with Softmax Loss

The original way of training sentence transformers like SBERT for semantic search.

Chapter 4

Training Sentence Transformers with Multiple Negatives Ranking Loss

How to create sentence transformers by fine-tuning with MNR loss.

Chapter 5

Multilingual Sentence Transformers

How to create multilingual sentence transformers with knowledge distillation.

Chapter 6

Question Answering

Illustrated guide to open domain question-answering.

Chapter 7

Q&A Retriever Models

Fine-tuning a retriever component for open domain question-answering.

Chapter 8

Q&A Reader Models

Fine-tuning a reader component for open domain question-answering.

Chapter 9

Unsupervised Training for Sentence Transformers

How to create sentence transformer models without labelled data.

Chapter 10

Data Augmentation with BERT

Augmented SBERT (AugSBERT) is a training strategy to enhance domain-specific datasets.

Chapter 11

Domain Transfer with BERT

Transfer information from an out-of-domain (or source) dataset to a target domain.

Chapter 12

Training Sentence Transformers

The most popular methods for training sentence transformers, and tips for each.

Chapter 13

Generative Pseudo Labeling (GPL)

The latest method of augmentation and unsupervised training by the creator of Sentence Transformers.

Chapter 14

And more...

What will you build?

Upgrade your search or recommendation systems with just a few lines of code, or contact us for help.