Learning Vector Space Models with SpaCy

Video description

Information representation is a fundamental aspect of computational linguistics and learning from unstructured data. This course explores vector space models, how they're used to represent the meaning of words and documents, and how to create them using Python-based spaCy. You'll learn about several types of vector space models, how they relate to each other, and how to determine which model is best for natural language processing applications like information retrieval, indexing, and relevancy rankings.

The course begins with a look at various encodings of sparse document-term matrices, moves on to dense vector representations that need to be learned, touches on latent semantic analysis, and finishes with an exploration of representation learning from neural network models with a focus on word2vec and Gensim. To get the most out of this course, learners should have intermediate level Python skills.

  • Understand how and why vector models are used in natural language processing
  • Discover the distributional hypothesis and its use in word and document vectors
  • Explore term-document tf-idf, latent semantic analysis, and neural embedding models
  • Gain experience integrating neural embedding models with spaCy
Aaron Kramer is a data scientist and engineer with Los Angeles based DataScience Inc. He is a spaCY contributor who holds a BA in Economics from Swarthmore College and is the author of multiple O'Reilly titles on the subject of natural language processing.

Product information

  • Title: Learning Vector Space Models with SpaCy
  • Author(s): Aaron Kramer
  • Release date: March 2017
  • Publisher(s): Infinite Skills
  • ISBN: 9781491986035