Take Stanford's Natural Language Understanding For Free
Written by Nikos Vaggalis   
Friday, 04 March 2022

The content of CS224u Natural Language Understanding by Stanford University has been made available in a self-paced version and to anyone for free.

A few months ago I covered its sister course CS224n Natural Language Processing with Deep Learning. What is the difference you say between CS224n and CS224u? The answer lies in the following diagram:

As you can see NLU is a subset of NLP which might be considered a prerequisite before tackling NLP. As per Wikipedia's definition:

Natural-language understanding (NLU) or natural-language interpretation is a subtopic of natural-language processing in artificial intelligence that deals with machine reading comprehension.

There is considerable commercial interest in the field because of its application to automated reasoning, machine translation, question answering, news-gathering, text categorization, voice-activation, archiving, and large-scale content analysis.

Specifically NLU is used today in products like

  • voice-driven assistants
  • natural-language search
  • question answering
  • sentiment analysis for automated trading
  • business intelligence
  • social media analytics
  • content summarization

This includes Amazon Alexa, Google Kome Assistant, Cortana and Siri, to name just a few implementations.

In this course students will learn how to develop systems and algorithms for the robust machine understanding of human language, and specifically:

  • Lexical semantics
  • Distributed representations of meaning
  • Relation extraction
  • Semantic parsing
  • Sentiment analysis
  • Dialogue agents

The online course's version is an adapted version of the Spring 2021 course which was taught on campus by an all star teaching team spearheaded by Professor Christopher Potts. Beside the lectures, it is heavy on projects and homework. 

The course begins by covering a wide range of models for distributed word representation and then moves onto topics like relation extraction, natural language inference, and grounded language understanding;important topics which highlight many of the central concepts in NLU.

NLU had its roots in the 1960s primarily utilized by the ELIZA algorithm. If you don't know about ELIZA see this account of "her" develpment and conversational output. Subseqent decades have witnessed ups and downs in the field but climaxing in 2011 with IBM Watson winning Jeopardy; a state of the art NLU question-answering system which was drawing on vast amounts of data and doing all sorts of clever things like parsing and distributional analysis of data to understand the question in a human way.

The computing power necessary to pull something like that off decreased with time, ending up from needing a supercomputer to that of a mediocre laptop, something that opened the way to the huge advancements that are happening right now;see GPT-3.

Dr. Potts explains that when you do a search into Google, various parts come into play

When you do a search into Google, you're not just finding the most relevant documents, but rather the most relevant documents as interpreted with your query in the context of things you search for where they take your query and do their best to understand what the intent behind the query is
and parse it and figure out whether it's a standard search,
or a request for directions, or a request to send a message, and so forth and so on.

The magic words here are context and intention. The computer should understand both of them in order to return an acceptable result. And this is what you will learn in taking CS224u.

As such, the complete syllabus:

  • Introduction and Welcome
  • Course Overview
  • Homework : Word Relatedness
  • High-level Goals & Guiding Hypotheses
  • Matrix Designs
  • Vector Comparison
  • Basic Reweighting
  • Dimensionality Reduction
  • Retrofitting
  • Static Representations
  • Homework : Sentiment Analysis
  • Sentiment Analysis
  • General Practical Tips
  • Stanford Sentiment Treebank
  • DynaSent
  • Hyperparameter Search
  • Feature Representation
  • RNN Classifiers
  • Contextual Representation Models
  • Transformers
  • BERT
  • RoBERTa
  • ELECTRA
  • Practical Fine-tuning
  • Homework : Colors
  • Grounded Language Understanding
  • Speakers
  • Listeners
  • Varieties of contextual grounding
  • The Rational Speech Acts Model
  • Neural RSA
  • Natural Language Inference
  • SNLI, MultiNLI, and Adversarial NLI
  • Adversarial Testing
  • Modeling Strategies
  • Attention
  • NLU and Information Retrieval
  • Classical IR
  • Neural IR, part
  • Neural IR, part
  • Neural IR, part
  • Relation Extraction
  • Data Resources
  • Problem Formulation
  • Evaluation
  • Simple Baselines
  • Directions to Explore
  • Overview of Analysis Methods in NLP
  • Adversarial Testing
  • Adversarial Training (and Testing)
    Probing
  • Feature Attribution
  • Overview of Methods and Metrics
  • Classifier Metrics
  • Natural Language Generation Metrics
  • Data Organization
  • Model Evaluation
  • Presenting Your Work: Final Papers
  • Writing NLP papers
  • NLP Conference Submissions
  • Giving Talks
  • Conclusion

Overall, it's a very intriguing course that begins with the history of NLU, showcases the crossover with NLP, teaches its practical applications in like business intelligence, speech recognition or text summarization from which both business and consumers benefit, as well as a looks at the advancements that the future will hold.

The full course material including notes can be found on the official course's website, while the recorded lectures can also be found as a complete youtube playlist.

 

More Information

CS224U: Natural Language Understanding

YouTube playlist

Related Articles

Take Stanford's Natural Language Processing with Deep Learning For Free

 

To be informed about new articles on I Programmer, sign up for our weekly newsletter, subscribe to the RSS feed and follow us on Twitter, Facebook or Linkedin.

Banner


GitHub Launches Enterprise Data Residency
30/09/2024

GitHub has announced an option offering tighter control over where data is stored to meet regional requirements. The GitHub Enterprise Cloud data residency feature will launch on October 29 for the Eu [ ... ]



TypeScript Improves Never-Initialized Variables Checks
21/10/2024

Microsoft has announced TypeScript 5.7 in beta, with improvements including stronger checks for variables that have never been initialized before use, and path rewriting for relative paths.


More News

espbook

 

Comments




or email your comment to: comments@i-programmer.info