Natural Language Processing with Deep Learning

Stanford Course , Prof. Chris Manning

413 students enrolled

Overview

Natural language processing (NLP) deals with the key artificial intelligence technology of understanding complex human language communication. This lecture series provides a thorough introduction to the cutting-edge research in deep learning applied to NLP, an approach that has recently obtained very high performance across many different NLP tasks including question answering and machine translation.

Lecture 1: Lecture 1 | Natural Language Processing with Deep Learning

Up Next
You can skip ad in
SKIP AD >
Advertisement
      • 2x
      • 1.5x
      • 1x
      • 0.5x
      • 0.25x
        EMBED LINK
        COPY
        DIRECT LINK
        PRIVATE CONTENT
        OK
        Enter password to view
        Please enter valid password!
        0:00
        0 (0 Ratings)

        Lecture Details

        Lecture 1 introduces the concept of Natural Language Processing (NLP) and the problems NLP faces today. The concept of representing words as numeric vectors is then introduced, and popular approaches to designing word vectors are discussed. Key phrases: Natural Language Processing. Word Vectors. Singular Value Decomposition. Skip-gram. Continuous Bag of Words (CBOW). Negative Sampling. Hierarchical Softmax. Word2Vec. ------------------------------------------------------------------------------- Natural Language Processing with Deep Learning Instructors: - Chris Manning - Richard Socher Natural language processing (NLP) deals with the key artificial intelligence technology of understanding complex human language communication. This lecture series provides a thorough introduction to the cutting-edge research in deep learning applied to NLP, an approach that has recently obtained very high performance across many different NLP tasks including question answering and machine translation. It emphasizes how to implement, train, debug, visualize, and design neural network models, covering the main technologies of word vectors, feed-forward models, recurrent neural networks, recursive neural networks, convolutional neural networks, and recent models involving a memory component. For additional learning opportunities please visit: http://stanfordonline.stanford.edu/

        LECTURES



        Review


        0

        0 Rates
        1
        0%
        0
        2
        0%
        0
        3
        0%
        0
        4
        0%
        0
        5
        0%
        0

        Comments Added Successfully!
        Please Enter Comments
        Please Enter CAPTCHA
        Invalid CAPTCHA
        Please Login and Submit Your Comment