80
Multimodality Learning from Text, Speech, and Vision CMU 11-4/611 Natural Language Processing Nov 19, 2020 Shruti Palaskar

Learning from Text, Speech, and Vision Multimodalitydemo.clab.cs.cmu.edu/NLP/F20/files/slides/Multimodality.pdf · 2020. 11. 19. · Learning from Text, Speech, and Vision CMU 11-4/611

  • Upload
    others

  • View
    3

  • Download
    0

Embed Size (px)

Citation preview

  • MultimodalityLearning from Text, Speech, and Vision

    CMU 11-4/611 Natural Language Processing

    Nov 19, 2020 Shruti Palaskar

  • OutlineI. What is multimodality?

    II. Types of modalities

    III. Commonly used Models

    IV. Multimodal Fusion and Representation Learning

    V. Multimodal Tasks: Use Cases

    2

  • I. What is Multimodality?

    3

  • Human Interaction is Inherently Multimodal

    4

  • How We Perceive

    5

  • How We Perceive

    6

  • The Dream: Sci-Fi Movies

    7

    JARVIS The Matrix

  • Reality?

    8

  • Give a caption.

    9

  • Give a caption.

    10

    Human: A Small Dogs Ears Stick Up As It Runs In The Grass.

    Model: A Black And White Dog Is Running On Grass With A Frisbee In Its Mouth

  • Single sentence image description -> Captioning

    11

  • Give a caption.

    12

  • Give a caption.

    13

    Human: A Young Girl In A White Dress Standing In Front Of A Fence And Fountain.

    Model: Two Men Are Standing In Front Of A Fountain

  • Reality?

    14

  • Watch the video and answer questions.

    15

    Q. is there only one person ?

    Q. does she walk in with a towel around her neck ?

    Q. does she interact with the dog ?

    Q. does she drop the towel on the floor ?

    QUESTIONS

    https://docs.google.com/file/d/1oKOK0IQLmD8z1G9LP8AEYhSBQhkWIusk/preview

  • Watch the video and answer questions.

    16

    Q. is there only one person ?A. there is only one person and a dog .

    Q. does she walk in with a towel around her neck ?A. she walks in from outside with the towel around her

    neck .

    Q. does she interact with the dog ?A. she does not interact with the dog

    Q. does she drop the towel on the floor ?A. she dropped the towel on the floor at the end of the

    video .

    QUESTIONS

    https://docs.google.com/file/d/1oKOK0IQLmD8z1G9LP8AEYhSBQhkWIusk/preview

  • Simple questions, simple answers -> Video Question Answering

    17

  • Reality? Baby Steps. Still a long way to go.

    18

  • ...ChallengesCommon challenges based on the tasks we just saw

    - Training Dataset bias- Very complicated tasks- Lack of commonsense reasoning within models- No world knowledge available like humans do

    - Physics, Nature, Memory, Experience

    How do we teach machines to perceive?

    19

  • OutlineI. What is multimodality?

    II. Types of modalities

    III. Commonly used Models

    IV. Multimodal Fusion and Representation Learning

    V. Multimodal Tasks: Use Cases

    20

  • II. Types of modalities

    21

  • Types of Modalities

    22

    IMAGE/VIDEO

    SPEECH/AUDIO

    TEXT

    EMOTION/AFFECT/SENTIMENT

  • Example Dataset: ImageNet

    ● Object Recognition● Image Tagging/Categorization● ~14M images● Knowledge Ontology● Hierarchical Tags

    ○ Mammal -> Placental -> Carnivore -> Canine -> Dog -> Working Dog -> Husky

    Deng et al. 2009 23

  • Example Dataset: How2 Dataset

    ● Speech● Video● English Transcript

    ● Portuguese Transcript● Summary

    Sanabria et al. 2018 24

  • Example Dataset: Open Pose● Action Recognition● Pose Estimation● Human Dynamic● Body Dynamics

    Wei et al. 2016 25

  • III. Commonly Used Models

    26

  • Multilayer Perceptrons

    27

    Single Perceptron

  • Multilayer Perceptrons

    28Single Perceptron

  • Multilayer Perceptrons: Uses in Multimedia

    29

  • Multilayer Perceptrons: Limitations

    30

    Limitation #1

    Very large amount of input data samples (xi), which requires a gigantic amount of model parameters.

  • Convolutional Neural Networks (CNNs)

    31

    Translation invariance: we can use same parameters to capture a specific “feature” in any area of the image. We can use different sets of parameters to capture different features.

    These operations are equivalent to perform convolutions with different filters.

  • Convolutional Neural Networks (CNNs)

    32LeCun et al. 1998

  • Convolutional Neural Networks (CNNs) for Image Encoding

    33Krizhevsky et al. 2012

  • Multilayer Perceptrons: Limitations

    34

    Limitation #1

    Very large amount of input data samples (xi), which requires a gigantic amount of model parameters.

    Limitation #2

    Does not naturally handle input data of variable dimension

    (eg. audio/video/word sequences)

  • Recurrent Neural Networks

    35

    Build specific connections capturing the temporal evolution

    → Shared weights in time

  • Recurrent Neural Networks

    36

  • Recurrent Neural Networks for Video Encoding

    37

    Combination is commonly implemented as a small NN on top of a pooling operation (e.g. max, sum, average).

    Recurrent Neural Networks are well suited for processing sequences.

    Donahue et al. 2015

  • Attention Mechanism

    38Bahdanau et al. 2014, Luong et al. 2015

  • Sequence-to-Sequence Models with Attention

    Image from Luong et al. 2015 39

  • Transformers

    40Vaswani et al. 2017

  • IV. Multimodal Fusion & Representation Learning

    41

  • Slide courtesy: LP Morency

    Fusion: Model Agnostic

    42

  • Fusion: Model Based

    43Slide courtesy: LP Morency

  • Representation Learning: Encoder-Decoder

    44

  • Representation Learning

    45

    Word2Vec

    Mikolov et al. 2013

  • Representation Learning: RNNs

    46Cho et al. 2014

  • Representation Learning: Self-Supervised

    47

  • Representation Learning: Transfer Learning

    48

  • Representation Learning: Joint Learning (Contrastive)

    49

  • Representation Learning: Joint Learning (Similarity)

    50

  • Large-scale Pre-training -- Vision+Language BERT

    51Zhou et al. 2019

  • V. Common Tasks, Use Cases

    52

  • V. Common Tasks1. Vision and Language2. Speech, Vision and Language3. Multimedia4. Emotion and Affect

    53

    ● Image/Video Captioning● Visual Question Answering● Visual Dialog ● Video Summarization● Lip Reading● Audio Visual Speech Recognition● Visual Speech Synthesis● …

  • 1. Vision and Language Common Tasks

    54

  • Image Captioning

    55Vinyals et al. 2015

  • Image Captioning

    56

    Karpathy et al. 2015Slide by Marc Bolaños

  • Image Captioning: Show, Attend and Tell

    57Xu et al. 2015

  • Image Captioning and Detection

    58Johnson et al. 2016

  • Video Captioning

    59Donahue et al. 2015

  • Video Captioning

    60

    Pan et al. 2016Slides by Marc Bolaños

  • Visual Question Answering

    61

  • Visual Question Answering

    62

  • Visual Question Answering

    63

  • Visual Question Answering

    64

  • Visual Question Answering

    65

  • Video Summarization

    on behalf of expert village my name is lizbeth muller and today we are going to show you how to make spanish omelet . i 'm going to dice a little bit of peppers here . i 'm not going to use a lot , i 'm going to use very very little . a little bit more then this maybe . you can use red peppers if you like to get a little bit color in your omelet . some people do and some people do n't . but i find that some of the people that are mexicans who are friends of mine that have a mexican she like to put red peppers and green peppers and yellow peppers in hers and with a lot of onions . that is the way they make there spanish omelets that is what she says . i loved it , it actually tasted really good . you are going to take the onion also and dice it really small . you do n't want big chunks of onion in there cause it is just pops out of the omelet . so we are going to dice the up also very very small . so we have small pieces of onions and peppers ready to go .

    how to cut peppers to make a spanish omelette ; get expert tips and advice on making cuban breakfast recipes in this free cooking video .

    Transcript (290 words on avg)

    “Teaser” (33 words on avg)

    ~1.5 minutes of audio and video

    http://www.youtube.com/watch?v=Fz-N1S0swh8

  • Video Summarization: Hierarchical Model

    Palaskar et al. 2019

  • Action Recognition

    68

  • 2. Speech, Vision and Language Common Tasks

    69

  • Audio Visual Speech Recognition: Lip Reading

    70

    Assael et al. 2016

  • Lip Reading: Watch, Listen, Attend and Spell

    71

    Chung et al. 2017

  • 3. Multimedia Common Tasks

    72

  • Multimedia Retrieval

    73

  • Multimedia Retrieval

    74

  • Multimedia Retrieval: Shared Multimodal Representation

    75

  • Multimedia Retrieval

    76

  • 4. Emotion and Affect

    77

  • Affect Recognition: Emotion, Sentiment, Persuasion, Personality

    78

  • OutlineI. What is multimodality?

    II. Types of modalities

    III. Commonly used Models

    IV. Multimodal Fusion and Representation Learning

    V. Multimodal Tasks: Use Cases

    79

  • Takeaways● Lots of multimodal data generated everyday● Need automatic ways to understand it

    ○ Privacy○ Security○ Regulation○ Storage

    ● Different models used for different downstream tasks○ Highly open-ended research!

    ● Try it out for fun on Kaggle!

    Thank [email protected] 80