Multi-Class Sentiment Classification for Short Text Sequences
|
|
- Arline Quinn
- 5 years ago
- Views:
Transcription
1 Multi-Class Sentiment Classification for Short Text Sequences TIMOTHY LIU KAIHUI SINGAPORE UNIVERSITY OF TECHNOLOGY AND DESIGN
2 What a selfless and courageous hero... Willing to give his life for a total stranger... Deepest sympathies and sincere condolences for his family and all in France NEUTRAL HAPPY SAD ANGRY HATE Sack the whole f***ing cabinet NEUTRAL HAPPY SAD ANGRY HATE
3 Motivation As part of a greater research project to tackle Fake News Team of students in SUTD
4 Motivation Evaluate people s reaction to online content Record it as a feature" Use this feature (among others) to flag suspicious content
5
6 Clickbait heading? Quality of text? Known hoax? Audience Reaction
7 Challenges Appropriate Data short mid length, highly informal sentences Expressing wide range of emotions IMDB Movie Reviews, Amazon product reviews Tweets? Label granularity (!!) Positive/Negative is not enough to gauge reaction Hard to find such labelled data
8 Dataset 40,000 tweets labelled with 14 emotion classes By CrowdFlower, hosted on data.world
9 Dataset
10 Dataset
11 Data Engineering Merging of classes Match previous papers (Bouazizi and Ohtsuki, 2017) 5 classes: Neutral, Happy, Sad, Angry, Hate Acquire additional data Crawl Twitter for tweets with tagged emotions i.e. #happy #angry etc. Not perfect :( After some rejection/cleaning: 47,288 tweets
12 Data Engineering
13 Data Engineering Not OK :(
14 Applying Deep Learning Artificial Neural Networks
15 Model Design Extract as many features/representations as possible Lack of data (in length of text and number of examples per class) Be as generalisable as possible Work on Singlish and different type of source medium Target application is comments on local news websites and blogs As opposed to tweets in training data
16 Model Design INPUT SEQUENCE WORD EMBEDDINGS Vanilla CNN 1D CONVOLUTION POOLING LAYER DENSE LAYER Keras name for fully-connected layer SOFTMAX
17 Model Design INPUT SEQUENCE WORD EMBEDDINGS Vanilla CNN 1D CONVOLUTION POOLING LAYER DENSE LAYER High dimensional vector space that captures relationships between terms SOFTMAX
18 CNN applied to words Extract features using 1D convolution
19 CNN applied to words Extract features using 1D convolution I salute you for the bravery and sacrifice! A true hero indeed. [ 1, 6097, 5, 12, 3, 6, 4841, 4, 515, 1476, 1238] salutation feature
20 Model Design INPUT SEQUENCE WORD EMBEDDINGS 1D CONVOLUTION POOLING LAYER DENSE LAYER SOFTMAX
21 Model Design Yoon Kim, 2014 INPUT SEQUENCE INPUT SEQUENCE WORD EMBEDDINGS WORD EMBEDDINGS WORD EMBEDDINGS 1D CONVOLUTION 1D CONV 1D CONV 1D CONV POOLING LAYER POOLING LAYER DENSE LAYER DENSE LAYER SOFTMAX SOFTMAX
22 Model Design Yoon Kim, 2014 INPUT SEQUENCE INPUT SEQUENCE WORD EMBEDDINGS WORD EMBEDDINGS WORD EMBEDDINGS 1D CONVOLUTION 1D CONV 1D CONV 1D CONV POOLING LAYER POOLING LAYER DENSE LAYER DENSE LAYER SOFTMAX SOFTMAX
23 Model Design Yoon Kim, 2014 INPUT SEQUENCE INPUT SEQUENCE WORD EMBEDDINGS WORD EMBEDDINGS WORD EMBEDDINGS 1D CONVOLUTION 1D CONV 1D CONV 1D CONV POOLING LAYER POOLING LAYER DENSE LAYER DENSE LAYER SOFTMAX SOFTMAX
24 CNN applied to words Extract features using 1D convolution I salute you for the bravery and sacrifice! A true hero indeed. [ 1, 6097, 5, 12, 3, 6, 4841, 4, 515, 1476, 1238] salutation Change Window/Filter Size
25 CNN applied to words Extract features using 1D convolution I salute you for the bravery and sacrifice! A true hero indeed. [ 1, 6097, 5, 12, 3, 6, 4841, 4, 515, 1476, 1238] salutation Change Window/Filter Size
26 CNN applied to words Extract features using 1D convolution I salute you for the bravery and sacrifice! A true hero indeed. [ 1, 6097, 5, 12, 3, 6, 4841, 4, 515, 1476, 1238] salutation Change Window/Filter Size
27 Downside of CNN Sequence matters Context is important CNN
28 LSTM Type of recurrent neural network (RNN) To learn sequences via series of hidden states Memory cells can keep information intact unless inputs makes them forget/overwrite Cell can decide to output this information or just store it (Zhou & Sun et al., 2015)
29 LSTM
30 LSTM [ 1, 6097, 5, 12, 3, 6, 4841, 4, 515, 1476, 1238]
31 LSTM [ 1, 6097, 5, 12, 3, 6, 4841, 4, 515, 1476, 1238]
32 LSTM [ 1, 6097, 5, 12, 3, 6, 4841, 4, 515, 1476, 1238]
33 LSTM [ 1, 6097, 5, 12, 3, 6, 4841, 4, 515, 1476, 1238]
34 LSTM [ 1, 6097, 5, 12, 3, 6, 4841, 4, 515, 1476, 1238]
35 LSTM [ 1, 6097, 5, 12, 3, 6, 4841, 4, 515, 1476, 1238]
36 LSTM [ 1, 6097, 5, 12, 3, 6, 4841, 4, 515, 1476, 1238]
37 LSTM [ 1, 6097, 5, 12, 3, 6, 4841, 4, 515, 1476, 1238]
38 LSTM [ 1, 6097, 5, 12, 3, 6, 4841, 4, 515, 1476, 1238]
39 LSTM [ 1, 6097, 5, 12, 3, 6, 4841, 4, 515, 1476, 1238]
40 LSTM [ 1, 6097, 5, 12, 3, 6, 4841, 4, 515, 1476, 1238]
41 LSTM [ 1, 6097, 5, 12, 3, 6, 4841, 4, 515, 1476, 1238] output
42 Model Design Why should we combine CNN and LSTM? To learn sequences of features and features of sequences. (Zhou & Sun et al., 2015)
43 Model Design Why should we combine CNN and LSTM? To learn sequences of features and features of sequences. Linear combination : overall validation accuracy < 60% How can we maximise feature extraction and improve classification accuracy to > 60%? Hyper-parameter tuning not enough tried
44 BalanceNet Make use of an inter-connected architecture of opposite architectures It does look like a balance
45
46 BalanceNet Make use of an inter-connected architecture of opposite architectures Captures all previous ideas on how to design the model Create multiple channels through the model to allow the model to select which channels work better for certain classes
47 BalanceNet INPUT SEQUENCE WORD EMBEDDINGS WORD EMBEDDINGS LSTM 1D CONV 1D CONV 1D CONV 1D CONV 1D CONV 1D CONV LSTM
48 BalanceNet INPUT SEQUENCE WORD EMBEDDINGS WORD EMBEDDINGS LSTM 1D CONV 1D CONV 1D CONV 1D CONV 1D CONV 1D CONV LSTM
49 BalanceNet INPUT SEQUENCE WORD EMBEDDINGS WORD EMBEDDINGS LSTM 1D CONV 1D CONV 1D CONV 1D CONV 1D CONV 1D CONV LSTM
50 BalanceNet INPUT SEQUENCE WORD EMBEDDINGS WORD EMBEDDINGS LSTM 1D CONV 1D CONV 1D CONV 1D CONV 1D CONV 1D CONV LSTM
51 BalanceNet INPUT SEQUENCE WORD EMBEDDINGS WORD EMBEDDINGS LSTM 1D CONV 1D CONV 1D CONV 1D CONV 1D CONV 1D CONV LSTM
52 BalanceNet INPUT SEQUENCE Channel WORD EMBEDDINGS WORD EMBEDDINGS LSTM 1D CONV 1D CONV 1D CONV 1D CONV 1D CONV 1D CONV LSTM Produce one of many outputs Model can learn which outputs represent which classes
53 BalanceNet MAX POOLING DROPOUT DENSE LAYER SOFTMAX
54 Results
55 Results
56 Results (happy) Haha look at all u dumbos cant read premium articles, im gonna becum premium nao to read their quality unbiased journalism Prediction: happy Wah gd deal. No much peepor there almost like whole area to urself. And if prepper still got boat for escape. Hope i win 2mr toto, HUAT ARH!!!!! Prediction: happy I salute you for the bravery and sacrifice! A true hero indeed. Prediction: happy
57 Results (sad) Truly I said, we are so greatly divided in this world. Indeed money can do wonder, no one can deny it. Don't remind me, I know I'm still the loser. Prediction: sad what a nuisance fk. a proper clean and flat footpath,,now obstructed by sharedbikes..! which idiotic MP allowed this to happen? Prediction: sad Billions of money spent end up become the words 'unable to find out what's wrong with the system' Prediction: sad
58 Results (hate) Thought he sold his kidney to buy it; Instead, he bought a kidney then bought the car Filthy rich This is why we need communism Prediction: hate Sack the whole fucking cabinet Prediction: hate For fuck s sake - news, stop calling it a machine. It was a guy in a fucking box. Like that then glory hole macam innovative blowjob machine. Prediction: hate
59 Results (neutral) ST really need to wake up. Prediction: neutral Dun need give birth?? Come in PRC ok Liao Prediction: neutral
60 Results (bonus?) Somebody needs to water Tharman's head, hair needs to be grown there Prediction: sad
61 What went wrong Distribution of training set Lower occurrences of angry tweets in training set results in less frequent/accurate predictions hate is more distinct so is easier to pick out
62 What went wrong What s up with the neutral class? Not easy to set boundaries for emotion-less text Hence the model picks up latent emotion Preprocessing could be better Mistake: did not process the same way as Stanford (for their pretrained GloVe vectors, which I used)
63 Future Plans Develop binary classifiers for emotion categories instead Develop sarcasm detector
64 Questions? Thank you for listening!!
65 Breakdown What does the data look like as it passes through the layers?
66 Breakdown INPUT SEQUENCE WORD EMBEDDINGS WORD EMBEDDINGS LSTM 1D CONV 1D CONV 1D CONV 1D CONV 1D CONV 1D CONV LSTM
67 Breakdown ===== Layer name: input_1 ===== Tensor: [[ ]] With shape: (1, 30) =====
68 Breakdown INPUT SEQUENCE WORD EMBEDDINGS WORD EMBEDDINGS LSTM 1D CONV 1D CONV 1D CONV 1D CONV 1D CONV 1D CONV LSTM
69 Breakdown ===== Layer name: input_1 ===== Tensor: [[ ]] With shape: (1, 30) =====
70 Breakdown ===== Layer name: input_1 ===== Tensor: [[ d ]] With shape: (1, 30) =====
71 Breakdown ===== Layer name: embedding_1 ===== Tensor: [[[ ] [ ] [ ]... [ ] [ ] [ ]]] With shape: (1, 30, 200) =====
72 Breakdown INPUT SEQUENCE WORD EMBEDDINGS WORD EMBEDDINGS LSTM 1D CONV 1D CONV 1D CONV 1D CONV 1D CONV 1D CONV LSTM
73 Breakdown ===== Layer name: bidirectional_1 ===== Tensor: [[[ e e e e e e e e e e e e-06] [ e e e e e e e e e e e e-06] [ e e e e e e e e e e e e+00] [ e e e e e e e e e e e e+00]]] With shape: (1, 30, 12) =====
74 Breakdown ===== Layer name: bidirectional_1 ===== Tensor: [[[ e e e e e e e e e e e e-06] [ e e e e e e e e e e e e-06] [ e e e e e e e e e e e e+00] [ e e e e e e e e e e e e+00]]] With shape: (1, 30, 12) =====
75 Breakdown INPUT SEQUENCE WORD EMBEDDINGS WORD EMBEDDINGS LSTM 1D CONV 1D CONV 1D CONV 1D CONV 1D CONV 1D CONV LSTM
76 Breakdown ===== Layer name: conv1d_4 ===== Tensor: [[[ ] [ ] [ ]... [ ] [ ] [ ]]] With shape: (1, 55, 24) =====
77 Breakdown INPUT SEQUENCE WORD EMBEDDINGS WORD EMBEDDINGS LSTM 1D CONV 1D CONV 1D CONV 1D CONV 1D CONV 1D CONV LSTM
78 Breakdown ===== Layer name: conv1d_6 ===== Tensor: [[[ ]... [ ] [ ] [ ] [ ]]] With shape: (1, 27, 12) =====
79 Breakdown ===== Layer name: conv1d_6 ===== Tensor: [[[ ]... [ ] [ ] [ ] [ ]]] With shape: (1, 27, 12) =====
80 Breakdown INPUT SEQUENCE WORD EMBEDDINGS WORD EMBEDDINGS LSTM 1D CONV 1D CONV 1D CONV 1D CONV 1D CONV 1D CONV LSTM
81 Breakdown ===== Layer name: bidirectional_3 ===== Tensor: [[[ e e e e e e-01] [ e e e e e e-01] [ e e e e e e-01]... [ e e e e e e-03] [ e e e e e e-05] [ e e e e e e-04]]] With shape: (1, 168, 24) =====
82 Breakdown INPUT SEQUENCE WORD EMBEDDINGS WORD EMBEDDINGS LSTM 1D CONV 1D CONV 1D CONV 1D CONV 1D CONV 1D CONV LSTM
83 Breakdown ===== Layer name: concatenate_4 ===== Tensor: [[[ e e e e e e+00] [ e e e e e e+00] [ e e e e e e+00]... [ e e e e e e-03] [ e e e e e e-05] [ e e e e e e-04]]] With shape: (1, 449, 24) =====
84 Breakdown MAX POOLING DROPOUT DENSE LAYER SOFTMAX
85 Breakdown ===== Layer name: max_pooling1d_1 ===== Tensor: [[[ ] [ ] [ ]... [ ] [ ] [ ]]] With shape: (1, 112, 24) =====
86 Breakdown ===== Layer name: dropout_12 ===== Tensor: [[[ ] [ ] [ ]... [ ] [ ] [ ]]] With shape: (1, 112, 24) =====
87 Breakdown ===== Layer name: dropout_12 ===== Tensor: [[[ ] [ ] [ ]... [ ] [ ] [ ]]] With shape: (1, 112, 24) =====
88 Breakdown ===== Layer name: dense_1 ===== Tensor: [[ ]] With shape: (1, 26) ===== ===== Layer name: dense_2 ===== Tensor: [[ ]] With shape: (1, 5) =====
89 Breakdown ===== Layer name: dense_1 ===== Tensor: [[ ]] With shape: (1, 26) ===== ===== Layer name: dense_2 ===== Tensor: [[ ]] With shape: (1, 5) =====
Deep Learning for Natural Language Processing. Sidharth Mudgal April 4, 2017
Deep Learning for Natural Language Processing Sidharth Mudgal April 4, 2017 Table of contents 1. Intro 2. Word Vectors 3. Word2Vec 4. Char Level Word Embeddings 5. Application: Entity Matching 6. Conclusion
More informationApplied Natural Language Processing
Applied Natural Language Processing Info 256 Lecture 20: Sequence labeling (April 9, 2019) David Bamman, UC Berkeley POS tagging NNP Labeling the tag that s correct for the context. IN JJ FW SYM IN JJ
More informationCSC321 Lecture 16: ResNets and Attention
CSC321 Lecture 16: ResNets and Attention Roger Grosse Roger Grosse CSC321 Lecture 16: ResNets and Attention 1 / 24 Overview Two topics for today: Topic 1: Deep Residual Networks (ResNets) This is the state-of-the
More information(
Class 15 - Long Short-Term Memory (LSTM) Study materials http://colah.github.io/posts/2015-08-understanding-lstms/ (http://colah.github.io/posts/2015-08-understanding-lstms/) http://karpathy.github.io/2015/05/21/rnn-effectiveness/
More informationSequence Modeling with Neural Networks
Sequence Modeling with Neural Networks Harini Suresh y 0 y 1 y 2 s 0 s 1 s 2... x 0 x 1 x 2 hat is a sequence? This morning I took the dog for a walk. sentence medical signals speech waveform Successes
More informationDeep Learning for NLP
Deep Learning for NLP Instructor: Wei Xu Ohio State University CSE 5525 Many slides from Greg Durrett Outline Motivation for neural networks Feedforward neural networks Applying feedforward neural networks
More informationArtificial Neural Networks D B M G. Data Base and Data Mining Group of Politecnico di Torino. Elena Baralis. Politecnico di Torino
Artificial Neural Networks Data Base and Data Mining Group of Politecnico di Torino Elena Baralis Politecnico di Torino Artificial Neural Networks Inspired to the structure of the human brain Neurons as
More informationNeural Architectures for Image, Language, and Speech Processing
Neural Architectures for Image, Language, and Speech Processing Karl Stratos June 26, 2018 1 / 31 Overview Feedforward Networks Need for Specialized Architectures Convolutional Neural Networks (CNNs) Recurrent
More informationNeural Networks 2. 2 Receptive fields and dealing with image inputs
CS 446 Machine Learning Fall 2016 Oct 04, 2016 Neural Networks 2 Professor: Dan Roth Scribe: C. Cheng, C. Cervantes Overview Convolutional Neural Networks Recurrent Neural Networks 1 Introduction There
More informationNatural Language Processing
Natural Language Processing Info 59/259 Lecture 4: Text classification 3 (Sept 5, 207) David Bamman, UC Berkeley . https://www.forbes.com/sites/kevinmurnane/206/04/0/what-is-deep-learning-and-how-is-it-useful
More informationCS 229 Project Final Report: Reinforcement Learning for Neural Network Architecture Category : Theory & Reinforcement Learning
CS 229 Project Final Report: Reinforcement Learning for Neural Network Architecture Category : Theory & Reinforcement Learning Lei Lei Ruoxuan Xiong December 16, 2017 1 Introduction Deep Neural Network
More informationPulses Characterization from Raw Data for CDMS
Pulses Characterization from Raw Data for CDMS Physical Sciences Francesco Insulla (06210287), Chris Stanford (05884854) December 14, 2018 Abstract In this project we seek to take a time series of current
More informationWaveNet: A Generative Model for Raw Audio
WaveNet: A Generative Model for Raw Audio Ido Guy & Daniel Brodeski Deep Learning Seminar 2017 TAU Outline Introduction WaveNet Experiments Introduction WaveNet is a deep generative model of raw audio
More informationNEURAL LANGUAGE MODELS
COMP90042 LECTURE 14 NEURAL LANGUAGE MODELS LANGUAGE MODELS Assign a probability to a sequence of words Framed as sliding a window over the sentence, predicting each word from finite context to left E.g.,
More informationMultimodal context analysis and prediction
Multimodal context analysis and prediction Valeria Tomaselli (valeria.tomaselli@st.com) Sebastiano Battiato Giovanni Maria Farinella Tiziana Rotondo (PhD student) Outline 2 Context analysis vs prediction
More informationMachine Learning for Large-Scale Data Analysis and Decision Making A. Neural Networks Week #6
Machine Learning for Large-Scale Data Analysis and Decision Making 80-629-17A Neural Networks Week #6 Today Neural Networks A. Modeling B. Fitting C. Deep neural networks Today s material is (adapted)
More informationMachine Learning: Chenhao Tan University of Colorado Boulder LECTURE 6
Machine Learning: Chenhao Tan University of Colorado Boulder LECTURE 6 Slides adapted from Jordan Boyd-Graber, Chris Ketelsen Machine Learning: Chenhao Tan Boulder 1 of 39 HW1 turned in HW2 released Office
More informationDeep Sequence Models. Context Representation, Regularization, and Application to Language. Adji Bousso Dieng
Deep Sequence Models Context Representation, Regularization, and Application to Language Adji Bousso Dieng All Data Are Born Sequential Time underlies many interesting human behaviors. Elman, 1990. Why
More informationAffect recognition from facial movements and body gestures by hierarchical deep spatio-temporal features and fusion strategy
Accepted Manuscript Affect recognition from facial movements and body gestures by hierarchical deep spatio-temporal features and fusion strategy Bo Sun, Siming Cao, Jun He, Lejun Yu PII: S0893-6080(17)30284-8
More informationRandom Coattention Forest for Question Answering
Random Coattention Forest for Question Answering Jheng-Hao Chen Stanford University jhenghao@stanford.edu Ting-Po Lee Stanford University tingpo@stanford.edu Yi-Chun Chen Stanford University yichunc@stanford.edu
More informationCaesar s Taxi Prediction Services
1 Caesar s Taxi Prediction Services Predicting NYC Taxi Fares, Trip Distance, and Activity Paul Jolly, Boxiao Pan, Varun Nambiar Abstract In this paper, we propose three models each predicting either taxi
More informationGenerating Sequences with Recurrent Neural Networks
Generating Sequences with Recurrent Neural Networks Alex Graves University of Toronto & Google DeepMind Presented by Zhe Gan, Duke University May 15, 2015 1 / 23 Outline Deep recurrent neural network based
More informationLecture 5 Neural models for NLP
CS546: Machine Learning in NLP (Spring 2018) http://courses.engr.illinois.edu/cs546/ Lecture 5 Neural models for NLP Julia Hockenmaier juliahmr@illinois.edu 3324 Siebel Center Office hours: Tue/Thu 2pm-3pm
More informationAdaptive Multi-Compositionality for Recursive Neural Models with Applications to Sentiment Analysis. July 31, 2014
Adaptive Multi-Compositionality for Recursive Neural Models with Applications to Sentiment Analysis July 31, 2014 Semantic Composition Principle of Compositionality The meaning of a complex expression
More informationOnline Videos FERPA. Sign waiver or sit on the sides or in the back. Off camera question time before and after lecture. Questions?
Online Videos FERPA Sign waiver or sit on the sides or in the back Off camera question time before and after lecture Questions? Lecture 1, Slide 1 CS224d Deep NLP Lecture 4: Word Window Classification
More informationLecture 17: Neural Networks and Deep Learning
UVA CS 6316 / CS 4501-004 Machine Learning Fall 2016 Lecture 17: Neural Networks and Deep Learning Jack Lanchantin Dr. Yanjun Qi 1 Neurons 1-Layer Neural Network Multi-layer Neural Network Loss Functions
More informationIntroduction to Deep Neural Networks
Introduction to Deep Neural Networks Presenter: Chunyuan Li Pattern Classification and Recognition (ECE 681.01) Duke University April, 2016 Outline 1 Background and Preliminaries Why DNNs? Model: Logistic
More informationDeep Learning: Pre- Requisites. Understanding gradient descent, autodiff, and softmax
Deep Learning: Pre- Requisites Understanding gradient descent, autodiff, and softmax Gradient Descent autodiff Gradient descent requires knowledge of, well, the gradient from your cost function (MSE) Mathematically
More informationConditional Language modeling with attention
Conditional Language modeling with attention 2017.08.25 Oxford Deep NLP 조수현 Review Conditional language model: assign probabilities to sequence of words given some conditioning context x What is the probability
More informationTracking the World State with Recurrent Entity Networks
Tracking the World State with Recurrent Entity Networks Mikael Henaff, Jason Weston, Arthur Szlam, Antoine Bordes, Yann LeCun Task At each timestep, get information (in the form of a sentence) about the
More informationGenerative Models for Sentences
Generative Models for Sentences Amjad Almahairi PhD student August 16 th 2014 Outline 1. Motivation Language modelling Full Sentence Embeddings 2. Approach Bayesian Networks Variational Autoencoders (VAE)
More informationIntroduction to Machine Learning Midterm Exam
10-701 Introduction to Machine Learning Midterm Exam Instructors: Eric Xing, Ziv Bar-Joseph 17 November, 2015 There are 11 questions, for a total of 100 points. This exam is open book, open notes, but
More informationStructured Neural Networks (I)
Structured Neural Networks (I) CS 690N, Spring 208 Advanced Natural Language Processing http://peoplecsumassedu/~brenocon/anlp208/ Brendan O Connor College of Information and Computer Sciences University
More informationAnalysis of the Learning Process of a Recurrent Neural Network on the Last k-bit Parity Function
Analysis of the Learning Process of a Recurrent Neural Network on the Last k-bit Parity Function Austin Wang Adviser: Xiuyuan Cheng May 4, 2017 1 Abstract This study analyzes how simple recurrent neural
More informationJakub Hajic Artificial Intelligence Seminar I
Jakub Hajic Artificial Intelligence Seminar I. 11. 11. 2014 Outline Key concepts Deep Belief Networks Convolutional Neural Networks A couple of questions Convolution Perceptron Feedforward Neural Network
More informationStance classification and Diffusion Modelling
Dr. Srijith P. K. CSE, IIT Hyderabad Outline Stance Classification 1 Stance Classification 2 Stance Classification in Twitter Rumour Stance Classification Classify tweets as supporting, denying, questioning,
More informationNeural networks and optimization
Neural networks and optimization Nicolas Le Roux Criteo 18/05/15 Nicolas Le Roux (Criteo) Neural networks and optimization 18/05/15 1 / 85 1 Introduction 2 Deep networks 3 Optimization 4 Convolutional
More informationMachine Learning for Signal Processing Neural Networks Continue. Instructor: Bhiksha Raj Slides by Najim Dehak 1 Dec 2016
Machine Learning for Signal Processing Neural Networks Continue Instructor: Bhiksha Raj Slides by Najim Dehak 1 Dec 2016 1 So what are neural networks?? Voice signal N.Net Transcription Image N.Net Text
More informationTackling the Limits of Deep Learning for NLP
Tackling the Limits of Deep Learning for NLP Richard Socher Salesforce Research Caiming Xiong, Romain Paulus, Stephen Merity, James Bradbury, Victor Zhong Einstein s Deep Learning Einstein Vision Einstein
More informationNeural Networks for Machine Learning. Lecture 2a An overview of the main types of neural network architecture
Neural Networks for Machine Learning Lecture 2a An overview of the main types of neural network architecture Geoffrey Hinton with Nitish Srivastava Kevin Swersky Feed-forward neural networks These are
More informationIntroduction to Machine Learning Midterm Exam Solutions
10-701 Introduction to Machine Learning Midterm Exam Solutions Instructors: Eric Xing, Ziv Bar-Joseph 17 November, 2015 There are 11 questions, for a total of 100 points. This exam is open book, open notes,
More informationFinal Examination CS 540-2: Introduction to Artificial Intelligence
Final Examination CS 540-2: Introduction to Artificial Intelligence May 7, 2017 LAST NAME: SOLUTIONS FIRST NAME: Problem Score Max Score 1 14 2 10 3 6 4 10 5 11 6 9 7 8 9 10 8 12 12 8 Total 100 1 of 11
More informationDeep Learning (CNNs)
10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Deep Learning (CNNs) Deep Learning Readings: Murphy 28 Bishop - - HTF - - Mitchell
More informationMemory-Augmented Attention Model for Scene Text Recognition
Memory-Augmented Attention Model for Scene Text Recognition Cong Wang 1,2, Fei Yin 1,2, Cheng-Lin Liu 1,2,3 1 National Laboratory of Pattern Recognition Institute of Automation, Chinese Academy of Sciences
More informationMachine Learning for Computer Vision 8. Neural Networks and Deep Learning. Vladimir Golkov Technical University of Munich Computer Vision Group
Machine Learning for Computer Vision 8. Neural Networks and Deep Learning Vladimir Golkov Technical University of Munich Computer Vision Group INTRODUCTION Nonlinear Coordinate Transformation http://cs.stanford.edu/people/karpathy/convnetjs/
More informationwith Local Dependencies
CS11-747 Neural Networks for NLP Structured Prediction with Local Dependencies Xuezhe Ma (Max) Site https://phontron.com/class/nn4nlp2017/ An Example Structured Prediction Problem: Sequence Labeling Sequence
More informationDeep Learning For Mathematical Functions
000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050
More informationAdministration. Registration Hw3 is out. Lecture Captioning (Extra-Credit) Scribing lectures. Questions. Due on Thursday 10/6
Administration Registration Hw3 is out Due on Thursday 10/6 Questions Lecture Captioning (Extra-Credit) Look at Piazza for details Scribing lectures With pay; come talk to me/send email. 1 Projects Projects
More informationPresented By: Omer Shmueli and Sivan Niv
Deep Speaker: an End-to-End Neural Speaker Embedding System Chao Li, Xiaokong Ma, Bing Jiang, Xiangang Li, Xuewei Zhang, Xiao Liu, Ying Cao, Ajay Kannan, Zhenyao Zhu Presented By: Omer Shmueli and Sivan
More informationConvolutional Neural Network Architecture
Convolutional Neural Network Architecture Zhisheng Zhong Feburary 2nd, 2018 Zhisheng Zhong Convolutional Neural Network Architecture Feburary 2nd, 2018 1 / 55 Outline 1 Introduction of Convolution Motivation
More informationDeep Learning Sequence to Sequence models: Attention Models. 17 March 2018
Deep Learning Sequence to Sequence models: Attention Models 17 March 2018 1 Sequence-to-sequence modelling Problem: E.g. A sequence X 1 X N goes in A different sequence Y 1 Y M comes out Speech recognition:
More informationNatural Language Processing and Recurrent Neural Networks
Natural Language Processing and Recurrent Neural Networks Pranay Tarafdar October 19 th, 2018 Outline Introduction to NLP Word2vec RNN GRU LSTM Demo What is NLP? Natural Language? : Huge amount of information
More informationStephen Scott.
1 / 35 (Adapted from Vinod Variyam and Ian Goodfellow) sscott@cse.unl.edu 2 / 35 All our architectures so far work on fixed-sized inputs neural networks work on sequences of inputs E.g., text, biological
More informationDeep learning / Ian Goodfellow, Yoshua Bengio and Aaron Courville. - Cambridge, MA ; London, Spis treści
Deep learning / Ian Goodfellow, Yoshua Bengio and Aaron Courville. - Cambridge, MA ; London, 2017 Spis treści Website Acknowledgments Notation xiii xv xix 1 Introduction 1 1.1 Who Should Read This Book?
More informationFace recognition for galaxies: Artificial intelligence brings new tools to astronomy
April 23, 2018 Contact: Tim Stephens (831) 459-4352; stephens@ucsc.edu Face recognition for galaxies: Artificial intelligence brings new tools to astronomy A 'deep learning' algorithm trained on images
More informationNotes on Deep Learning for NLP
Notes on Deep Learning for NLP Antoine J.-P. Tixier Computer Science Department (DaSciM team) École Polytechnique, Palaiseau, France antoine.tixier-1@colorado.edu Last updated Thursday 30 th August, 2018
More informationNeural Networks with Applications to Vision and Language. Feedforward Networks. Marco Kuhlmann
Neural Networks with Applications to Vision and Language Feedforward Networks Marco Kuhlmann Feedforward networks Linear separability x 2 x 2 0 1 0 1 0 0 x 1 1 0 x 1 linearly separable not linearly separable
More informationWays to make neural networks generalize better
Ways to make neural networks generalize better Seminar in Deep Learning University of Tartu 04 / 10 / 2014 Pihel Saatmann Topics Overview of ways to improve generalization Limiting the size of the weights
More informationEfficientLow-rank Multimodal Fusion With Modality-specific Factors
EfficientLow-rank Multimodal Fusion With Modality-specific Factors Zhun Liu, Ying Shen, Varun Bharadwaj, Paul Pu Liang, Amir Zadeh, Louis-Philippe Morency Artificial Intelligence Multimodal Sentiment and
More informationRecurrent Neural Networks. deeplearning.ai. Why sequence models?
Recurrent Neural Networks deeplearning.ai Why sequence models? Examples of sequence data The quick brown fox jumped over the lazy dog. Speech recognition Music generation Sentiment classification There
More informationNatural Language Processing
Natural Language Processing Pushpak Bhattacharyya CSE Dept, IIT Patna and Bombay LSTM 15 jun, 2017 lgsoft:nlp:lstm:pushpak 1 Recap 15 jun, 2017 lgsoft:nlp:lstm:pushpak 2 Feedforward Network and Backpropagation
More informationDeep Learning Recurrent Networks 2/28/2018
Deep Learning Recurrent Networks /8/8 Recap: Recurrent networks can be incredibly effective Story so far Y(t+) Stock vector X(t) X(t+) X(t+) X(t+) X(t+) X(t+5) X(t+) X(t+7) Iterated structures are good
More informationArtificial Neural Networks. Introduction to Computational Neuroscience Tambet Matiisen
Artificial Neural Networks Introduction to Computational Neuroscience Tambet Matiisen 2.04.2018 Artificial neural network NB! Inspired by biology, not based on biology! Applications Automatic speech recognition
More informationNatural Language Processing with Deep Learning CS224N/Ling284
Natural Language Processing with Deep Learning CS224N/Ling284 Lecture 4: Word Window Classification and Neural Networks Richard Socher Organization Main midterm: Feb 13 Alternative midterm: Friday Feb
More informationFrom perceptrons to word embeddings. Simon Šuster University of Groningen
From perceptrons to word embeddings Simon Šuster University of Groningen Outline A basic computational unit Weighting some input to produce an output: classification Perceptron Classify tweets Written
More informationContents. (75pts) COS495 Midterm. (15pts) Short answers
Contents (75pts) COS495 Midterm 1 (15pts) Short answers........................... 1 (5pts) Unequal loss............................. 2 (15pts) About LSTMs........................... 3 (25pts) Modular
More information<Special Topics in VLSI> Learning for Deep Neural Networks (Back-propagation)
Learning for Deep Neural Networks (Back-propagation) Outline Summary of Previous Standford Lecture Universal Approximation Theorem Inference vs Training Gradient Descent Back-Propagation
More informationIntroduction to Convolutional Neural Networks 2018 / 02 / 23
Introduction to Convolutional Neural Networks 2018 / 02 / 23 Buzzword: CNN Convolutional neural networks (CNN, ConvNet) is a class of deep, feed-forward (not recurrent) artificial neural networks that
More informationMachine Learning Basics
Security and Fairness of Deep Learning Machine Learning Basics Anupam Datta CMU Spring 2019 Image Classification Image Classification Image classification pipeline Input: A training set of N images, each
More informationThe Origin of Deep Learning. Lili Mou Jan, 2015
The Origin of Deep Learning Lili Mou Jan, 2015 Acknowledgment Most of the materials come from G. E. Hinton s online course. Outline Introduction Preliminary Boltzmann Machines and RBMs Deep Belief Nets
More informationIndex. Santanu Pattanayak 2017 S. Pattanayak, Pro Deep Learning with TensorFlow,
Index A Activation functions, neuron/perceptron binary threshold activation function, 102 103 linear activation function, 102 rectified linear unit, 106 sigmoid activation function, 103 104 SoftMax activation
More informationRecurrent Neural Networks. Jian Tang
Recurrent Neural Networks Jian Tang tangjianpku@gmail.com 1 RNN: Recurrent neural networks Neural networks for sequence modeling Summarize a sequence with fix-sized vector through recursively updating
More informationA Convolutional Neural Network-based
A Convolutional Neural Network-based Model for Knowledge Base Completion Dat Quoc Nguyen Joint work with: Dai Quoc Nguyen, Tu Dinh Nguyen and Dinh Phung April 16, 2018 Introduction Word vectors learned
More informationRegularization Introduction to Machine Learning. Matt Gormley Lecture 10 Feb. 19, 2018
1-61 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Regularization Matt Gormley Lecture 1 Feb. 19, 218 1 Reminders Homework 4: Logistic
More informationBayesian Networks (Part I)
10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Bayesian Networks (Part I) Graphical Model Readings: Murphy 10 10.2.1 Bishop 8.1,
More informationLecture 11 Recurrent Neural Networks I
Lecture 11 Recurrent Neural Networks I CMSC 35246: Deep Learning Shubhendu Trivedi & Risi Kondor University of Chicago May 01, 2017 Introduction Sequence Learning with Neural Networks Some Sequence Tasks
More informationIntroduction to Deep Learning CMPT 733. Steven Bergner
Introduction to Deep Learning CMPT 733 Steven Bergner Overview Renaissance of artificial neural networks Representation learning vs feature engineering Background Linear Algebra, Optimization Regularization
More informationDeep Convolutional Neural Networks for Pairwise Causality
Deep Convolutional Neural Networks for Pairwise Causality Karamjit Singh, Garima Gupta, Lovekesh Vig, Gautam Shroff, and Puneet Agarwal TCS Research, Delhi Tata Consultancy Services Ltd. {karamjit.singh,
More information(Feed-Forward) Neural Networks Dr. Hajira Jabeen, Prof. Jens Lehmann
(Feed-Forward) Neural Networks 2016-12-06 Dr. Hajira Jabeen, Prof. Jens Lehmann Outline In the previous lectures we have learned about tensors and factorization methods. RESCAL is a bilinear model for
More informationDeep Spatio-Temporal Time Series Land Cover Classification
Bachelor s Thesis Exposé Submitter: Arik Ermshaus Submission Date: Thursday 29 th March, 2018 Supervisor: Supervisor: Institution: Dr. rer. nat. Patrick Schäfer Prof. Dr. Ulf Leser Department of Computer
More informationRegression Adjustment with Artificial Neural Networks
Regression Adjustment with Artificial Neural Networks Age of Big Data: data comes in a rate and in a variety of types that exceed our ability to analyse it Texts, image, speech, video Real motivation:
More informationIncorporating Pre-Training in Long Short-Term Memory Networks for Tweets Classification
Incorporating Pre-Training in Long Short-Term Memory Networks for Tweets Classification Shuhan Yuan, Xintao Wu and Yang Xiang Tongji University, Shanghai, China, Email: {4e66,shxiangyang}@tongji.edu.cn
More informationModelling Time Series with Neural Networks. Volker Tresp Summer 2017
Modelling Time Series with Neural Networks Volker Tresp Summer 2017 1 Modelling of Time Series The next figure shows a time series (DAX) Other interesting time-series: energy prize, energy consumption,
More informationJointly Extracting Event Triggers and Arguments by Dependency-Bridge RNN and Tensor-Based Argument Interaction
Jointly Extracting Event Triggers and Arguments by Dependency-Bridge RNN and Tensor-Based Argument Interaction Feng Qian,LeiSha, Baobao Chang, Zhifang Sui Institute of Computational Linguistics, Peking
More informationStatistical Machine Learning
Statistical Machine Learning Lecture 9 Numerical optimization and deep learning Niklas Wahlström Division of Systems and Control Department of Information Technology Uppsala University niklas.wahlstrom@it.uu.se
More informationPart-of-Speech Tagging + Neural Networks 3: Word Embeddings CS 287
Part-of-Speech Tagging + Neural Networks 3: Word Embeddings CS 287 Review: Neural Networks One-layer multi-layer perceptron architecture, NN MLP1 (x) = g(xw 1 + b 1 )W 2 + b 2 xw + b; perceptron x is the
More informationClassification of Higgs Boson Tau-Tau decays using GPU accelerated Neural Networks
Classification of Higgs Boson Tau-Tau decays using GPU accelerated Neural Networks Mohit Shridhar Stanford University mohits@stanford.edu, mohit@u.nus.edu Abstract In particle physics, Higgs Boson to tau-tau
More informationEE-559 Deep learning Recurrent Neural Networks
EE-559 Deep learning 11.1. Recurrent Neural Networks François Fleuret https://fleuret.org/ee559/ Sun Feb 24 20:33:31 UTC 2019 Inference from sequences François Fleuret EE-559 Deep learning / 11.1. Recurrent
More informationDeep Learning intro and hands-on tutorial
Deep Learning intro and hands-on tutorial Π passalis@csd.auth.gr Ε. ώ Π ώ. Π ΠΘ 1 / 53 Deep Learning 2 / 53 Δ Έ ( ) ω π μ, ώπ ώ π ω (,...), π ώ π π π π ω ω ώπ ώ π, (biologically plausible) Δ π ώώ π ώ Ε
More informationIdentifying QCD transition using Deep Learning
Identifying QCD transition using Deep Learning Kai Zhou Long-Gang Pang, Nan Su, Hannah Peterson, Horst Stoecker, Xin-Nian Wang Collaborators: arxiv:1612.04262 Outline 2 What is deep learning? Artificial
More informationMachine Learning Algorithm. Heejun Kim
Machine Learning Algorithm Heejun Kim June 12, 2018 Machine Learning Algorithms Machine Learning algorithm: a procedure in developing computer programs that improve their performance with experience. Types
More informationWhat s so Hard about Natural Language Understanding?
What s so Hard about Natural Language Understanding? Alan Ritter Computer Science and Engineering The Ohio State University Collaborators: Jiwei Li, Dan Jurafsky (Stanford) Bill Dolan, Michel Galley, Jianfeng
More informationTwo-Stream Bidirectional Long Short-Term Memory for Mitosis Event Detection and Stage Localization in Phase-Contrast Microscopy Images
Two-Stream Bidirectional Long Short-Term Memory for Mitosis Event Detection and Stage Localization in Phase-Contrast Microscopy Images Yunxiang Mao and Zhaozheng Yin (B) Computer Science, Missouri University
More informationRecurrent neural networks
12-1: Recurrent neural networks Prof. J.C. Kao, UCLA Recurrent neural networks Motivation Network unrollwing Backpropagation through time Vanishing and exploding gradients LSTMs GRUs 12-2: Recurrent neural
More informationRecurrent Neural Networks Deep Learning Lecture 5. Efstratios Gavves
Recurrent Neural Networks Deep Learning Lecture 5 Efstratios Gavves Sequential Data So far, all tasks assumed stationary data Neither all data, nor all tasks are stationary though Sequential Data: Text
More informationMultilayer Neural Networks. (sometimes called Multilayer Perceptrons or MLPs)
Multilayer Neural Networks (sometimes called Multilayer Perceptrons or MLPs) Linear separability Hyperplane In 2D: w x + w 2 x 2 + w 0 = 0 Feature x 2 = w w 2 x w 0 w 2 Feature 2 A perceptron can separate
More informationAnalysis of Railway Accidents Narratives Using Deep Learning
Analysis of Railway Accidents Narratives Using Deep Learning Mojtaba Heidarysafa, Kamran Kowsari, Laura E Barnes, and Donald E Brown Department of System and Information Engineering, University of Virginia,
More informationIntroduction to RNNs!
Introduction to RNNs Arun Mallya Best viewed with Computer Modern fonts installed Outline Why Recurrent Neural Networks (RNNs)? The Vanilla RNN unit The RNN forward pass Backpropagation refresher The RNN
More informationCS 179: LECTURE 16 MODEL COMPLEXITY, REGULARIZATION, AND CONVOLUTIONAL NETS
CS 179: LECTURE 16 MODEL COMPLEXITY, REGULARIZATION, AND CONVOLUTIONAL NETS LAST TIME Intro to cudnn Deep neural nets using cublas and cudnn TODAY Building a better model for image classification Overfitting
More informationAPPLIED DEEP LEARNING PROF ALEXIEI DINGLI
APPLIED DEEP LEARNING PROF ALEXIEI DINGLI TECH NEWS TECH NEWS HOW TO DO IT? TECH NEWS APPLICATIONS TECH NEWS TECH NEWS NEURAL NETWORKS Interconnected set of nodes and edges Designed to perform complex
More information