Demo 2: Deep Learning for NLP - Sentiment Analysis

Christophe Servan


Introduction

In this Demo, you will build a sentiment analysis system for Twitter. The task consists in predicting whether a tweet is positive, negative or neutral regarding its content. Typically, sentiment polarity is conveyed by a combination of factors: Therefore, typical sentiment analysis systems rely on word lists for capturing subjectivity and sentiment bearing words, and word n-grams or parse trees for capturing modulation effects. Since the quantity of data available for training is rather scarce, a number of approaches can be applied for making a system that generalizes well.
For instance, words can be replaced by part-of-speech tags, or word classes. Sentiment-bearing word-lists can be extended through collocation statistics. Tweets have a few specificities that make the sentiment analysis task more challenging. They contain very informal language with a lot of typos and colloquial language. Smileys can be an extra source of information but they are highly variable. Hashtags and user names are marked (along with other Twitter specific abbreviations). Word morphology can be modified to express sentiment (such as "loooooooool"). Therefore, formal linguistic analysis such as parsing and word lists are unlikely to be very helpful, but deep learning approaches can help dealing with that variability.

Data


Report

You should send me your report the day before the next session using PDF format at the address christophe_dot_servan_at_epita_dot_fr (replace _dot_ and _at_ by dots and at, respectively)

Important remarks:


Work to do

      This demo will train you to manipulate a classifier using a neural network approach

I. Read data

    In data downloaded you will use the file sanders-twitter-sentiment.csv. This file contains tweets associated to 4 classes: "positive", "negative", "neutral" or "irrelevant". The main idea is to train a sentiment analysis system considering this task as a classification task. The file contains 5,513 annotated tweets in comma separated values format. Only the fourth and the fifth columns interest us in this demo.

      Exercice 1: Load and process data.

        Q1: Load the CSV file using the csv import in python. (1pt)

        Q2: Produce a python list of pre-processed tweets using the script preprocess_twitter.py (1pt)

      Exercice 1: Vectorize data.

      For vectorizing tweet texts, first you can use the tokenizer provided in keras. It splits sentences into words and maps them to ids between 1 and the number of words in the lexicon. The tokenizer can be restricted to the most frequent words in the texts with the nb_words parameter.

      Then, the next stage consists in padding the sequences with word 0 to a given maximum length (trimming them if they are longer than the provided maxlen parameter) with the pad_sequences function. It returns a matrix of shape (number of examples, max length) which contains integers (word ids). The word mapping can be found in tokenizer.word_index.

        Q1: Use these python lines to process data and labels: (1pt)

        from keras.preprocessing.text import Tokenizer
        from keras.preprocessing.sequence import pad_sequences
        
        tokenizer = Tokenizer(nb_words=10000)
        tokenizer.fit_on_texts(texts)
        sequences = tokenizer.texts_to_sequences(texts)
        data = pad_sequences(sequences, maxlen=32)
        vocab = tokenizer.word_index
        vocab['<eos>'] = 0                
        

        Q2: Now you should have two numpy arrays which shape are (5513, 32) and (5513, 4) for tweets and labels respectively. (1pt)

        Now cut this coprus in two parts one for the training and one for the evaluation. The training set (x_train and y_train) should contains 4000 utterances and the validation set (x_val, y_val) should have the rest of the data. Feel free to shuffle data.

II. Train and evaluate the model

    Unlike in the previous demo, you will not train word embeddings (WE), but use pre-trained WE. These WE are contained in the file glove.twitter.27B.100d.filtered.txt They have been provided by Stanford though the GloVe project.

    Use the embeddings.py module to load the WE into a numpy matrix:

    weights = embedding.load(vocab, 100,'glove.twitter.27B.100d.filtered.txt')
    

    The embedding layer can be initialized with the embeddings we loaded from the web. We won't finetune them as we want to use them with words that are not in the training data. For recurrent unit, we will use Gated Recurrent Units (GRU) because they have less parameters and are faster to train than LSTM, but the later would also be a good choice. Their size is fixed to 64, but that's an hyperparameter that could be modified to improve the classifier.

      Exercice 1: Use RNN to train the first classifier:

      from keras.layers import Embedding, Input, GRU, Dense
      from keras.models import Model
      # 100-dim embeddings initialized with GloVe,
      # over sequences of size 32, and not fine tuneable
      embedding_layer = Embedding(len(vocab), 100, weights=[weights],input_length=32, trainable=False)
      sequence_input = Input(shape=(32,), dtype='int32')
      embedded_sequences = embedding_layer(sequence_input)
      x = GRU(64)(embedded_sequences)
      preds = Dense(labels.shape[1], activation='softmax')(x)
      model = Model(sequence_input, preds)
      model.compile(loss='categorical_crossentropy',optimizer='Nadam', metrics=['acc'])
      model.fit(x_train, y_train, validation_data=(x_val, y_val),nb_epoch=10, batch_size=64, shuffle=True)            
      

        Q1: write in Python the necessary script train the model (3pt)

        Q2: evaluate the model using precision, recall, F1-measure, and accuracy metrics. (4pt)

        Q3: How many epochs are needed to reach at least 80% of accuracy? (2pt)

      Exercice 2: Use CNN to train the second classifier:

      The recurrent model has the downside that its hidden state of fixed size might be suboptimal for building representations of the beginning of the input, and might be too order-dependent. An alternative to this model is convolutional neural networks. They consist of a convolution filter which is repeated over a window moving along the input. They act a bit like a bag of n-grams which can select relevant word n-grams. In the following example, we use 128 filters of size of 3 over the input, with global max pooling, and then atten the results and pass it through a dense layer with a softmax activation to generate the probability distribution over labels :

      from keras.layers import Conv1D, MaxPooling1D, Flatten
      
      sequence_input = Input(shape=(32,), dtype='int32')
      embedded_sequences = embedding_layer(sequence_input)
      x = Conv1D(128, 3, activation='relu')(embedded_sequences)
      x = MaxPooling1D(3)(x) # global max pooling
      x = Flatten()(x)
      preds = Dense(labels.shape[1], activation='softmax')(x)
      
      model = Model(sequence_input, preds)
                 
      

      The model should be a bit faster to train than the recurrent one, but it leads to a lower accuracy. In particular, as accuracy on the training set raises to 99%, it reaches a high value on the validation set and then diminishes in later epochs. Clearly, the model is over fitting the training data, leading to poor generalization performance. It might be a good idea to extend that model with a regularization technique, such as for instance dropout.

        Q1: write in Python the necessary script train the model (3pts)

        Q2: evaluate the model using precision, recall, F1-measure, and accuracy metrics. (4pts)

        Q3: (optional) Add a regularization method to the code and observe the results. (3pts)



The End.