Can we avoid repeated questions by identifying similar intent?
Inhibiting repeated questions can help avoid duplication of efforts
- Introduction/Background
- Problem Definition
- Methods
- Final results and analysis
- Conclusion
- Future experiments
- Timeline and Responsibilities
- References
Course project for CS 7641
Introduction/Background
With public discussion forums becoming highly popular in recent times, especially with classes and work moving to an online setup, websites like Edstem, Piazza, Stack Overflow and Quora have a constant influx of a large volume of questions. Multiple questions with the same intent leads to redundancy, and waste of time, space and resources.
Problem Definition
The aim is to identify and flag questions with a high similarity index, and retain only canonical questions in order to make the work of administrative staff such as TAs/Professors easier in terms of answering repeated questions or questions of the same intent.
Methods
- Dataset: The dataset that we will be using is the Quora Question Pairs dataset[1]. It consists of 404,290 pairs of questions. Each datapoint consists of a pair of questions and whether or not they are similar.
In order to gain useful insights, we visualize and understand patterns in our data using the following plots.
Fig 1: Word Cloud depicting which words appear more frequently in the dataset
Fig 2: Histogram representing the number of characters in each question
Fig 3: Histogram representing the number of words in each question
Fig 4: Plots representing the distribution of ratio of words shared between similar and dissimlar question
- Data Augmentation:
- We exploit the transitive property of similarity to generate new datapoints.
- If Question Q1 is similar to Question Q2 and Question Q2 is similar to Question Q3, then we can infer that Q1 is similar to Q3
- If Question Q1 is similar to Question Q2 and Question Q2 is not similar to Question Q3, then we can infer that Q1 is not similar to Q3
Through this method, we generated 25% more training datapoints.
- We exploit the transitive property of similarity to generate new datapoints.
- Data preprocessing:
- We used the following preprocessing techniques:
- Sample Sentence: “How can internet speed be increased by hacking through DNS?”
- Removing punctuation
- After Punctuation Removal: “How can internet speed be increased by hacking through DNS” 2. Converting characters to Lower Case
- After conversion: “how can internet speed be increased by hacking through dns” 3. Tokenization
- Tokenization is the process of converting the sentence into a list of it’s constituent words/phrases. It helps reduce the input to a finite set, making it easier to process.
- After Tokenization: [‘how’, ‘can’, ‘internet’, ‘speed’, ‘be’, ‘increased’, ‘by’, ‘hacking’, ‘through’, ‘dns’] 4. Stop Word Removal
- The idea of Stop Word Removal is to remove commonly occuring words in the corpus.
- After Stop Word Removal: [‘internet’, ‘speed’, ‘increased’, ‘hacking’, ‘dns’] 5. POS Tagging
- POS tagging refers to categorizing words in a corpus corresponding to a particular part of speech (eg: Noun, Verb, Adjective etc). This process helps improve the performance of the Lemmatizer.
- After POS Tagging: [(‘internet’, ‘a’), (‘speed’, ‘n’), (‘increased’, ‘v’), (‘hacking’, ‘v’), (‘dns’, ‘n’)] 6. Lemmatization
- Lemmatization is grouping together different inflected forms of a word so that they can be analyzed as a single item.
- After Lemmatization: [‘internet’, ‘speed’, ‘increase’, ‘hack’, ‘dns’]
- Sample Sentence: “How can internet speed be increased by hacking through DNS?”
- We used the following preprocessing techniques:
- Training:
- For each training iteration, we input questions in a pairwise fashion - 𝑄𝑖, 𝑄𝑗.
- The model learns representations of building blocks of both sentences𝑄𝑖−>ν𝑖, 𝑄𝑗−>ν𝑗.
- Next, these representations are concatenated 𝑣𝑐𝑜𝑛 = 𝑐𝑜𝑛𝑐𝑎𝑡(ν𝑖, ν𝑗), and passed on to a feedforward neural network or a machine learning model 𝐹 (𝑣𝑐𝑜𝑛) that predicts whether two questions are similar or not.
- ∀ questions 𝑄𝑖 ε question bank 𝑄𝐵, we group them into clusters 𝑐1,…,𝑐n for efficient inference.
Supervised method: Classifying sentences as similar or not using BERT-based model (Added for midpoint report)
- The tokenized version of the candidate sentence that we get from the preprocessing stage is first passed to a pre-trained BERT model.
- The output of this pre-trained BERT model is the sentence-level embedding of the tokenized version - E1.
- The same process is done on the second candidate sentence to retrieve its sentence-level embedding - E2.
- The two embeddings are concatenated, and then passed to a feed-forward neural network which consists of a fully connected layer, a ReLu layer, another fully connected layer, and finally a softmax layer in the order specified.
- The softmax layer then outputs the probabilities of the pair of candidate questions being similar. Based on the probabilities we classify the questions as similar or not similar.
Fig 5: BERT takes in sentences as inputs and gives sentence-level embeddings as the output
- Inference:
- For a query question 𝑄𝑞 we identify the cluster 𝑐𝑖 it belongs to.
- For all candidate questions 𝑄𝑑 belonging to cluster 𝑐𝑖, we find similarity 𝑠𝑖𝑚(𝑄𝑑, 𝑄𝑞). With clustering we avoid finding similarities with all questions in the question bank, making inference efficient.
- If for any 𝑄𝑑 ε 𝑐𝑖, if 𝑠𝑖𝑚(𝑄𝑑, 𝑄𝑞) > 𝑡ℎ𝑟𝑒𝑠ℎ𝑜𝑙𝑑, we flag that question as similar.
- We also output 𝑡𝑜𝑝 − 𝑘 similar questions based on 𝑠𝑖𝑚(𝑄𝑑, 𝑄𝑞).
- Models in consideration:
- Models for learning representations
- BERT
- GPT
- XLNet
- Models for clustering
- K-means
- Models for finding similarity
- Feedforward neural network (FFN)
- Models for learning representations
Final results and analysis
Supervised Learning Pipeline
Experiment 1 : Data Preprocessing and Augmentation Choices Ablation
Through the first set of experiments, we try to observe the effects of data preprocessing and data augmentation on the performance of the model. We also apply three different types of models on a sample dataset of 1000 data points from the entire original dataset. The results of these are seen in the tables below:
Data Preprocessing | Data Augmentation | Model | Train accuracy | Train F1 score | Test accuracy | Test F1 score |
---|---|---|---|---|---|---|
No | No | Baseline | 0.57 | 0.61 | 0.42 | 0.58 |
Yes | No | Baseline | 0.61 | 0.61 | 0.57 | 0.62 |
Yes | Yes | Baseline | 0.6 | 0.62 | 0.61 | 0.6 |
Data Preprocessing | Data Augmentation | Model | Train accuracy | Train F1 score | Test accuracy | Test F1 score |
---|---|---|---|---|---|---|
No | No | Static BERT | 0.66 | 0.61 | 0.57 | 0.53 |
Yes | No | Static BERT | 0.67 | 0.62 | 0.59 | 0.57 |
Yes | Yes | Static BERT | 0.66 | 0.62 | 0.59 | 0.55 |
Data Preprocessing | Data Augmentation | Model | Train accuracy | Train F1 score | Test accuracy | Test F1 score |
---|---|---|---|---|---|---|
No | No | Fine-tuned BERT | 0.98 | 0.97 | 0.57 | 0.41 |
Yes | No | Fine-tuned BERT | 0.97 | 0.97 | 0.58 | 0.5 |
Yes | Yes | Fine-tuned BERT | 0.97 | 0.96 | 0.64 | 0.52 |
Here, the baseline model is nothing but the pre-trained BERT-based model, and the embeddings of the sentences obtained from this model are compared using a cosine similarity. This provided us a good baseline result and starting point, which we tried to improve upon.
For the other two models, the dataset was then split into training and testing data with a ratio of 80:20 respectively. The models were trained over this data for 5 epochs with a learning rate of 0.0001 and batch size of 8. The architecture used for these models in this experiment : BERT -> FFN -> Sigmoid layer -> FFN -> Softmax layer. In the Static BERT model the pre-trained BERT model is kept static and the remaining layers are trained, whereas in the Fine-tuned BERT, all the layers are trained and the pre-trained BERT-based model is fine-tuned on our dataset.
Inference : As we can see from the table, applying data preprocessing techniques (mentioned in section 2) and data augmentation techniques increases the accuracy on the test data. This empirically backs our decision to do data preprocessing and augmentation.
Experiment 2 : Fine Tuning Ablation
Through the second set of experiments, we tried to gain insight on which model would work best for our problem. As a part of this, we ran our baseline, Static BERT, and Fine-tuned BERT on a sample dataset of 10000 datapoints randomly chosen from the original dataset. For the Static BERT and Fine-tuned BERT models we again split the data into train and test with a 80:20 ratio, and ran them for 10 epochs with a learning rate of 0.0001 and batch size 32. The architecture used for these 2 models in this experiment : BERT -> FFN -> PRelu -> FFN -> PRelu -> FFN -> Softmax Layer. The results are seen in the table below:
Data Preprocessing | Data Augmentation | Model | Train accuracy | Train F1 score | Test accuracy | Test F1 score |
---|---|---|---|---|---|---|
Yes | Yes | Baseline | 0.59 | 0.6 | 0.6 | 0.61 |
Yes | Yes | Static BERT | 0.7 | 0.63 | 0.68 | 0.61 |
Yes | Yes | Fine-tuned BERT | 0.99 | 0.98 | 0.71 | 0.62 |
Inference : As it can be seen in the table, the Fine-tuned BERT model produces the best accuracy results on the data as compared to the baseline and Static-BERT. This again empirically backs our decision to choose a pre-trained BERT model and fine-tune it on our dataset.
Experiment 3 : Model Selection Ablation
We chose to fine-tune our models based on the results from experiment 2. Next, we did some hyperparamter tuning to pick the best combination of hyperparameters for performance. We found that a batch size of 128 and a training over 5 epochs gave the best test results for our Fine-tuned BERT model. We then trained all 400,000 question pairs with this hyperparameter setting. We repeated the same process with GPT and XLNet architectures. The results for this experiment are shown below:
Model | Train accuracy | Test accuracy |
---|---|---|
Fine-tuned BERT | 0.94 | 0.86 |
Fine-tuned GPT-2 | 0.65 | 0.63 |
Fine-tuned XLNet | 0.64 | 0.64 |
Inference : As it can be seen in the table, the Fine-tuned BERT model produces the best accuracy results on the test data as compared to the GPT-2 and XLNet models. Thus, we present the Fine-tuned BERT model as our final supervised learning model with a test accuracy of 86%.
Unsupervised Learning Pipeline for Efficient Inference
We build an unsupervised learning pipeline in order to cluster similar questions. The pipeline consists of 3 steps :
- The first part of the pipeline consists of creating BERT embeddings of all the questions after preprocessing them.
-
We then perform Principal Component Analysis (PCA) in order to reduce the dimensionality of our dataset. From the below graph, we can infer that with just 15 components, 99 percent of the variance is retained.
-
We cluster the preprocessed dataset using Kmeans clustering algorithm. The elbow method plotted below, is used to find out the ideal number of clusters after analysing the tradeoff between performance and computation cost. From the graph, we can infer that the elbow is at 100 clusters
This unsupervised learning pipeline is used to create a smaller question space from which we can choose the candidate reference questions. This in turn reduces the time overhead by 100 times during inference. Here is an example illustrating potential canonical questions:
Conclusion
In this project, we propose a method to classify questions as similar or not similar using a combination of a supervised method and unsupervised method.
For the supervised method, we implemented multiple transformer based models (such as BERT, GPT-2, and XLNet) as backbones over a feed forward neural network to classify whether two questions are similar or not. After extensive results and ablations we were able to finalize our model as the pre-trained BERT model which is fine-tuned on our data. This supervised model classifies the given query question as similar or not after comparison with the top K reference questions obtained from the unsupervised pipeline based on their sentence embeddings. We achieve a test accuracy of 86% for this task.
For the unsupervised method, we use K-means clustering to group the questions in the dataset based on their sentence level embeddings, in order to narrow down the question space from which the reference questions are picked for comparison. The unsupervised pipeline gives us the candidate similar questions for each of the clusters. Through this, we were able to form 100 clusters which reduces our overhead significantly.
Future experiments
Future work may include downstream tasks such as automatic question tagging, and personalized recommendation of questions based on the field of interest. This can be very advantageous for forums such as Ed, Quora, Piazza, etc. and can help towards making the work of teaching staff easier.
Timeline and Responsibilities
References
- https://quoradata.quora.com/First-Quora-Dataset-Release-Question-Pairs
- D. A. Prabowo and G. Budi Herwanto, “Duplicate Question Detection in Question Answer Website using Convolutional Neural Network,” 2019 5th International Conference on Science and Technology (ICST), 2019, pp. 1-6, doi: 10.1109/ICST47872.2019.9166343.
- Jacob Devlin, Ming-Wei Chang, Kenton Lee, Kristina Toutanova, “BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding”, arXiv:1810.
- Tianqi Chen andCarlos Guestrin.2016.XGBoost: A Scalable Tree Boosting System. Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD ‘16). Association for Computing Machinery, New York,NY,USA,785–794.DOI:https://doi.org/10.1145/2939672.785.
- Reynolds D.(2009)GaussianMixture Models. In:Li S.Z.,Jain A. (eds) Encyclopedia of Biometrics. Springer, Boston, MA. https://doi.org/10.1007/978-0-387-73003-5_