Please use this identifier to cite or link to this item:
https://elibrary.tucl.edu.np/handle/123456789/7686
Title: | HYBRID NEURAL NETWORK FOR FAKE NEWS STANCE DETECTION |
Authors: | GHIMIRE, NIROJ |
Keywords: | Stance Detection;;Fake News;;NLP;Hybrid Neural Network;;Encoder Decoder Architecture |
Issue Date: | Aug-2021 |
Publisher: | Pulchowk Campus |
Institute Name: | Institute of Engineering |
Level: | Masters |
Citation: | MASTER OF SCIENCE IN INFORMATION AND COMMUNICATION ENGINEERING |
Abstract: | Fake news is becoming more readily available as technology advances, which sometime mislead the readers and leads to inaccurate social opinions. Fake news may be found on the Internet, news sources and social media platforms. The spread of low-quality news has harmed both individuals and society. In this thesis work, we analyze three hybrid models, CNN+simple RNN, CNN+GRU and CNN+BiLSTM in encoder decoder architecture to predict the stance between headline and article of the news. Pre-trained GloVe word embedding is used for word to vector representation as it can capture the inter-word semantic information. The CNN-RNN combination had been shown efficient in deep learning applications because they can capture sequential and local features of input data. The models were successfully trained and tested on both binary (ISOT) and multiclass (FNC-1) fake news datasets. It is found that the CNN+ BiLSTM model had better results than other two hybrid models in both binary and multiclass classification task for the fake news stance detection system. |
Description: | Fake news is becoming more readily available as technology advances, which sometime mislead the readers and leads to inaccurate social opinions. |
URI: | https://elibrary.tucl.edu.np/handle/123456789/7686 |
Appears in Collections: | Electronics and Computer Engineering |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
Thesis_Final_Report__Copy_final.pdf | 1.78 MB | Adobe PDF | View/Open |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.