Skip to content

FrancescoBriuolo/FakeNewsPropagation

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

20 Commits
 
 
 
 
 
 

Repository files navigation

Unisannio

Python Unisannio

Evaluation of graph-based techniques for the analysis of fake news and propagation metrics

This thesis project entitled “Evaluation of graph-based techniques for the analysis of fake news and propagation metrics” involved two parallel projects. The first consists of extracting topic-specific data from the fact-checking site PolitiFact by web scraping. These data are used to construct a bipartite graph of sources and keywords, which is analyzed by considering relative projections in order to extract information about the veracity of the news. The second experiment is a study of a hierarchical propagation network for fake news detection. The goal is to use network features to implement multiple machine learning models and a Feedforward neural network for fake news detection using data provided by the FakeNewsNet dataset and perform a sensitivity analysis of the classification algorithms used with respect to the features considered. ​

Bipartite graph

This project contains three Python's scripts. Two of these (init_archives.py and update_archives.py) are for the scraping activity from the fact-checking website "Politifact". The "init_archives.py" script enables the creation of two CSV files. Entering a keyword creates the 'archive_with_keywords.csv' file, which contains the latest news related to the first n pages (where n is an input parameter) for search with the keyword entered. If a keyword is not entered (""), the 'archive.csv' file is created and populated with the most recent statements for the first n pages specified. The "update_archives.py" script enables to update the previously created files. Entering a keyword updates the 'archive_with_keywords.csv' file adding to the top of the list all the most recent stuff relative to the keywords with which the search was performed. If a keyword is not entered (""), the 'archive.csv' file is updated with the most recent statements for the first n pages specified. It is necessary to specify that for the purposes of this thesis project only searches using keywords were conducted, although the other option is useful for possible future development. The CSV file 'archive_with_keywords.csv' consists of 6 fields: author (the fact-checker), statement (news content), source (source of the news - person, social post, organization), date (of the review), target (label assigned by the fact-checker: True, Mostly True, Half True, Mostly False, False, Pants on Fire), keyword (with which the search is executed).

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published