In the present era of the distributed system where almost the complete world has been engaged in social networking, how one can claim that he/she get the real authenticated content. Since the content on the internet is not verified especially the social media content where the people post mostly the doubtful information. The main difficult problem is the filtering of truth from such contents. In such a situation social media find the new challenge of establishing veracity (doubtable data). So a new system PHEME is going to establish for analyzing the content in social sites, blogs and socially related posts based on the language and determine the uncertainty or the doubts in the content. This system will help not only in medical information systems (where causes serious damages if the wrong information held) but also in digital journalism.
Table of Contents
1. Introduction
2. Identification
3. Classification of Information
4. Architecture
5. Detection
6. Conclusion
7. References
Research Objectives and Themes
The primary objective of this research is to address the challenge of filtering authentic information from doubtful content on social media networks. By introducing the PHEME system, the study aims to analyze social posts based on language, semantics, and history to identify veracity and mitigate the risks of misinformation in fields such as medical systems and digital journalism.
- Analysis of veracity in social media and online content.
- Classification of information into speculation, controversy, misinformation, and disinformation.
- Technical implementation of the PHEME system architecture.
- Use of Natural Language Processing (NLP) and data mining for real-time detection.
- Demonstration of information verification through query-based visual analysis.
Excerpt from the Book
1. Introduction
In the present era of the distributed system where almost the complete world has been engaged in social networking, how one can claim that he/she get the real authenticated content. Since the content on the internet is not verified especially the social media content where the people post mostly the doubtful information. The main difficult problem is the filtering of truth from such contents. In such a situation social media find the new challenge of establishing veracity (doubtable data). So a new system PHEME is going to establish for analyzing the content in social sites, blogs and socially related posts based on the language and determine the uncertainty or the doubts in the content. This system will help not only in medical information systems (where causes serious damages if the wrong information held) but also in digital journalism.
The system is under developing phase which claims to check the posts on social networks like Facebook and Twitter and then only permit to publish on Webpage. The project is running by International Group of Researches under the University of Sheffield as part of GATE, the University of Warwick, and King’s College London, Saarland University in Germany and MODUL University Vienna.
Summary of Chapters
1. Introduction: Discusses the prevalence of doubtful information on social networks and introduces the PHEME project as a solution for verifying content veracity.
2. Identification: Outlines the combination of technologies like NLP, Data Mining, and Information Visualization used to identify the realism of information.
3. Classification of Information: Defines the four primary categories of fake information: speculation, controversy, misinformation, and disinformation.
4. Architecture: Explains the three-layer system design, comprising the front end, back end, and API, along with the integration of MVC and bootstrap framework.
5. Detection: Describes the practical application of the system using Twitter to demonstrate how queries reveal information credibility via visual graphs.
6. Conclusion: Summarizes the potential of the project to assist government and emergency services in responding to content and notes the future development path.
7. References: Lists the academic papers, technical documentation, and web sources utilized throughout the research.
Keywords
Veracity, Fake Posts, Social Networking, PHEME, NLP, Data Mining, Misinformation, Disinformation, WebLyzard, GATE, OWLIM, Semantic Analysis, Digital Journalism, Information Diffusion, System Architecture.
Frequently Asked Questions
What is the fundamental goal of this research?
The work aims to develop a system capable of filtering truth from unverified and potentially harmful content on social media platforms to ensure safer online communication.
What are the primary thematic areas covered?
The research covers social media veracity, information classification, system architecture for data analysis, and the application of intelligent algorithms to detect misinformation.
What is the core research question?
The research asks how to effectively establish the authenticity of information in a distributed environment where users frequently post unverified or doubtful content.
Which scientific methods are employed?
The study utilizes a combination of Natural Language Processing (NLP), Data Mining, Linked Open Data (LOD) cross-referencing, and semantic analysis integrated within the GATE text mining platform.
What content is discussed in the main part of the paper?
The main part details the identification steps, the categorization of fake information, the multi-layered technical architecture (front-end, back-end, API), and the demonstration of the detection process.
Which keywords characterize this work?
Key terms include Veracity, Social Networking, PHEME, Misinformation, Disinformation, Semantic Analysis, and System Architecture.
How does the PHEME system classify fake information?
The system classifies information into four categories based on semantics and language: speculation (unverified statements), controversy (disputed messages), misinformation (untrue content spread unwittingly), and disinformation (content spread with malicious intent).
What role does the architecture play in the system's flexibility?
The architecture is built on a modular three-layer design (front-end, back-end, API) that allows for future enhancements through an event bus and enables the system to be easily deployed with existing applications.
Why is Twitter used for the demonstration in the detection section?
Twitter is used as a case study for the sake of simplicity to demonstrate how a user can input a query and receive a comprehensive visual analysis of the information's credibility.
- Citation du texte
- M.Tech Hemant Kumar Saini (Auteur), Nitesh Chouhan (Auteur), 2014, Detecting Veracity, Munich, GRIN Verlag, https://www.grin.com/document/285280