Mitigating/Grading Misinformation and Disinformation with Hybrid Model AI

By David Sun* Eric Sun* Eric Gu* from University of California San Diego
Advisor: Ali Arsanjani from Google Cloud

*indicates equal contribution

GenAI For Good

About the project

The proliferation of fake news and misinformation, often amplified by large language models (LLMs), poses a significant threat to societal trust and stability. This paper introduces a hybrid veracity detection and scoring framework that leverages both generative AI and traditional machine learning to detect, rank, and mitigate misinformation and disinformation across diverse media formats. Our approach decomposes content into structured analytical components, using an ensemble of factuality factors such as frequency heuristics, malicious account indicators, and psychological manipulation cues to identify and assess deceptive patterns. By employing advanced techniques such as Retrieval-Augmented Generation (RAG), fractal chain-of-thought prompting, and function calling, our system dynamically refines predictions, enhancing transparency and reducing hallucinations. This hybridized LLM-based veracity machine not only facilitates precise misinformation detection but also provides a scalable and interpretable solution for managing the complexities of content veracity in an evolving digital landscape.

Codebase  github Report 📑 Poster 📈 Demo ▶️

Introduction

In today’s digital era, the rapid spread of misinformation and disinformation poses a significant societal challenge. Enabled by the rise of advanced technologies such as large language models and artificial intelligence tools, these phenomena undermine mutual trust and can have serious consequences on democratic processes and public safety. Individuals and entities can now easily create and disseminate unchecked information, reaching vast audiences at an alarming rate. This ease of spreading falsehoods not only threatens social harmony but also necessitates an urgent call for effective detection, evaluation, and mitigation strategies. This project aims to explore the growing impact of digital misinformation and disinformation, highlighting how emerging technologies facilitate their spread. It will also propose new solutions to enhance the resilience of information ecosystems against the onslaught of digital falsehoods.

Why is our project unique?

Our veracity engine uses a suite of latest tools and techniques to power the analysis. This includes:

Roadmap

Figure 1. Roadmap of our project

Datasets

Liar PLUS: Integration with Predictive AI

Politifact & Snopes: Integration with Generative AI

Methodology

Predictive AI

Combines traditional predictive AI models for statistical rigor and generative AI for nuanced content analysis. Anchors analysis with structured factuality scoring.

Generative AI

Comparison to Traditional Prompting:
Traditional: One-off evaluations, limited depth.
FCOT: Recursive, multi-factor, transparent reasoning.
Use factuality factors as objective functions.
Update score at each iteration with the usage of function calling.

By embedding these real-time search results into the prompt, the GenAI gains access to a broader and more dynamic set of data, enabling it to cross-reference claims made in the inputted news article with credible external sources.

Function calls are strategically used to dynamically adjust analysis parameters based on real-time feedback. This adaptability is essential for calculating the effectiveness of various thought patterns generated by our algorithm, ensuring that the most logical and factually consistent chains are prioritized.

Results

Prediction / Generative / Hybrid

Predictive Performance on Liar PLUS Dataset:

Model Description Score (%)
BERT Embedding Model 43.7
XGBoost/LightGBM (Boosting algorithm) 33.1
Random Forest Classifier (Bagging algorithm) 67.8
Sentiment Analysis (TF-IDF) 45.9
Word2Vec 55.2

Table 1. Predictive Performance on Liar PLUS dataset

Overall Model Performance:

Model Description Score (%)
Baseline (Feeding straight into Gemini Flash 2.0) 19
Hybrid (Random Forest + Gemini) 34.3
Hybrid + RAG 40
Hybrid + RAG + Web Search 56.9
Hybrid + RAG + Web Search + FCOT Prompting 67.2
Hybrid (50/50) + RAG + Web Search + FCOT Prompting + Function Calling 65.3
Hybrid (70/30) + RAG + Web Search + FCOT Prompting + Function Calling 85.1

Table 2. Overall Model Performance

Precision & Recall Result:

Table 3. predictive vs generative vs hybrid accuracy

Prompting Comparison Result: Prompting Comparison Link

Prompts Constructed

Normal Prompting

Use 3 iterations to check the veracity score of this news article. Factors to consider are Frequency Heuristic and Misleading Intentions. In each, determine what you missed in the previous iteration. Also put the result from RAG into consideration/rerank. RAG: Here, out of six potential labels (true, mostly-true, half-true, barely-true, false, pants-fire), this is the truthfulness label predicted using a classifier model: {predict_score}. These are the top 100 related statement in LiarPLUS dataset that related to this news article: {get_top_100_statements(input_text)}. Provide a percentage score and explanation for each iteration and its microfactors. Final Evaluation: Return an exact numeric veracity score for the text, and provide a matching label out of these six [true, mostly-true, half-true, barely-true, false, pants-fire]

Chain of Thought: Chain of Thought Full Prompt Link

Fractal Chain of Thought: Fractal Chain of Thought Full Prompt Link

Further Discussion

Ethical Consideration