Loki is our open-source solution designed to automate the process of verifying factuality. It provides a comprehensive pipeline for dissecting long texts into individual claims, assessing their worthiness for verification, generating queries for evidence search, crawling for evidence, and ultimately verifying the claims. This tool is especially useful for journalists, researchers, and anyone interested in the factuality of information. To stay updated, please subscribe to our newsletter at our website or join us on Discord!
- Decomposer: Breaks down extensive texts into digestible, independent claims, setting the stage for detailed analysis.
- Checkworthy: Assesses each claim's potential significance, filtering out vague or ambiguous statements to focus on those that truly matter. For example, vague claims like "MBZUAI has a vast campus" are considered unworthy because of the ambiguous nature of "vast."
- Query Generator: Transforms check-worthy claims into precise queries, ready to navigate the vast expanse of the internet in search of truth.
- Evidence Crawler: Ventures into the digital realm, retrieving relevant evidence that forms the foundation of informed verification.
- ClaimVerify: Examines the gathered evidence, determining the veracity of each claim to uphold the integrity of information.
git clone https://github.com/Libr-AI/OpenFactVerification.git
cd OpenFactVerification
- Install Poetry by following it installation guideline.
- Install all dependencies by running:
poetry install
-
Create a Python environment at version 3.9 or newer and activate it.
-
Navigate to the project directory and install the required packages:
pip install -r requirements.txt
You can choose to export essential api key to the environment
- Example: Export essential api key to the environment
export SERPER_API_KEY=... # this is required in evidence retrieval if serper being used
export OPENAI_API_KEY=... # this is required in all tasks
export ANTHROPIC_API_KEY=... # this is required only if you want to replace openai with anthropic
export LOCAL_API_KEY=... # this is required only if you want to use local LLM
export LOCAL_API_URL=... # this is required only if you want to use local LLM
Alternatively, you can save the api information in a yaml file with the same key names as the environment variables and pass the path to the yaml file as an argument to the check_response
method.
See demo_data\api_config.yaml
as an example of the api configuration file.
- Example: Pass the path to the api configuration file
python -m factcheck --modal string --input "MBZUAI is the first AI university in the world" --api_config demo_data/api_config.yaml
To test the project, you can run the factcheck.py
script:
# String
python -m factcheck --modal string --input "MBZUAI is the first AI university in the world"
# Text
python -m factcheck --modal text --input demo_data/text.txt
# Speech
python -m factcheck --modal speech --input demo_data/speech.mp3
# Image
python -m factcheck --modal image --input demo_data/image.webp
# Video
python -m factcheck --modal video --input demo_data/video.m4v
The main interface of the Fact-check Pipeline is located in factcheck/core/FactCheck.py
, which contains the check_response
method. This method integrates the complete pipeline, where each functionality is encapsulated in its class as described in the Features section.
Example usage:
from factcheck import FactCheck
factcheck_instance = FactCheck()
# Example text
text = "Your text here"
# Run the fact-check pipeline
results = factcheck_instance.check_response(text)
print(results)
Web app usage:
python webapp.py --api_config demo_data/api_config.yaml
We welcome contributions from the community! If you'd like to contribute, please follow these steps:
- Fork the repository.
- Create a new branch for your feature (
git checkout -b feature/AmazingFeature
). - Commit your changes (
git commit -m 'Add some AmazingFeature'
). - Push to the branch (
git push origin feature/AmazingFeature
). - Open a pull request.
python -m factcheck --modal string --input "MBZUAI is the first AI university in the world" --api_config demo_data/api_config.yaml --model claude-3-opus-20240229 --prompt claude_prompt
python -m factcheck --modal string --input "MBZUAI is the first AI university in the world" --api_config demo_data/test_api_config.yaml --retriever google
python -m factcheck --modal string --input "MBZUAI is the first AI university in the world" --api_config demo_data/test_api_config.yaml --prompt demo_data/sample_prompt.yaml
πͺ Join Our Journey to Innovation with the Supporter Edition
As we continue to evolve and enhance our fact-checking solution, we're excited to invite you to become an integral part of our journey. By registering for our Supporter Edition, you're not just unlocking a suite of advanced features and benefits; you're also fueling the future of trustworthy information.
Your support enables us to:
π Innovate continuously: Develop new, cutting-edge features that keep you ahead in the fight against misinformation.
π‘ Improve and refine: Enhance the user experience, making our app not just powerful, but also a joy to use.
π± Grow our community: Invest in the resources and tools our community needs to thrive and expand.
π And as a token of our gratitude, registering now grants you complimentary token creditsβa little thank you from us to you, for believing in our mission and supporting our growth!
Feature | Open-Source Edition | Supporter Edition |
---|---|---|
Trustworthy Verification Results | β | β |
Diverse Evidence from the Open Web | β | β |
Automated Correction of Misinformation | β | β |
Privacy and Data Security | β | β |
Multimodal Input | β | β |
One-Stop Custom Solution | β | β |
Customizable Verification Data Sources | β | β |
Enhanced User Experience | β | β |
Faster Efficiency and Higher Accuracy | β | β |
Donβt miss out on the latest updates, feature releases, and community insights! We invite you to subscribe to our newsletter and become a part of our growing community.
π Subscribe now at our website!
- API Key Handling: Transitioned from creating key files via copying to dynamically reading all API keys from a YAML file, streamlining configuration processes.
- Unified Configuration Dictionary: Replaced platform-specific dictionaries with a unified dictionary that aligns with environmental variable naming conventions, enhancing consistency and maintainability.
- Model Switching: Introduced a
--model
parameter that allows switching between different models, currently supporting OpenAI and Anthropic. - Modular Architecture: Restructured the codebase into one Base class file and individual class files for each model, enhancing modularity and clarity.
- Base Class Redefinition: Redefined the Base class to abstract asynchronous operations and other functionalities. Users customizing models need only override three functions.
- Prompt Switching: Added a
--prompt
parameter for switching between predefined prompts, initially supporting prompts for OpenAI and Anthropic. - Prompt Definitions via YAML and JSON: Enabled prompt definitions using YAML and JSON, allowing prompts to be automatically read from corresponding YAML or JSON files when the prompt parameter ends with
.yaml
or.json
. - Search Engine Switching: Introduced a
--retriever
parameter to switch between different search engines, currently supporting Serper and Google. - Webapp Frontend Optimization: Optimized the web application frontend to prevent duplicate requests during processing, including disabling the submit button after a click and displaying a timer during processing.
As Loki continues to evolve, our development plan focuses on broadening capabilities and enhancing flexibility to meet the diverse needs of our users. Here are the key areas we are working on:
- Broader Model Compatibility:
- Integration with leading AI models besides ChatGPT and Claude to diversify fact-checking capabilities, including Command R and Gemini.
- Implementation of self-hosted model options for enhanced privacy and control, e.g., FastChat, TGI, and vLLM.
- Unit Testing for Prompts:
- Develop robust unit tests for each step to ensure prompt reliability and accuracy across different scenarios.
- Diverse Search Engines:
- Incorporate a variety of search engines including Bing, scraperapi to broaden search capabilities.
- Integration with Searxng, an open-source metasearch engine.
- Support for specialized indexes like LlamaIndex and Langchain, and the ability to search local documents.
- Dockerization:
- Packaging Loki into Docker containers to simplify deployment and scale-up operations, ensuring Loki can be easily set up and maintained across different environments.
We are committed to these enhancements to make Loki not just more powerful, but also more adaptable to the needs of a global user base. Stay tuned as we roll out these exciting developments!
This project is licensed under the MIT license - see the LICENSE file for details.
- Special thanks to all contributors who have helped in shaping this project.
@misc{Loki,
author = {Wang, Hao and Wang, Yuxia and Wang, Minghan and Geng, Yilin and Zhao, Zhen and Zhai, Zenan and Nakov, Preslav and Baldwin, Timothy and Han, Xudong and Li, Haonan},
title = {Loki: An Open-source Tool for Fact Verification},
year = {2024},
publisher = {GitHub},
journal = {GitHub repository},
howpublished = {\url{https://github.com/Libr-AI/Loki}},
}