Skip to content
This repository has been archived by the owner on Aug 10, 2024. It is now read-only.

Code and data for paper "Large language models can rate news outlet credibility"

License

Notifications You must be signed in to change notification settings

osome-iu/ChatGPT_domain_rating

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

[Note]: We released a new version of the paper with substantial modifications. Please refer to https://github.com/osome-iu/llm_domain_rating for the updated materials.

Introduction

This repo contains the code and data for an older version of our paper with the name Large language models can rate news outlet credibility.

We use the following data in our study:

Data Location Note
Aggregate domain rating list from Lin et al. https://github.com/hauselin/domain-quality-ratings Please download the data in their repo.
MBFC ratings /data/mbfc_ratings.csv We collected the data and share it here.
NewsGuard ratings N/A The data is proprietary, please contact newsguardtech.com to license the data.
ChatGPT ratings /data/chatgpt_ratings.csv.gz We share the responses from ChatGPT here.
Tranco list https://tranco-list.eu Please download the data from their website.

We also share the script we used to query the ChatGPT API at /scripts/query_domain_credibility.py. You will need an OpenAI API key, which can be applied at https://platform.openai.com .

Citation

If you use our data or code in your research, please cite our work as follows:

@article{yang2023large,
  title={Large language models can rate news outlet credibility},
  author={Yang, Kai-Cheng and Menczer, Filippo},
  journal={Preprint arXiv:2304.00228},
  year={2023}
}

About

Code and data for paper "Large language models can rate news outlet credibility"

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages