@inproceedings{4986cad3d6ff42078e5d33908ee99d78,
title = "Can ChatGPT Understand Causal Language in Science Claims?",
abstract = "This study evaluated ChatGPT{\textquoteright}s ability to understand causal language in science papers and news by testing its accuracy in a task of labeling the strength of a claim as causal, conditional causal, correlational, or no relationship. The results show that ChatGPT is still behind the existing fine-tuned BERT models by a large margin. ChatGPT also had difficulty understanding conditional causal claims mitigated by hedges. However, its weakness may be utilized to improve the clarity of human annotation guideline. Chain-of-thought prompting was faithful and helpful for improving prompt performance, but finding the optimal prompt is difficult with inconsistent results and the lack of effective method to establish cause-effect between prompts and outcomes, suggesting caution when generalizing prompt engineering results across tasks or models.",
author = "Yuheun Kim and Lu Guo and Bei Yu and Yingya Li",
note = "Publisher Copyright: {\textcopyright} 2023 Association for Computational Linguistics.; 13th Workshop on Computational Approaches to Subjectivity, Sentiment and Social Media Analysis, WASSA 2023 ; Conference date: 14-07-2023",
year = "2023",
language = "English (US)",
series = "Proceedings of the Annual Meeting of the Association for Computational Linguistics",
publisher = "Association for Computational Linguistics (ACL)",
pages = "379--389",
editor = "Jeremy Barnes and {De Clercq}, Orphee and Roman Klinger",
booktitle = "WASSA 2023 - 13th Workshop on Computational Approaches to Subjectivity, Sentiment and Social Media Analysis, Proceedings of the Workshop",
address = "United States",
}