This repository contains scripts for running the experiments illustrated in the paper:
"Under the Surface: Tracking the Artifactuality of LLM-Generated Data"
Debarati Das†¶, Karin de Langis¶, Anna Martin-Boyle¶, Jaehyung Kim¶, Minhwa Lee¶, Zae Myung Kim¶, Shirley Anugrah Hayati, Risako Owan, Bin Hu, Ritik Sachin Parkar, Ryan Koo, Jong Inn Park, Aahan Tyagi, Libby Ferland, Sanjali Roy, Vincent Liu, Dongyeop Kang
Minnesota NLP, University of Minnesota Twin Cities
† Project Lead, ¶ Core Contribution
Our website can be accessed at this Link.
The paper can be accessed at arXiv.
The datasets used in the paper can be downloaded from HuggingFace Hub.
This research project collects diverse text data from large language models (LLMs), encompassing both structured "task labels" and open-ended "free-form text." This extensive dataset allows for a holistic examination of LLM outputs, offering insights into their performance under varying degrees of structure and freedom. The research underscores the importance of responsible and ethical practices in LLM-generated data creation and usage, advocating for collaborative efforts to address biases, enhance diversity, and deepen the understanding of complex human opinions in LLM outputs for ethical and sustainable development.
The structure of the repository closely follows the stress testing methods applied to the five different data types: Task Labels, Preferences, Instructions, Simulations, and Free-Form Text
More specifically, the stress testing experiments are categorized as either "first-order" or "second-order" experiments. In short, the first-order experiments investigate the data "as-is," for example, focusing on their distributional differences and correlation patterns among human- and LLM-generated data; validating and analyzing using manual inspection; and counting how often labels flip between the original human and the resulting machine text. The second-order experiments involve fine-tuning LLMs on the machine-generated data and investigating whether the existing artifacts or biases are amplified.
![image](https://private-user-images.githubusercontent.com/3746478/299826993-ae37ef70-78fe-4142-8cc5-8eb02a2c8efd.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3Mzg4MzU1MzIsIm5iZiI6MTczODgzNTIzMiwicGF0aCI6Ii8zNzQ2NDc4LzI5OTgyNjk5My1hZTM3ZWY3MC03OGZlLTQxNDItOGNjNS04ZWIwMmEyYzhlZmQucG5nP1gtQW16LUFsZ29yaXRobT1BV1M0LUhNQUMtU0hBMjU2JlgtQW16LUNyZWRlbnRpYWw9QUtJQVZDT0RZTFNBNTNQUUs0WkElMkYyMDI1MDIwNiUyRnVzLWVhc3QtMSUyRnMzJTJGYXdzNF9yZXF1ZXN0JlgtQW16LURhdGU9MjAyNTAyMDZUMDk0NzExWiZYLUFtei1FeHBpcmVzPTMwMCZYLUFtei1TaWduYXR1cmU9MDNmNzVjMjFlZGE0YzUzMDFlNjc0YjE4MGJjZjc3MzMzZTE3ODkxODg0ZjU3MDk2MmI5NzE3NzdlYTE3YWY2YSZYLUFtei1TaWduZWRIZWFkZXJzPWhvc3QifQ.XFQItMrgicXhLGRK-GOAQhJfhM33Pk__xDufdAWaGuE)
The code for the corresponding first-order and second-order experiments are placed under two directories of the same names, respectively, except for Simulation data type where we did not perform a second-order experiment.
Under each data type directory, a corresponding README.md
file is located for further details:
@misc{das2024surface,
title={Under the Surface: Tracking the Artifactuality of LLM-Generated Data},
author={Debarati Das and Karin De Langis and Anna Martin and Jaehyung Kim and Minhwa Lee and Zae Myung Kim and Shirley Hayati and Risako Owan and Bin Hu and Ritik Parkar and Ryan Koo and Jonginn Park and Aahan Tyagi and Libby Ferland and Sanjali Roy and Vincent Liu and Dongyeop Kang},
year={2024},
eprint={2401.14698},
archivePrefix={arXiv},
primaryClass={cs.CL}
}