Skip to content

The official repository for our paper, "How You Prompt Matters! Even Task-Oriented Constraints in Instructions Affect LLM-Generated Text Detection"

License

Notifications You must be signed in to change notification settings

ryuryukke/HowYouPromptMatters

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

11 Commits
 
 
 
 
 
 
 
 

Repository files navigation

How You Prompt Matters! Even Task-Oriented Constraints in Instructions Affect LLM-Generated Text Detection

arXiv LICENSE

This is the official repository for our paper, "How You Prompt Matters! Even Task-Oriented Constraints in Instructions Affect LLM-Generated Text Detection"

📖 Introduction

To combat the misuse of Large Language Models (LLMs), many recent studies have presented LLM-generated-text detectors with highly promising performance. When users instruct LLMs to generate texts, the instruction can include different constraints depending on the user’s need. However, most recent studies do not cover such diverse instruction patterns when creating datasets for LLM detection.

In this paper, we find that even task-oriented constraints --- constraints that would naturally be included in an instruction and are not related to detection-evasion --- cause existing powerful detectors to have a large variance in detection performance. We focus on student essay writing as a realistic domain and manually create task-oriented constraints based on several factors for essay quality.

Our experiments and analysis show that:

  • A task-oriented constraint has a more significant effect on the detection performance than the baseline randomness caused by generating texts multiple times (via sampling) or paraphrasing the instruction.
    • The standard deviation (SD) of current detector performance on texts generated by an instruction with such a constraint reaches up to 14.4 points in F1-score.
  • The constraints, overall, make LLM detection more challenging than without them (up to a 40.3 drop in F1-score).
  • The high instruction-following ability of LLMs fosters the large impact of such constraints on detection performance.

Considering the remarkable speed of recent LLM development, the instruction-following ability of LLMs would be much better, amplifying the effects of the constraints. Therefore, in an era of evolving LLMs, our finding more strongly calls for further development of robust LLM detectors against such distribution shifts caused by a constraint in instruction.

📢 Updates

  • March 2024: Our constrained essay datasets are now available!

📄 Dataset Info

Our dataset is based on 500 pairs of essay problem statements and human(native-student)-written essays that are part of OUTFOX dataset. The native students range from 6th to 12th grade in the U.S.

We instruct three LMs to generate essays: GPT-4(gpt-4-0613), ChatGPT(gpt-3.5-turbo-0613), and GPT-3(davinci-002).

data/llm_essays/Multiple/ includes essays in the Multiple setting: generating texts multiple times (via sampling).

data/llm_essays/Paraphrase/ includes essays in the Paraphrase setting: generating texts via each paraphrased instruction.

data/llm_essays/Constraint/ includes essays in the Constraint setting: generating texts via instruction with each different constraint.

📚 Citation

If you find our data or ideas useful in your research, please consider citing our work as follows:

@misc{koike2024prompt,
      title={How You Prompt Matters! Even Task-Oriented Constraints in Instructions Affect LLM-Generated Text Detection}, 
      author={Ryuto Koike and Masahiro Kaneko and Naoaki Okazaki},
      year={2024},
      eprint={2311.08369},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}

About

The official repository for our paper, "How You Prompt Matters! Even Task-Oriented Constraints in Instructions Affect LLM-Generated Text Detection"

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published