Collective Inquiry • 2025

hybrid stories

A research study analyzing the collaborative writing process between humans and generative AI tools in short story creation.

How does co-working with AI affect the fiction-writing process, and the fiction-rating process?

Transformation sequence showing degrees of human-AI collaboration
Like mermaids—part human, part something else—these stories blend human creativity with AI assistance.

Participate in Research

Submit a Story

Upload a story you wrote with any degree of LLM or multi-modal AI assistance

Submit a Story →

Tag Story Elements

Help identify which parts were written by humans, AI, or through collaboration

Start Tagging →

Score the StoriesComing Soon

Rate stories on creativity, structure, and artistic achievement

Start Scoring →

Research Methodology

The study takes part in three phrases:

1

Story Submission & Documentation

Writers submit completed short flash fiction stories, alongside complete chat logs or interaction histories with AI tools.

2

Mixed Authorship Attribution Tagging

Volunteer data taggers examine each story and the provided chat links to identify and categorize text snippets as purely human-authored, AI-generated, or collaboratively produced.

Volunteers also categorize the key elements of each story - characters, plot elements, and literary devices.

In parallel, an LLM extracts the key elements of the story.

3

Literary Merit Evaluation

Volunteer readers evaluate stories on their literary merit and enjoyability, including narrative structure, character development, and plot points.

Then, the stories also undergo a mixed human-AI rating process, using techniques from Google DeepMind Rater Assist. Finally, an AI will generate an independent rating of the story's literary merit and technical execution.

Research Objectives

We're analyzing how writers integrate AI assistance into their creative workflow and decision-making processes1. By comparing stories with different collaboration patterns, from minimal AI assistance to extensive co-creation, we aim to understand what new workflows are emerging.2.

We also aim to understand how both human- and AI-perceived quality vary with the degree and type of AI assistance or collaboration. What patterns can we observe?

When given free rein to use AI assistance, what parts of a story do writers use assistance on? What is the difference between what an AI tags as important in a story, versus what humans do? How do human ratings of a story change when they are prompted by an AI on what qualities to rate on?

This research contributes to understanding how AI tools affect creative quality, authorship attribution3, and the future of collaborative writing. Results will inform both academic research and practical guidelines for writers working with AI assistance4.

References

1. Bridgers, S., Jain, R., Greig, R., Shah, R. (2024) "Human-AI Complementarity: A Goal for Amplified Oversight." DeepMind Safety Research. doi:10.1016/j.ccr.2023.03.012
2. Nguyen, A., Hong, Y., Dang, B., & Huang, X.(2023). "Human-AI collaboration patterns in AI-assisted academic writing" Studies in Higher Education., Vol. 49, 2024. doi:10.1080/03075079.2024.2323593
3. Mirowski, P., Mathewson, K. W., Pittman, J., & Evans, R. (2022). "Co-Writing with Opinionated Language Models Affects Users' Views." Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems, 1-15. doi:10.1145/3491102.3517582
4. Yuan, A., Coenen, A., Reif, E., & Ippolito, D. (2022). "Wordcraft: Story Writing With Large Language Models." 27th International Conference on Intelligent User Interfaces, 841-852. doi:10.1145/3490099.3511105
5. Clark, E., Ross, A. S., Tan, C., Yangfeng, J., & Smith, N. (2023). "Creative Writing with a Machine in the Loop: Case Studies on Slogans and Stories." Journal of Digital Humanities, 8(1), 45-67. doi:10.1145/3172944.3172983