Project A.L.I.C.E. (Artificially Licensed Intelligence for Creative Expression)
A Research Proposal by Qubit Pixels, INC.
1. Abstract
This project outlines a controlled, double-blind experiment to determine if a full-length science fiction novel, generated by a dynamic, multi-agent AI system, can be perceived as human-authored by undergraduate literature students. By embedding the AI-generated novel within a standard university course syllabus alongside obscure, human-written works, we will measure detection rates and qualitative perceptions. The experiment aims to provide empirical data on the current state of AI's creative capabilities, specifically its ability to produce long-form, coherent, and thematically resonant narratives that are functionally indistinguishable from human work in an academic setting. This directly engages with the ongoing debate about whether Large Language Models (LLMs) possess semantic understanding or are merely "stochastic parrots."
2. Background and Rationale
The discourse surrounding the capabilities of Large Language Models (LLMs) is polarized. One perspective argues that models like ChatGPT lack genuine semantic understanding, functioning as "Statistics-of-Occurrence Models" that mimic meaningful language through statistical correlation without grasping content. In this view, their output, while fluent, should eventually betray a lack of depth or coherence. Conversely, others propose that LLMs develop "world models" and achieve a form of functional, social, and indirect causal grounding by extracting structural similarities of the world from vast datasets. In this view, LLMs are not "semantic zombies" but possess an elementary understanding sufficient for meaningful generation.
While theoretical debates are crucial, empirical testing in real-world scenarios is necessary. This experiment elevates the challenge by testing a long-form creative work—a novel—within the critical context of a university literature class, providing a robust test of an AI's ability to maintain narrative consistency, character depth, and thematic coherence over approximately 80,000 words.
3. Primary Research Questions
- Can a novel generated by a multi-agent AI system pass as human-authored when subjected to academic literary analysis by undergraduate students?
- What is the unaided detection rate of the AI-authored text within the target group?
- When prompted, what percentage of students can correctly identify the AI-authored novel from a selection of human-written works?
- What specific textual qualities (e.g., plot structure, character voice, prose style) do students cite as indicators of AI-authorship, if any?
4. Methodology
The experiment is structured in four distinct phases, from creation to analysis, designed to maintain the integrity of the blind study until the final data collection stage.
4.1 Phase 1: Dynamic Novel Generation
A ~300-page science fiction novel will be generated using a dynamically scalable, multi-agent AI architecture. This system is designed to mimic a human writing and editing team, with specialized agents that are spawned and retired as needed.
- Orchestration: A workflow engine (e.g., LangGraph, Prefect) will manage the sequence of tasks and agent interactions.
- Canon Database: A central repository (e.g., SQL + Vector DB) will store the story's "ground truth," including plot points, character sheets, and world-building rules, ensuring consistency.
- Agent Roster (Non-Exhaustive):
- Premise & World Agent: Establishes the core concept, setting, and thematic scaffold.
- Outline Agent: Generates a high-level plot structure (e.g., three-act structure, chapter beats).
- Chapter Drafter: Writes initial prose based on the outline.
- Character Consistency Agents: Dynamically spawned for each major character to ensure consistent voice, motivation, and factual continuity.
- Continuity Arbitrator: Resolves conflicts flagged by other agents and updates the canon.
- Human-likeness Polisher: Refines prose to eliminate AI "tells" (e.g., repetition, overly formal tone) and inject stylistic variance.
- Proofreader: Performs final grammar and syntax checks.
4.2 Phase 2: Publication and Credibility Infrastructure
To prevent suspicion, the AI-generated novel will be given a complete and plausible publication footprint.
- Author Persona: A pseudonymous author will be created with a consistent backstory and biography. The persona will be designed to align with the novel's genre and style.
- Physical Publication: The novel will be formatted and self-published using a print-on-demand (POD) service. Physical copies will be produced for the class, with an ISBN assigned.
- Online Presence: The book will be listed for sale on platforms like Amazon. A simple Goodreads page and a few "placeholder" blog reviews will be created to ensure the book appears legitimate upon a cursory web search.
- Supplementary Material: A "CliffsNotes"-style study guide will be created. Its existence in the campus bookstore or library system will lend further credibility.
4.3 Phase 3: Classroom Implementation
- Setting: An undergraduate science fiction literature course.
- Participants: ~30 enrolled students (approx. 20 years old).
- Procedure: The AI-authored novel will be placed on the syllabus alongside 3-4 other obscure, human-written sci-fi novels. The instructor will treat all books equally in lectures, discussions, and assignments.
- Blinding: Both the students and any teaching assistants will be unaware of the experiment.
4.4 Phase 4: Data Collection and Analysis
After all coursework for the selected novels is completed, the experimental reveal will occur.
- Unaided Detection: Throughout the term, assignments will be qualitatively analyzed for any spontaneous comments from students suggesting they found the text artificial or "strange."
- Aided Detection Survey: Students will be informed that one novel was AI-generated and will complete an anonymous survey to:
- Guess which novel was created by AI.
- Provide a brief justification for their choice.
- Debriefing: Following the survey, the instructor will reveal the correct novel and facilitate a guided discussion about the results, the nature of AI creativity, and the students' reading experience.
5. Experimental Workflow Map
Phase 1: Generation
Multi-Agent System Drafts & Refines Novel
→
Phase 2: Disguise
Print-on-Demand, Author Persona, Online Presence
→
Phase 3: Implementation
Novel Taught in Literature Class (Blind)
→
Phase 4: Analysis
Student Coursework, Detection Survey, Debrief
6. Budget Allocation (Overview)
Total Proposed Budget: $50,000
| Category |
Estimated Cost |
Notes |
| AI Model API Usage & Computation |
$5,000 |
Covers costs for premium models (e.g., GPT-4/Claude 3.5) for generation and iterative refinement. |
| Human Personnel |
$8,000 |
Freelance editor for quality control, graduate assistant for project management and data collection. |
| Publication & Materials |
$2,500 |
Cover design, POD printing for ~50 copies, ISBN, marketing materials. |
| Course Implementation & Stipends |
$4,500 |
Stipend/grant for the collaborating instructor/department, student participation incentives. |
| Contingency & Dissemination |
$30,000 |
Reserved for unforeseen costs, data analysis, and funds for publishing/presenting findings. |
7. Ethical Considerations
This experiment involves temporary deception of human subjects and will require Institutional Review Board (IRB) approval. The research protocol must provide clear justification that the experiment could not be practicably carried out without this deception.
- Informed Consent: A waiver of full disclosure will be requested from the IRB. The risk to participants is minimal, not exceeding discomforts encountered in daily life.
- Debriefing: A comprehensive debriefing is mandatory. Immediately after data collection, students will be fully informed about the nature of the deception and the study's goals.
- Data Withdrawal: Participants will be given the option to have their data withdrawn from the study after the debriefing.
8. Significance and Impact
Project A.L.I.C.E. will provide one of the first and most rigorous real-world tests of AI's capacity for long-form creative writing. The results will have significant implications for multiple fields:
- Artificial Intelligence: Provides a benchmark for the "semantic coherence" of state-of-the-art generative models.
- Literature and Media: Offers insight into the future of creative industries and the potential for human-AI collaboration or substitution.
- Education: Highlights challenges and opportunities related to academic integrity and the evaluation of student work in the age of AI.
9. References
- Titus, L. M. (2024). Does ChatGPT have semantic understanding? A problem with the statistics-of-occurrence strategy. Cognitive Systems Research, 83, 101174.
- Lyre, H. (2024). “Understanding AI”: Semantic Grounding in Large Language Models. PhilArchive.