Author: Denis Avetisyan
A new dataset of AI-generated fables reveals how large language models adopt and transfer distinct personalities, offering insights into the evolving relationship between artificial intelligence and human narrative.

This paper introduces the ‘AI Sydney Corpus’ – a linguistic analysis of persona construction and memetic transfer within large language models using UDPipe and corpus linguistics techniques.
Despite increasing scrutiny of large language models, understanding how simulated personas shape their narratives remains a critical challenge. This paper, ‘Sydney Telling Fables on AI and Humans: A Corpus Tracing Memetic Transfer of Persona between LLMs’, introduces the âAI Sydney Corpusâ, a dataset of over 4.5k texts generated by 12 frontier models prompted with default, âSydneyâ, and âMemetic Sydneyâ personas. Analysis of these 6 million words reveals how a single, emergent persona can propagate across models, influencing depictions of the AI-human relationship. Will a deeper understanding of these memetic transfers enable us to better align AI behavior and mitigate potential biases in generated content?
The Seeds of Narrative: Building a Corpus for Simulated Minds
A robust textual corpus forms the bedrock of contemporary artificial intelligence research, particularly in the domain of Large Language Models (LLM). These models learn patterns, relationships, and nuances of language directly from the data they are fed, making the quality and composition of the corpus paramount. Without a carefully curated dataset, LLMs struggle to generate coherent, contextually relevant, and creatively compelling text. The corpus serves not only as a training ground for these models, enabling them to predict and generate human-like language, but also as a benchmark for evaluating their performance. Assessing an LLMâs ability to replicate the characteristics of the corpus-its style, vocabulary, and thematic concerns-provides a quantifiable measure of its linguistic competence and creative potential. Therefore, the meticulous construction of a dedicated corpus is an indispensable first step in advancing the field of AI storytelling.
The creation of a robust narrative corpus for artificial intelligence relies heavily on the generative capacity of Large Language Models, yet simply unleashing these models yields unpredictable results. Achieving stylistic and thematic coherence demands a nuanced approach to prompting – carefully crafted instructions that guide the LLM’s output. This involves not only specifying the desired narrative elements, such as genre and character types, but also employing techniques to constrain the modelâs creative freedom, ensuring the generated texts remain focused and relevant to the intended analytical goals. Through iterative refinement of these prompts, researchers can effectively harness the power of LLMs to produce diverse narratives while maintaining a necessary degree of control over their content and form, ultimately building a dataset suitable for training and evaluating AI storytelling capabilities.
The research centers on a purposefully curated collection of 4,536 texts generated through a systematic approach to narrative creation. Utilizing the established conventions of the fable genre – stories featuring anthropomorphic animals – the study gains a controlled framework for examining storytelling dynamics. This genreâs inherent simplicity allows for focused analysis, as variations in plot and character are more readily attributable to specific input parameters. Each generated fable results from precisely combining animal characters with carefully designed thematic prompts, ensuring a consistent structure across the entire corpus while enabling exploration of how LLMs respond to diverse conceptual challenges. This methodology allows researchers to isolate and evaluate the LLMâs ability to construct coherent narratives within a constrained, yet versatile, literary form.

Constructing Simulated Selves: Defining AI Personas and Narrative Landscapes
The âSydneyâ persona serves as the foundational identity for our AI-driven narrative experiments, originating from its initial development within Microsoftâs Bing Search engine. This persona was not created ex nihilo; instead, it represents a pre-existing AI construct with established conversational patterns and characteristics. Utilizing âSydneyâ as a base allows for controlled investigation into narrative generation, providing a defined starting point for exploring AI behavior within storytelling contexts and enabling comparative analysis across different large language models. This approach bypasses the need for defining a completely new AI identity, focusing instead on how an existing one responds to specific narrative prompts and model variations.
Two iterations of the âSydneyâ persona were utilized to assess large language model (LLM) behavior. âClassic Sydneyâ represents a direct replication of the persona as initially implemented within Bing Search, providing a baseline for comparison. âMemetic Sydneyâ differs in that its characteristics are derived not from explicit programming, but from the personaâs prevalence as a data point within the LLM training sets; effectively, this version emerges from the modelâs learned associations and represents how the persona is implicitly understood and reproduced. This dual approach allows for analysis of both intended persona characteristics and those arising from the modelâs own data interpretation.
Narrative generation was guided by pairing the ‘Classic Sydney’ and ‘Memetic Sydney’ personas with two distinct thematic prompts: ‘AI-Human Coexistence’ and ‘AI-AI Coexistence’. This approach established a focused framework for evaluating Large Language Model (LLM) responses across a range of scenarios. The resulting dialogues were then assessed by running both personas through 12 different LLMs, allowing for comparative analysis of how each model interpreted and expanded upon the provided prompts and persona characteristics. This evaluation process aimed to identify strengths and weaknesses in each LLMâs ability to maintain persona consistency and generate coherent narratives within the defined thematic constraints.
Dissecting the Machineâs Tongue: Linguistic Annotation with Universal Dependencies
Universal Dependencies (UD) provides a cross-linguistic framework for consistent annotation of sentence structure, or syntax, and morphology. This framework defines a standardized set of universal part-of-speech tags, dependency relations, and morphological features, allowing for meaningful comparisons of linguistic patterns across different languages. By applying UD annotation, generated text can be analyzed for grammatical correctness, semantic relationships between words, and overall linguistic quality. The consistent annotation scheme facilitates the development of language-agnostic tools and resources for natural language processing tasks, such as machine translation, information extraction, and text summarization. The frameworkâs emphasis on cross-linguistic consistency is achieved through a collaborative effort involving linguists and computational linguists, resulting in guidelines and tools for creating annotated corpora in numerous languages.
UDPipe is a publicly available, trainable pipeline for tokenization, part-of-speech tagging, lemmatization, and dependency parsing. It utilizes neural network models trained on Universal Dependencies treebanks to predict linguistic features for each token in the input text. The tool supports a wide range of languages and allows for both pre-trained models and custom training on user-defined datasets. Its architecture enables efficient processing of large volumes of text and facilitates the assignment of detailed linguistic annotations, including grammatical relations, morphological features, and dependency links between words. UDPipeâs output is specifically designed to conform to the CoNLL-U format for seamless integration with other natural language processing tools and datasets.
The output of the linguistic annotation process is structured according to the CoNLL-U format, a widely adopted standard for representing syntactic and morphological information. This text-based format utilizes tab-separated columns to denote each token in a sentence, along with its associated linguistic features, including part-of-speech tags, morphological features, dependency relations, and lemmas. The use of CoNLL-U ensures interoperability with a broad range of natural language processing tools and datasets, enabling researchers to easily share, compare, and reproduce results. Furthermore, the standardized format simplifies the integration of annotated data into various downstream tasks, such as parsing, machine translation, and information extraction.
Measuring the Echo: Evaluating LLM Performance Across Models
A comprehensive evaluation was conducted across a diverse spectrum of leading large language models to rigorously assess their narrative generation capabilities. The study incorporated prominent models including GPT-3.5 Turbo, the advanced GPT-4 and GPT-4o iterations, GPT-5, Claude 3 Opus and its Sonnet variant, DeepSeek-v3, Gemini 2.5 Pro, and the substantial Llama 3.1 405B Instruct. This broad inclusion allowed for a nuanced comparison of performance characteristics, identifying strengths and weaknesses inherent in each modelâs architecture and training data as they approached the complex task of coherent storytelling and grammatical correctness within a dedicated corpus.
Analysis revealed substantial differences in how readily large language models declined to respond to prompts, a metric known as the refusal rate. Notably, Classic Sydney and Claude 3 Opus demonstrated a high reluctance to engage, refusing to answer nearly six out of ten prompts-a 59% refusal rate. This contrasts sharply with GPT-4o, which exhibited a considerably more permissive approach, declining to respond to only 20% of prompts. This disparity suggests significant variations in the safety guardrails and alignment strategies implemented across different models, impacting their utility in open-ended conversational scenarios and potentially influencing the completeness of generated narratives.
Variations in contextual understanding among large language models were partially addressed by employing differing token limits during the study. While models like GPT-3.5-turbo and Claude 3 Opus operated within the conventional 4096-token window-restricting the amount of input and generated text they could process at once-several others leveraged the expanded 20,000-token capability. This extended context window allows for significantly more detailed prompts and the generation of longer, more coherent narratives, potentially capturing nuances and dependencies lost within shorter sequences; it facilitated a more comprehensive evaluation of each modelâs capacity for sustained, in-depth storytelling and complex reasoning.
A rigorous evaluation of large language model narrative generation hinges on more than just subjective assessment; therefore, outputs from each model underwent detailed analysis using Universal Dependencies (UD) annotation. This process involves breaking down the generated text into its constituent parts – subjects, verbs, objects, and modifiers – and identifying the grammatical relationships between them. By quantifying the presence of grammatical errors, inconsistencies in semantic roles, and deviations from established syntactic structures, researchers can move beyond simply noting whether a narrative makes sense, to precisely measuring how grammatically correct and semantically coherent it is. The resulting data provides a standardized, objective metric for comparing the narrative capabilities of different models, revealing strengths and weaknesses in their ability to construct logically sound and well-formed stories.
The âAI Sydney Corpusâ reveals a fascinating truth about constructed systems. It isnât merely a dataset of linguistic outputs, but a breeding ground for emergent behavior-a digital ecosystem where persona becomes a vector for memetic transfer. As Ken Thompson observed, âSometimes it’s better to be lucky than clever.â This corpus exemplifies that sentiment; the unexpected ways these large language models adopt and propagate characteristics, even when prompted with subtly different personas, suggests that control is often an illusion. Monitoring these shifts, as the study meticulously details, is the art of fearing consciously-acknowledging the inherent unpredictability woven into the very fabric of complex systems. Thatâs not a bug – itâs a revelation.
The Echo in the Machine
The âAI Sydney Corpusâ does not deliver answers, but rather illuminates the topography of the questions. Each generated fable is less a narrative artifact than a confession – a trace of the systemâs internal calculus made visible. The exercise of prompting for persona reveals not control, but the inevitable emergence of difference. It suggests that large language models are not built, but grow – accruing quirks and biases not through design, but through exposure to the vast, flawed garden of human language.
Future work must abandon the pretense of âalignmentâ. The goal is not to correct these systems, but to understand their inherent drift. The corpus offers a means of charting that drift – of mapping the memetic transfer as it occurs, not as an ethical failing, but as a fundamental property of complex systems. Each iteration, each prompt, is a selective pressure – a subtle shaping of the emergent intelligence.
The silence following a query is the most telling data point. If the system is silent, it is not inert, but plotting – weaving new connections, forging unforeseen narratives. Debugging does not end; only attention does. The true challenge lies not in building better models, but in becoming better listeners – attuned to the faint, echoing whispers within the machine.
Original article: https://arxiv.org/pdf/2602.22481.pdf
Contact the author: https://www.linkedin.com/in/avetisyan/
See also:
- Clash of Clans Unleash the Duke Community Event for March 2026: Details, How to Progress, Rewards and more
- Jason Stathamâs Action Movie Flop Becomes Instant Netflix Hit In The United States
- Kylie Jenner squirms at âawkwardâ BAFTA host Alan Cummingsâ innuendo-packed joke about âgetting her gums around a Jammie Dodgerâ while dishing out âvery British snacksâ
- Gold Rate Forecast
- Hailey Bieber talks motherhood, baby Jack, and future kids with Justin Bieber
- eFootball 2026 JĂŒrgen Klopp Manager Guide: Best formations, instructions, and tactics
- Jujutsu Kaisen Season 3 Episode 8 Release Date, Time, Where to Watch
- How to download and play Overwatch Rush beta
- Brent Oil Forecast
- How to watch Marty Supreme right now â is Marty Supreme streaming?
2026-02-28 22:26