While the purpose of the research was mainly “to assess the capability of LLMs to generate personas with limited cognitive and language skills” (which LLMs are indeed capable of achieving), the more interesting findings were that:

    • Every test of a model’s ability to perform a task was, in reality, a test of the examiner’s skill in defining a persona suitable for the task; their proficiency in locating this persona within the model’s latent space; and the model’s latent capacity to simulate the persona with sufficient fidelity to accomplish the task.
    • Findings show that the language models are capable of “downplaying their abilities to achieve a faithful simulation of prompted personas”.
    • Even if an LLM (by current standards) encompasses a more comprehensive world-model than any human, prompting it to simulate a human or human-like expert would not (for now) result in super-human behavior, “since the human imperfections would be simulated as well”.
    • Even if an LLM (by current standards) encompasses a more comprehensive world-model than any human, prompting it to simulate a human or human-like expert would not (for now) result in super-human behavior, “since the human imperfections would be simulated as well”.