I’m interested in automatically generating lengthy, coherent stories of 10,000+ words from a single prompt using an open source local large language model (LLM) on low-spec hardware like a laptop without GPU and with i5-8250U, 16GB DDR4-2400MHz. I came across the “Awesome-Story-Generation” repository which lists relevant papers describing promising methods like “Re3: Generating Longer Stories With Recursive Reprompting and Revision”, announced in this Twitter thread from October 2022 and “DOC: Improving Long Story Coherence With Detailed Outline Control”, announced in this Twitter thread from December 2022. However, these papers used GPT-3, and I was hoping to find similar techniques implemented with open source tools that I could run locally. If anyone has experience or knows of resources that could help me achieve long, coherent story generation with an open source LLM on low-spec hardware, I would greatly appreciate any advice or guidance.

  • Possibly linux@lemmy.zip
    link
    fedilink
    English
    arrow-up
    17
    arrow-down
    1
    ·
    7 months ago

    You aren’t going to get a response that long. That is just the limitations of LLM’s. If you do manage to get something that long it won’t make sense as it can’t hold enough context as it generates.

  • Julian@lemm.ee
    link
    fedilink
    English
    arrow-up
    18
    arrow-down
    8
    ·
    7 months ago

    You can get a really cool, coherent story of any length you want by writing one or hiring a writer.

    • AlligatorBlizzard@sh.itjust.works
      link
      fedilink
      arrow-up
      9
      arrow-down
      1
      ·
      7 months ago

      I’ve won NaNoWriMo twice and I can confirm that writing your own does not necessarily result in a cool or coherent story. One of the two is likely better than an LLM could come up with, though.

  • d416@lemmy.world
    link
    fedilink
    arrow-up
    9
    ·
    7 months ago

    The limited context lengths for local LLMs will be a barrier to write 10k words in a single prompt. Approaches to this is to have the LLM have a conversation with itself or other LLMs. There are prompts out there that can simulate this, but you will need to intervene every few hundred words or so. Check out ‘AutoGen’ frameworks that can orchestrate this for you. CrewAI is one of the better ones. hope this helps

  • Irdial@lemmy.sdf.org
    link
    fedilink
    English
    arrow-up
    5
    ·
    7 months ago

    Ollama provides a Python API which may be useful. You could have it generate the story in chunks, having it generate a list of key points which get passed to subsequent prompts. Maybe…

  • tallricefarmer
    link
    fedilink
    arrow-up
    1
    ·
    7 months ago

    I feel like i’ve seen someone ask this exact question here not too long ago. Was it you?