Skip to content

Getting started

Rickard Edén edited this page Oct 3, 2023 · 5 revisions

By default it uses KoboldCpp, but if you're feeling adventurous you can change llm_config.yaml in the tale/ folder to another one.

  • Get and install KoboldCpp: https://github.com/LostRuins/koboldcpp/releases (There is also support for OpenAI api) Download a suitable llm model for koboldcpp. I recommend llama-2-7b-chat-codeCherryPop.ggmlv3.q5_K_M(https://huggingface.co/TokenBender/llama2-7b-chat-hf-codeCherryPop-qLoRA-merged), a 7b model which is both fast and very accurate (relatively speaking) when generating JSON.

  • Download repo, either with git clone git@github.com:neph1/LlamaTale.git or as a zip. Master branch should be stable.

  • Run pip install -r requirements.txt

  • Start your backend, KoboldCpp or OpenAI compatible. (port 5001 by default, or change in llm_config.yaml)

  • Start the demo with python -m stories.prancingllama.story

Optional:

  • If you'd rather play in a browser, add the --web flag and connect to http://localhost:8180/tale/story
  • There's also a GUI version, which is launched with --gui
  • If you have a v2 character card and want to skip character creation, add --character path_to_character
  • If you want to load a v2 character as a follower, type load_character path_to_character_relative_to_story_folder in the game prompt
  • Check https://github.com/neph1/LlamaTale/wiki/Creating-a-character if you want to make a 'full featured' LlamaTale character.