Skip to content

LLM does not recover after context overflow #385

@3inary

Description

@3inary

Describe the bug

When the context limit is exceeded, the system does not recover gracefully.
After a context overflow error, the LLM fails to restart and the Unity Editor becomes unstable.

Steps to reproduce

  1. ChatBot Demo Scene
  2. Model: Llama 3.1 8B
  3. Set Context Size = 512
  4. Num Predict / Batch Size = 512
  5. Play and enter somtihing like "Tell a absurd sifi story" and "continue" until limit is hit
  6. LlamaLib error 400: request (523 tokens) exceeds the available context size (512 tokens), try increasing it
  7. Stopping and starting Play mode does not recover the model
  8. Editor restart required and Editor may crash on exit

LLMUnity version

3.0.1

Operating System

macOs

Metadata

Metadata

Assignees

No one assigned

    Labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions