Skip to content

LLM Execution Optimization and Benchmark Discussion #232

@Scriptwonder

Description

@Scriptwonder

The major pain point I am experiencing is that LLM does not have a spatial/contextual understanding of certain commands. This happens when I ask it to generate a populated city, a Dust 2 map prototype, or a garden. User prompt is meant to be simple, so I think we need to add Metaprompt and guidance for LLM, especially on Unity execution, to make it smarter.

What I envision right now is a simple GUI input/button to start, so the user can describe their intent for the further prompt / LLM can use ManageScene to gather the current JSON data and try to have an understanding of what to do. It can go more complex than that, but a starting point with LLM knowing the explicit general answer would be good.

Also, hard to find a benchmark for this kind of testing scenario! We should think of possible testing cases and generation cases to test out our tool in the future.

This can be a very long discussion in terms of optimizing LLM execution; maybe Coplay has experience tackling this issue, but I am very open to the discussion.

Metadata

Metadata

Assignees

No one assigned

    Labels

    enhancementNew feature or request

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions