2025-09-04 23:26:11 -06:00
2025-09-04 23:00:01 -06:00
2025-09-04 03:53:27 -06:00
2025-09-04 23:26:11 -06:00
2025-09-04 23:00:01 -06:00
2025-09-04 23:00:01 -06:00
2025-09-04 23:02:52 -06:00
2025-09-04 23:00:01 -06:00
2025-09-04 03:53:27 -06:00

LSFBench

Minimal Luau/Lune benchmark to evaluate LLMs: one model answers questions, another model scores the answers against the reference key.

Quick Start

Prereqs

  • Install Lune (0.10.x)
  • Start Ollama at http://localhost:11434 and pull the models referenced in config.luau (e.g. qwen3:4b)

Notice

The evaluator model must support structured JSON outputs.

Description
luausoftworksbench - a LLM benchmark for testing Luau knowledge and comprehension, written in Luau
Readme MIT 70 KiB
Languages
Luau 100%