EleutherAI / lm-evaluation-harness

A framework for few-shot evaluation of language models.

Home Page:https://www.eleuther.ai

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Output constrained support

Mihaiii opened this issue · comments

I'd like to build a benchmark and I'd like to set up some rules based on which I can constrain the responses the LLM provides - then eval with pre-determined code (taking into consideration the predetermined constrained rules) the response.

I'm thinking at integrations with libraries like outlines or guidance.