post https://api.monsterapi.ai/v1/deploy/evaluation/llm/lm_eval
Use Eluether Evaluation Harness to evaluate on lm_eval engine. Supported evals are as follow for engine:
- lm_eval: mmlu, gsm8k, hellaswag, arc, truthfulqa, winogrande
Models >8B and context more than 8k are not currently supported. Support will be added shortly.