What's Changed
Initial release of GuideLLM with version 0.1.0! This core release adds the basic structure, infrastructure, and code for benchmarking LLM deployments across several different use cases utilizing a CLI interface and terminal output. Further improvements are coming soon!
- Support added for general OpenAI backends and any text-input-based model served through those
- Support added for emulated, transformers, and file-based datasets
- Support added for general file storage of the full benchmark/evaluation that was run
- Full support for different benchmark types including sweeps, synchronous, throughput, constant, and poison enabled through new scheduler and executor interfaces built on top of Python's asyncio
New Contributors
- @DaltheCow made their first contribution in #4
- @markurtz made their first contribution in #3
- @rgreenberg1 made their first contribution in #21
- @jennyyangyi-magic made their first contribution in #35
Full Changelog: https://github.com/neuralmagic/guidellm/commits/v0.1.0