branch: externals/minuet commit 8f25ad253843cd996a43cc7c3ec1b8788affc47a Author: Milan Glacier <d...@milanglacier.com> Commit: Milan Glacier <d...@milanglacier.com>
doc: update README. --- README.md | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/README.md b/README.md index 6fe1289050..0e1889048e 100644 --- a/README.md +++ b/README.md @@ -293,9 +293,9 @@ For cloud-based providers, offers a valuable resource for comparing the speed of both closed-source and open-source models hosted by various cloud inference providers. -When assessing model speed via Openrouter, two key metrics are latency (time to -first token) and throughput (tokens per second). Latency is often a more -critical factor than throughput. +When assessing model speed, two key metrics are latency (time to first token) +and throughput (tokens per second). Latency is often a more critical factor +than throughput. Ideally, one would aim for a latency of less than 1 second and a throughput exceeding 100 tokens per second.