branch: externals/minuet
commit 8f25ad253843cd996a43cc7c3ec1b8788affc47a
Author: Milan Glacier <d...@milanglacier.com>
Commit: Milan Glacier <d...@milanglacier.com>

    doc: update README.
---
 README.md | 6 +++---
 1 file changed, 3 insertions(+), 3 deletions(-)

diff --git a/README.md b/README.md
index 6fe1289050..0e1889048e 100644
--- a/README.md
+++ b/README.md
@@ -293,9 +293,9 @@ For cloud-based providers,
 offers a valuable resource for comparing the speed of both closed-source and
 open-source models hosted by various cloud inference providers.
 
-When assessing model speed via Openrouter, two key metrics are latency (time to
-first token) and throughput (tokens per second). Latency is often a more
-critical factor than throughput.
+When assessing model speed, two key metrics are latency (time to first token)
+and throughput (tokens per second). Latency is often a more critical factor
+than throughput.
 
 Ideally, one would aim for a latency of less than 1 second and a throughput
 exceeding 100 tokens per second.

Reply via email to