Login
From:
www.together.ai
(Uncensored)
subscribe
Announcing Together Inference Engine – the fastest inference available
https://www.together.ai/blog/together-inference-engine-v1
links
backlinks
Roast topics
Find topics
Find it!
The Together Inference Engine is multiple times faster than any other inference service, with 117 tokens per second on Llama-2-70B-Chat and 171 tokens per second on Llama-2-13B-Chat