An LPU Inference Engine, with LPU standing for Language Processing Unit™, is a new type of end-to-end processing unit system that provides the fastest inference at ~500 tokens/second.
Replies
Best
This seems extremely interesting- I’m curious what you’ve seen to be the biggest use case for this LLM?
Report
It is fast, that is for sure. Where can I get more information about the chips and hardware? Is there a GPU cloud service?
Replies
Tradepost.ai
Locofy.ai
Videotok
Locofy.ai
Avatars by Studio Neiro AI
Crustdata