Show HN: We made our own inference engine for Apple Silicon

We wrote our inference engine on Rust, it is faster than llama cpp in all of the use cases. Your feedback is very welcomed. Written from scratch with idea that you can add support of any kernel and platform.


Comments URL: https://news.ycombinator.com/item?id=44570048

Points: 72

# Comments: 23

https://github.com/trymirai/uzu

Creato 7h | 15 lug 2025, 16:50:31


Accedi per aggiungere un commento