Article URL: https://openai.com/index/introducing-codex/
Comments URL: https://news.ycombinator.com/item?id=44006345
Points: 169
# Comments: 93
Utworzony
6h
|
16 maj 2025, 17:20:16
Zaloguj się, aby dodać komentarz
Inne posty w tej grupie

I discovered that in LLM inference, keys and values in the KV cache have very different quantization sensitivities. Keys need higher precision than values to maintain quality.
I patched llama.cp
Article URL: https://clojurescript.org/news/2025-05-16-release




Article URL: https://bobacollection.staxmuseum.org/
Comments URL: https://news.y