Use Language Models in Your Rust Application (Free, Open-Weight, Self-Hosted)

preview_player
Показать описание
A concise look at the Kalosm crate, including how it fits in with the rest of the libraries of the LLM ecosystem.

Рекомендации по теме
Комментарии
Автор

Wow, there really are so many applications! Amazing video❤

virusblitz
Автор

Awesome content! Excited to try this crate. Thank you for sharing

northicewind
Автор

if you are an engineer and serious about LLMs in Rust learn Candle. Ive been able to research and develop optimizers and architectures with Candle. Its an involved framework but is feature rich and complete in more places than any other framework. It makes design decisions that are reminiscent of torch and tensorflow and is very low level and not polyglot compared to alternatives.

first-thoughtgiver-of-will
Автор

Technically if you want to deploy an LLM powered application you would not just rely on a library like Transformers. You also need a performant server that can A. Do adaptive batching and B. Implement efficient “multiplications” to work optimally on multiple batches. So while kalosm looks dope, you would not want to use Rust if you actually need a python library like vLLM or NVIDIA’s triton (it’s oss). Folks at Kyutai (the company founded by the same guy who contributes to Candle) said they use rust for inference. So they must have implemented all of that in rust :)

But if you don’t need that much load, of course use kalosm. I’m pretty sure you can find a crate that implements batching for tokio.

__sassan__
Автор

I think the sweet spot for using a library like Kalosm would be integrating it into a Tauri app and downloading + executing the ML model on the user's local machine. Pair Kalosm & Tauri with a Leptos frontend, and you have a really compelling native desktop (or mobile) app!

An installable PWA using a similar stack (minus Tauri of course) would also be a great use case for Kalosm + Leptos..

DaM_Cdn
join shbcf.ru