<oembed><type>rich</type><version>1.0</version><title>utxo the webmaster 🧑‍💻 wrote</title><author_name>utxo the webmaster 🧑‍💻 (npub1ut…r50e8)</author_name><author_url>https://yabu.me/npub1utx00neqgqln72j22kej3ux7803c2k986henvvha4thuwfkper4s7r50e8</author_url><provider_name>njump</provider_name><provider_url>https://yabu.me</provider_url><html>For any local AI maxis, here is my current setup and models:&#xA;&#xA;4x 3090s&#xA;&#xA;2x - qwen3.5-35b q4 256k - 60-80 t/s&#xA;2x - gemma4-27b q4 256k - 50-70 t/s&#xA;&#xA;Running on vLLM via docker &#xA;&#xA;Working mint openclaw, Gemma struggling a bit in open webui (reasoning and tool calling still struggle a bit with Gemma)&#xA;&#xA;Quality and speed are actually amazing, very surprising... Just coding is not very good (compared to opus)</html></oembed>