Haya Barlas
3 posts
We ran Qwen3.6-27B on $800 of consumer GPUs, day one: llama.cpp vs vLLM
Originally published at llmkube.com/blog/qwen3-6-27b-bakeoff. Cross-posted here for the dev.to audience. A Kubernetes-native bake-off on 2× RTX 5060 Ti,…
The magic of messages that have always been with us
If you’re interested in how this magic is implemented in practice as a Python CLI tool, you’ll find…
JustGiving’s playbook for strategic influence in PMM
This article is based on Haya Barlas’s brilliant talk at the London Product Marketing Summit. PMA members can…