I just built a quick demo showing the Llama 3.2B chatbot running on our Metis® platform, totally offline. This model packs 3 billion parameters and runs smoothly on both a standard Lenovo P360 with our PCIe card and even on an Arduino-based dev board (Portenta X8).We hit 6+ tokens/sec with a single core – which means real-time chat. Perfect for smart customer support bot, digital concierge systems, any edge AI assistant application really, all running fully on-device. No cloud needed.Check out the video and let me know what you think. Any projects you can think of where you could use a self-contained, power-efficient, offline AI chatbot like this?