Vision-Language Model in Your Browser
This demo showcases in-browser vision-language inference with LFM2-VL-450M, powered by ONNX Runtime and WebGPU.
Everything runs entirely in your browser with WebGPU acceleration, meaning no data is sent to a server.