The Gemma 4 Vision-Language-Action (VLA) model now runs on the Jetson Orin Nano Super. This deployment enables real-time robotic control and spatial reasoning on low-power edge hardware. Developers can now execute complex multimodal instructions without relying on cloud inference. It proves that high-capability VLAs can fit within tight memory constraints.