LENS 22 — LEARNING STAIRCASE Core question: What's the path from "what is this?" to "I can extend this?" THE STAIRCASE HAS SIX LEVELS. Level 1, CURIOUS, takes fifteen minutes and requires nothing. You watch Annie drive toward a kitchen counter at 54 frames per second, guided entirely by a vision-language model, with no map. The command is two tokens: LEFT MEDIUM. That's it. Level 2, TINKERER, takes fifteen minutes to two hours and requires only Python and an API key — no robot. You run the VLM goal-tracking loop against a laptop webcam. You ask "Where is the coffee mug?" every 18 milliseconds. You print LEFT, CENTER, or RIGHT. You see the multi-query pipeline cycle through scene, obstacle, and path queries on alternating frames. Level 3, BUILDER, takes one to three days. You add hardware: a Raspberry Pi 5, an edge GPU like Panda or Jetson, a USB camera, and a sonar sensor. You deploy the NavController. Phase 2a and 2b are fully achievable here. You have not yet touched ROS2. Level 4 is THE PLATEAU — and it has two sibling rungs, not one. RUNG 4A is SLAM deployment. You want SLAM. SLAM needs ROS2. ROS2 needs Docker. Docker needs Zenoh. And Zenoh — the apt package — ships the wrong wire protocol version. You must build rmw_zenoh from source, which needs Rust, which needs a multi-stage Dockerfile. Then the IMU frame_id: one string wrong, six hours of debugging. Then slam_toolbox's lifecycle activation requires a TF gate that is not documented in a single place. Then MessageFilter drops 13 percent of scans under load with no error message. One to four weeks of debugging. Skill-type discontinuity — not harder ML, different domain entirely. RUNG 4B is the rung most practitioners never see. ACTIVATE THE IDLE NPU ON THE ROBOT YOU ALREADY BUILT. The Hailo-8 AI HAT-plus on the Pi 5 — 26 TOPS of neural processing, physically installed on the robot, idle for navigation the entire time the VLM pipeline was under construction. YOLOv8n runs on it at 430 frames per second with zero WiFi dependency. Roughly one to two engineering sessions to learn HailoRT, TAPPAS GStreamer pipelines, and dot-h-e-f compilation from ONNX. Same difficulty tier as SLAM deployment — it is a new ecosystem, not harder machine learning — but with no procurement blocker. The hardware is already in your hand. Level 5, INTEGRATOR, becomes the dual-process composer. Compose Hailo L1 — fast reactive, 30-plus hertz, local, no WiFi — with VLM L2 — slow semantic, 15 to 27 hertz, on Panda. This is exactly the architecture validated by the IROS paper arXiv twenty-six-oh-one dot twenty-one-five-oh-six: 66 percent latency reduction, 67.5 percent task success versus 5.83 percent for VLM-only. Layer SLAM-plus-VLM semantic-map fusion on top. Annie gains a safety floor that survives WiFi drops. Level 6, EXTENDER, is where you do original work. AnyLoc visual loop closure, SigLIP 2 place recognition, voice queries against the semantic map. THE KEY INSIGHT: The plateau is not a difficulty increase. It is a domain transition. You are not a bad ML practitioner. You have entered robotics middleware, which has twenty years of sharp edges accumulated in places no tutorial points to. THE META-LESSON — THE INVISIBLE-RUNG PRINCIPLE. The Learning Staircase has invisible rungs corresponding to dormant hardware you already own. The Hailo-8 on the Pi 5 is idle. The second DGX Spark — the Beast — sits dormant while Titan does the work of both. An Orin NX 16-gigabyte is owned and earmarked for a future robot that has not yet been assembled. Each is a ready-made Level 4 rung hidden by how roadmaps are drawn. Research roadmaps list MODELS and ALGORITHMS, not IDLE SILICON, so a practitioner feels stuck between "VLM working" and "buy a better GPU" and misses the fact that the better rung is already mounted to the chassis. The next step up is not always "buy more compute." It is often "activate what you bought months ago." Audit your hardware inventory every time you feel plateaued. THREE THINGS UNSTICK PEOPLE AT THE PLATEAU. First: a working Docker Compose that someone has already debugged. Second: a sensor validation script that prints four lines — IMU OK, Lidar OK, TF OK, EKF OK. Third: accepting that the transition is real, and checking whether the next rung is already in your hand. Nova's frame: the Phase 2 roadmap reads as a clean linear progression — 90, 85, 65, 55, 50. The twenty-point cliff between phases 2b and 2c is not harder machine learning. It is a skill-type discontinuity into robotics middleware. And the shortest path across it may be activating hardware you already own.