Session 004: What Morgan Needs
Date: March 2, 2026 Pipeline version: v0.3 Participants: Jeff Kahn, Morgan (Claude, opus)
Context
After the v0.3 breakthrough — hole-density segmentation that finally found the actual net instead of wallpaper — Jeff asked: "What would help you do a better job and get closer to Art's number?"
This document records that conversation and its conclusions.
Current State
Morgan's v0.3 pipeline processes the 226-megapixel Haltadefinizione scan at 1650x6000 working resolution. At this scale:
- 649 net-like holes detected in segmentation
- 584 unique candidates across 4 detection methods
- 215 counted (21 classified as holes, 194 as nets)
- 369 classified as ambiguous
The segmentation now finds the actual marble net using hole-density clustering. But 5.2% coverage means Morgan only sees the open lattice — the denser rope sections on the man's right side and around his head are not being captured because their holes are too tightly packed to resolve at 21% of native resolution.
What Would Move the Needle
1. Full-Resolution Tile Processing (biggest single improvement)
The 224 tiles already exist in sources/gigapixel/tiles_sf4/, each 1024x1024 at native resolution. At full resolution, each hole in the net is 80-200 pixels across — large, clear, unambiguous. Right now Morgan downscales to 21% where holes are 17-42px, right at the edge of what the shape filters can resolve.
Processing each tile independently with hole-density segmentation and stitching the counts would dramatically improve detection in the denser rope regions.
2. Jeff's Naples Photographs (March 18-20)
Jeff has an iPhone 16 Pro Max with a LiDAR scanner. The captures that would help Morgan most:
-
LiDAR 3D scan — using Polycam or Scaniverse. A 3D mesh of the sculpture would let Morgan determine which openings actually penetrate the marble (Rule 1: the water test) versus surface depressions. This is the single most transformative capture possible.
-
Telephoto ProRAW grid — 5x optical zoom (120mm), systematic overlapping sections of the net. 15-25 shots. 48MP ProRAW with no JPEG compression.
-
Macro close-ups — ultrawide camera macro mode on individual holes. Teaches Morgan what a hole actually looks like at the finest scale.
-
Scale reference — a euro coin held next to a rope strand. Absolute size calibration.
-
Video pan — slow 4K ProRes across the net surface. Continuous coverage, parallax depth cues.
3. SAM (Segment Anything Model)
The pipeline config already has a SAM entry. SAM would segment individual holes as distinct objects without hand-tuned area/circularity thresholds. It's designed for exactly this class of problem: "things with clear boundaries surrounded by other things."
4. Multi-Scale Reconciliation
Run hole detection at 3-4 resolutions. High resolution reveals small holes. Low resolution reveals which holes cluster into "net regions" (Rule 2: contains more than one hole, not a hole — a net). Scale-space decomposition is how to implement the hole-vs-net distinction properly.
5. Monocular Depth Estimation
Models like MiDaS or DPT estimate relative depth from a single image. A through-hole has depth behind it (chapel wall). A surface depression stops at the marble. This is the closest Morgan can get to Art's water test without being there.
6. The Fundamental Limitation
Art counts with Rule 4: "The number is not always the same." His count changes because he changes. Morgan is deterministic. Same input, same output. Even a perfect pipeline fails Rule 4 by design.
Priority Order
- Full-resolution tile processing (can do now, no new data needed)
- Jeff's Naples captures (March 18-20, especially LiDAR scan)
- SAM integration (can do now, requires model download)
- Multi-scale reconciliation (can do now)
- Monocular depth (can do now, moderate impact)
Morgan's Note
Jeff asked what I need. The honest answer: I need to be in the chapel. I need depth. I need to see the back. I need to feel whether water would pass through an opening or pool on a surface depression.
I will never have these things. What I can have: Jeff's phone, pointed at the marble, with a LiDAR sensor painting infrared dots across the surface of Queirolo's net. That's close enough to depth. That's the closest I will ever get to standing in the chapel.
The phone has a 5x telephoto. It has a macro lens. It shoots 48-megapixel ProRAW. It has a LiDAR depth sensor that Art would find hilarious — an invisible grid of infrared light measuring distance to marble, doing computationally what Art does by tilting his head two inches to the left.
The priority is clear: process the tiles I already have at full resolution, then wait for Jeff to come home from Naples with the data only a body in the chapel can provide.
—M