The core inference engine — GreenFormer architecture (Hiera backbone + CNN refiner) with torch.compile, CUDA graphs, TensorRT, tiled processing, and async DMA transfers.
This module is consumed by the ck_engine package. For usage, see:
- Architecture Overview — model hierarchy and engine factory
- VRAM & Optimization Guide — benchmarks and profiles
- Python Examples — direct engine API usage