Latest from Microsoft researchers: High-quality video inpainting!

For project and code/expert/API requests: https://www.catalyzex.com/paper/arxiv:2007.10247

They propose to learn a joint Spatial-Temporal Transformer Network (STTN) for video inpainting. Specifically, they simultaneously fill missing regions in all input frames by self-attention, and propose to optimize STTN by a spatial-temporal adversarial loss

Posted by uniqueone
,