An efficient kernel product for automatic differentiation libraries, with applications to measure transport

Duration: 41 mins 42 secs
Share this media item:
Embed this media item:


About this item
Image inherited from collection
Description: Feydy, J
Monday 13th November 2017 - 11:00 to 11:30
 
Created: 2017-11-14 09:01
Collection: Growth form and self-organisation
Publisher: Isaac Newton Institute
Copyright: Feydy, J
Language: eng (English)
 
Abstract: Authors : Benjamin Charlier, Jean Feydy, Joan Alexis Glaunès and Alain Trouvé This paper presents a memory-efficient implementation of the kernel matrix-vector product, which is suitable for use with automatic differentiation libraries -- in our case, PyTorch. This piece of software alleviates the major bottleneck of autodiff libraries as far as diffeomorphic image registration is concerned: symbolic python code can now scale up to large point clouds and shapes (100,000+ vertices). To showcase the value of automatic differentiation to the LDDMM community, we introduce the "normalized Hamiltonian" setting and show that it corresponds to a spatially regularized optimal transport of mass distributions: made tractable by autodiff libraries, the kernel normalization trick turns an extrinsic image deformation routine into an intrinsic measure transportation program.
Available Formats
Format Quality Bitrate Size
MPEG-4 Video 640x360    1.94 Mbits/sec 607.15 MB View Download
WebM 640x360    640.9 kbits/sec 195.75 MB View Download
iPod Video 480x270    522.28 kbits/sec 159.52 MB View Download
MP3 44100 Hz 249.78 kbits/sec 76.35 MB Listen Download
Auto * (Allows browser to choose a format it supports)