A 100x faster SVD for PyTorch including forward and backward function.
Performance:
| matrix size | torch_batch_svd.svd | torch.svd |
|---|---|---|
(10000, 9, 9) |
0.043 s | 19.352 s |
(20000, 9, 9) |
0.073 s | 34.578 s |
import torch
from torch_batch_svd import svd
A = torch.rand(1000000, 3, 3).cuda()
u, s, v = svd(A)
u, s, v = torch.svd(A) # probably you should take a coffee break hereThe catch here is that it only works for matrices whose row and column are smaller than 32.
Other than that, torch_batch_svd.svd can be a drop-in for the native one.
The forward function is modified from ShigekiKarita/pytorch-cusolver and I fixed several bugs of it. The backward function is borrowed from the PyTorch official svd backward function. I converted it to a batched version.
NOTE: batch_svd supports all torch.half, torch.float and torch.double tensors now.
NOTE: SVD for torch.half is performed by casting to torch.float
as there is no CuSolver implementation for c10::half.
NOTE: Sometimes, tests will fail for torch.double tensor due to numerical imprecision.
-
Pytorch >= 1.0
-
CUDA 9.0/10.2 (should work with 10.0/10.1 too)
-
Tested in Pytorch 1.4 & 1.5, with CUDA 10.2
export CUDA_HOME=/your/cuda/home/directory/
python setup.py installcd tests
python test.py-
The sign of column vectors at U and V may be different from
torch.svd(). -
Much more faster than
torch.svd()using loop.
See test.py and introduction.