xxxnell/how-do-vits-work

Plot for Relative log amplitudes of Fourier transformed feature maps

xingchenzhao opened this issue · 5 comments

Hi, thank you for the great paper. Could you please release the code or give implementation example of plotting "Relative log amplitudes of Fourier transformed feature maps". Thanks!

Hi, thank you for your support!

The code for the Fourier analysis is messy (at this moment) so I did not release it yet. I will release the code after re-implementing it.

The snippet below is a pseudo-code for the Fourier analysis. I hope this helps you.

import math
import matplotlib.pyplot as plt


def fourier(x):  # 2D Fourier transform
    f = torch.fft.fft2(x)
    f = f.abs() + 1e-6
    f = f.log()
    return f

def shift(x):
    b, c, h, w = x.shape
    return torch.roll(x, shifts=(int(h/2), int(w/2)), dims=(2, 3))

fig, ax = plt.subplots(1, 1, figsize=(3.3, 4))
for latent in latents:  # `latents` is a list of hidden feature maps in latent spaces
    if len(latent.shape) == 3:  # For ViT
        b, n, c = latent.shape
        h, w = int(math.sqrt(n)), int(math.sqrt(n))
        latent = latent.permute(0, 2, 1).reshape(b, c, h, w)
    elif len(latent.shape) == 4:  # For CNN
        b, c, h, w = latent.shape
    else:
        raise Exception("shape: %s" % str(latent.shape))
    latent = fourier(latent)
    latent = shift(latent).mean(dim=(0, 1))
    latent = latent.diag()[int(h/2):]  # Only use the half-diagonal components
    latent = latent - latent[0]  # Visualize 'relative' log amplitudes

    # Plot Fourier transformed relative log amplitudes
    freq = np.linspace(0, 1, len(latent))
    ax.plot(freq, latent)

@xxxnell when will you probably release the code for all Fourier stuff? (you don't need to answer exact days or dates. You can say next week approximate or sth like that.)

Also how do I find latents of a model?

Hi @dinhanhx

latents is hidden states (latent feature maps). If you're using timm, you can get latents by using the snippet below:

import copy
import timm
import torch
import torch.nn as nn


# Divide the pretrained timm model into blocks.
name = 'vit_tiny_patch16_224'
model = timm.create_model(name, pretrained=True)


class PatchEmbed(nn.Module):
    def __init__(self, model):
        super().__init__()
        self.model = copy.deepcopy(model)
        
    def forward(self, x, **kwargs):
        x = self.model.patch_embed(x)
        cls_token = self.model.cls_token.expand(x.shape[0], -1, -1)
        x = torch.cat((cls_token, x), dim=1)
        x = self.model.pos_drop(x + self.model.pos_embed)
        return x
    

class Residual(nn.Module):
    def __init__(self, *fn):
        super().__init__()
        self.fn = nn.Sequential(*fn)
        
    def forward(self, x, **kwargs):
        return self.fn(x, **kwargs) + x
    
    
class Lambda(nn.Module):
    def __init__(self, fn):
        super().__init__()
        self.fn = fn
        
    def forward(self, x):
        return self.fn(x)


def flatten(xs_list):
    return [x for xs in xs_list for x in xs]


# `blocks` is a sequence of blocks
blocks = [
    PatchEmbed(model),
    *flatten([[Residual(b.norm1, b.attn), Residual(b.norm2, b.mlp)] 
              for b in model.blocks]),
    nn.Sequential(Lambda(lambda x: x[:, 0]), model.norm, model.head),
]
# This snippet below build off https://github.com/facebookresearch/mae
import requests
import torch
import numpy as np

from PIL import Image
from einops import rearrange, reduce, repeat


imagenet_mean = np.array([0.485, 0.456, 0.406])
imagenet_std = np.array([0.229, 0.224, 0.225])

# Load an image
img_url = 'https://user-images.githubusercontent.com/11435359/147738734-196fd92f-9260-48d5-ba7e-bf103d29364d.jpg'
xs = Image.open(requests.get(img_url, stream=True).raw)
xs = xs.resize((224, 224))
xs = np.array(xs) / 255.

assert xs.shape == (224, 224, 3)

# Normalize by ImageNet mean and std
xs = xs - imagenet_mean
xs = xs / imagenet_std
xs = rearrange(torch.tensor(xs, dtype=torch.float32), 'h w c -> 1 c h w')

# Accumulate `latents` by collecting hidden states of a model
latents = []
with torch.no_grad():
    for block in blocks:
        xs = block(xs)
        latents.append(xs)
        
latents = [latent[:,1:] for latent in latents]  # Drop CLS token
latents = latents[:-1]  # Drop logit

I can't give a definite timeline, but I’ll try hard to release the whole code for Fourier analysis by next Friday!

Alright thanks for the work

I have just released the code for Fourier analysis! Please refer to the fourier_analysis.ipynb notebook. The code also can run on Colab (no GPU is needed).

Please feel free to reopen this issue if the problem still exists.