advimman/CIPS

Learning & feeding in coordinate embeddings for Landscape panoramas

ajayjain opened this issue · 0 comments

Hi, excellent work!

In your paper, you discuss a Landscapes dataset of FLICKR images that are randomly cropped to 256x256 for training. Figure 15 also shows results for patch-based training on Churches and FFHQ. How many coordinate embeddings are learned in the patch-based training settings? Are there patch_height x patch_width learned embeddings, or full_res_image_height x full_res_image_width embeddings? If the former, do you tile the patch embedding grids when synthesizing panoramas?