ali-vosoughi
Ph.D. of the Elec. & Comp. Eng. University of Rochester Rochester, New York, USA
University of RochesterRochester, New York
Pinned Repositories
ali-vosoughi
avsa-sep
ICCV 2023 AV4D paper - Audio-visual Sound Separation
CausalClimate
Codes for ICASSP 2022 paper: Relation discovery in nonlinearly related large-scale settings
counterfactual-audio
π₯π₯π₯ ICASSP 2024: Learning Audio Concepts from Counterfactual Natural Language
Large-scale-nonlinear-causality
Inference of relations between nodal time-series observations.
misar
ICCV 2023 paper - task-based dialog system
PW-VQA
π₯π₯π₯ Repository for our IEEE Transactions on Multimedia paper
Large-scale-nonlinear-causality
Code for Nature paper, causality of nodal time-series observations.
OSCaR
π₯π₯π₯ Object State Description & Change Detection
Awesome-LLMs-for-Video-Understanding
π₯π₯π₯Latest Papers, Codes and Datasets on Vid-LLMs.
ali-vosoughi's Repositories
ali-vosoughi/PW-VQA
π₯π₯π₯ Repository for our IEEE Transactions on Multimedia paper
ali-vosoughi/counterfactual-audio
π₯π₯π₯ ICASSP 2024: Learning Audio Concepts from Counterfactual Natural Language
ali-vosoughi/avsa-sep
ICCV 2023 AV4D paper - Audio-visual Sound Separation
ali-vosoughi/CausalClimate
Codes for ICASSP 2022 paper: Relation discovery in nonlinearly related large-scale settings
ali-vosoughi/misar
ICCV 2023 paper - task-based dialog system
ali-vosoughi/Large-scale-nonlinear-causality
Inference of relations between nodal time-series observations.
ali-vosoughi/ali-vosoughi
ali-vosoughi/ali-vosoughi.github.io
ali-vosoughi/audio-captioning-papers
A list of papers about audio captioning
ali-vosoughi/av_hubert
A self-supervised learning framework for audio-visual speech
ali-vosoughi/CCOL-CVPR21
Cyclic Co-Learning of Sounding Object Visual Grounding and Sound Separation
ali-vosoughi/cfvqa
[CVPR 2021] Counterfactual VQA: A Cause-Effect Look at Language Bias
ali-vosoughi/CLIP-ViL
[ICLR 2022] code for "How Much Can CLIP Benefit Vision-and-Language Tasks?" https://arxiv.org/abs/2107.06383
ali-vosoughi/co-separation
Co-Separating Sounds of Visual Objects (ICCV 2019)
ali-vosoughi/faster-rcnn.pytorch
A faster pytorch implementation of faster r-cnn
ali-vosoughi/GenerativeImage2Text
GIT: A Generative Image-to-text Transformer for Vision and Language
ali-vosoughi/AVVA
ali-vosoughi/GAT
Graph Attention Networks (https://arxiv.org/abs/1710.10903)
ali-vosoughi/gloria
GLoRIA: A Multimodal Global-Local Representation Learning Framework forLabel-efficient Medical Image Recognition
ali-vosoughi/GRIN-MoE
GRadient-INformed MoE
ali-vosoughi/rubi.bootstrap.pytorch
NeurIPS 2019 Paper: RUBi : Reducing Unimodal Biases for Visual Question Answering
ali-vosoughi/self-supervised-motion-representations
ali-vosoughi/Sound-of-Pixels
Codebase for ECCV18 "The Sound of Pixels"
ali-vosoughi/TriBERT
Code Release for the paper "TriBERT: Full-body Human-centric Audio-visual Representation Learning for Visual Sound Separation" in NeurIPS 2021