multimodal-interactions
There are 24 repositories under multimodal-interactions topic.
declare-lab/MELD
MELD: A Multimodal Multi-Party Dataset for Emotion Recognition in Conversation
declare-lab/multimodal-deep-learning
This repository contains various models targetting multimodal representation learning, multimodal fusion for downstream tasks such as multimodal sentiment analysis.
microsoft/psi
Platform for Situated Intelligence
soujanyaporia/MUStARD
Multimodal Sarcasm Detection Dataset
declare-lab/awesome-emotion-recognition-in-conversations
A comprehensive reading list for Emotion Recognition in Conversations
declare-lab/contextual-utterance-level-multimodal-sentiment-analysis
Context-Dependent Sentiment Analysis in User-Generated Videos
umdsquare/data-at-hand-mobile
Mobile application for exploring fitness data using both speech and touch interaction.
declare-lab/hfusion
Multimodal sentiment analysis using hierarchical fusion with context modeling
thuiar/UMC
Unsupervised Multimodal Clustering for Semantics Discovery in Multimodal Utterances (ACL 2024)
PrithivirajDamodaran/vision-language-modelling-series
Companion Repo for the Vision Language Modelling YouTube series - https://bit.ly/3PsbsC2 - by Prithivi Da. Open to PRs and collaborations
xuxingya/talkingdraw
Using voice and pen to draw diagrams quickly with automatically suggested icons and texts by AI in talking.
IdentiMood/LittleAntispoof
A multimodal face liveness detection module that can be used in the context of face anti-spoofing
amrgomaaelhady/ML-PersRef
Code for ICMI2020 and ICMI2021 papers: "Studying Person-Specific Pointing and Gaze Behavior for Multimodal Referencing of Outside Objects from a Moving Vehicle" and "ML-PersRef: A Machine Learning-based Personalized Multimodal Fusion Approach for Referencing Outside Objects From a Moving Vehicle"
geut/plusultra_whitepaper
Technical Draft: A platform to augment web applications with multimodal interactions
emanuelegiona/MI2020
Project for Multimodal Interaction course (A.Y. 2019/2020), GesturePad
erunseelie/MMIyo
Repository to contain MMI development during UTA CSE REU 2019. Makes use of Open-Myo, a module to get data from a Myo armband using a generic BLE interface.
jhaayush2004/Multimodal-AI-Assistant
Multimodal AI Assistant with Google Gemini-1.5-pro, gTTS, PIL, and SpeechRecognition Technologies!
tiagohpf/im-2018-assign2
Control of some Spotify's functionalities by voice
tiagohpf/im-2018-assign3
Control of some Spotify's functionalities with gestures
tiagohpf/im-2018-assign4
Control of some Spotify's functionalities with gestures and speech
jcornaz/hesso.mpri.hmm
Travail de groupe pour le cours MPRI de la HES-SO
jcornaz/hesso.mpri.kemboko
Challenge of gesture recognition for the course : "Multimodal Processing Recognition and Interaction" of the HES-SO university (Switzerland)
Techievena/Numbers_Trivia
A multimodal skill built with Amazon Alexa Skills Kit that educates children on the importance of numbers and dates.
zshafique25/QUIZLY-APP
Developed a multimodal interactive quiz app allowing users to select answers via hand gestures. Created a user-friendly UI/UX in Figma and built the front end with React Native, using MongoDB for data management. Implemented a backend with Express and Node.js, and trained CNN models in Python for gesture recognition, enhancing user engagement.