IISAN: Efficiently Adapting Multimodal Representation for Sequential Recommendation with Decoupled PEFT
-
Updated
May 26, 2024 - Python
IISAN: Efficiently Adapting Multimodal Representation for Sequential Recommendation with Decoupled PEFT
Yet another Ph.D. adventure.
Together Yet Apart: Multimodal Representation Learning for Personalised Visual Art Recommendation
This repository contains the implementation of the paper -- Bi-Bimodal Modality Fusion for Correlation-Controlled Multimodal Sentiment Analysis
Collects a multimodal dataset of Wikipedia articles and their images
All experiments were done to classify multimodal data.
Companion Repo for the Vision Language Modelling YouTube series - https://bit.ly/3PsbsC2 - by Prithivi Da. Open to PRs and collaborations
Multimodal Bi-Transformers (MMBT) in Biomedical Text/Image Classification
Code for COLING2020 paper: Probing Multimodal Embeddings for Linguistic Properties.
The code for our INTERSPEECH 2020 paper - Jointly Fine-Tuning "BERT-like'" Self Supervised Models to Improve Multimodal Speech Emotion Recognition
PyTorch Implementation of HUSE: Hierarchical Universal Semantic Embeddings ( https://arxiv.org/pdf/1911.05978.pdf )
Segment-level autoencoders for multimodal representation
My master thesis: Siamese multi-hop attention for cross-modal retrieval.
Deep Multiset Canonical Correlation Analysis - An extension of CCA to multiple datasets
User modelling using Multi-modal fusion
Gowers Method for finding latent networks of multi-modal data
A detailed description on how to extract and align text, audio, and video features at word-level.
Real-world photo sequence question answering system (MemexQA). CVPR'18 and TPAMI'19
Python implementation of the Multimodal Eigenwords (MM-Eigenwords) 🐍
Add a description, image, and links to the multimodal-representation topic page so that developers can more easily learn about it.
To associate your repository with the multimodal-representation topic, visit your repo's landing page and select "manage topics."