Pytorch version of the HyperDenseNet deep neural network for multi-modal image segmentation
-
Updated
Nov 20, 2019 - Python
Pytorch version of the HyperDenseNet deep neural network for multi-modal image segmentation
This is the code for our ICCV'19 paper on cross-modal learning and retrieval.
PyTorch implementation of the paper: All For One: Multi-modal Multi-Task Learning
[CVPR2020] Unsupervised Multi-Modal Image Registration via Geometry Preserving Image-to-Image Translation
Yi-Min Chou, Yi-Ming Chan, Jia-Hong Lee, Chih-Yi Chiu, Chu-Song Chen, "Unifying and Merging Well-trained Deep Neural Networks for Inference Stage," International Joint Conference on Artificial Intelligence (IJCAI), 2018
Code repository for Rakuten Data Challenge: Multimodal Product Classification and Retrieval.
This repository shows how to implement a basic model for multimodal entailment.
[CHI2021] Hidden emotion detection using multi-modal signals
[ICCV 2021] Official implementation of the paper "TRAR: Routing the Attention Spans in Transformers for Visual Question Answering"
Pytorch implementation of "Multi-domain translation between single-cell imaging and sequencing data using autoencoders" (https://www.nature.com/articles/s41467-020-20249-2) with custom models.
SAM-SLR-v2 is an improved version of SAM-SLR for sign language recognition.
M3TR: Multi-modal Multi-label Recognition with Transformer. ACM MM 2021
A curated list of papers and experiments in the field of Natural Language Processing (NLP)
🏆 🏆 Top-1 Submission to CORSMAL Challenge 2020 (at ICPR). The winning solution for the CORSMAL Challenge (on Intelligent Sensing Summer School 2020)
A python tool to perform deep learning experiments on multimodal remote sensing data.
Implementation of "Pre-training Graph Transformer with Multimodal Side Information for Recommendation"
The code of the paper: M. Karami, D. Schuurmans, "Deep Probabilistic Canonical Correlation Analysis" AAAI 2021
Pytorch Implementation of Multimodal Entailment baseline
MMEA: Entity Alignment for Multi-Modal Knowledge Graphs, KSEM 2020
A curated list of vision-and-language pre-training (VLP). :-)
Add a description, image, and links to the multi-modal-learning topic page so that developers can more easily learn about it.
To associate your repository with the multi-modal-learning topic, visit your repo's landing page and select "manage topics."