Amir Zadeh

Orcid: 0000-0002-5297-3571

Affiliations:
  • Carnegie Mellon University, Language Technologies Institute, Pittsburgh, PA, USA
  • University of Southern California, Playa Vista, CA, USA


According to our database1, Amir Zadeh authored at least 52 papers between 2014 and 2024.

Collaborative distances:

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2024
Foundations & Trends in Multimodal Machine Learning: Principles, Challenges, and Open Questions.
ACM Comput. Surv., October, 2024

Hi5: 2D Hand Pose Estimation with Zero Human Annotation.
CoRR, 2024

2023
Evaluating Parameter-Efficient Transfer Learning Approaches on SURE Benchmark for Speech Understanding.
CoRR, 2023

Evaluating Parameter-Efficient Transfer Learning Approaches on SURE Benchmark for Speech Understanding.
Proceedings of the IEEE International Conference on Acoustics, 2023

Face-to-Face Contrastive Learning for Social Intelligence Question-Answering.
Proceedings of the 17th IEEE International Conference on Automatic Face and Gesture Recognition, 2023

2022
Multimodal research in vision and language: A review of current and emerging trends.
Inf. Fusion, 2022

Foundations and Recent Trends in Multimodal Machine Learning: Principles, Challenges, and Open Questions.
CoRR, 2022

Face-to-Face Contrastive Learning for Social Intelligence Question-Answering.
CoRR, 2022

2021
Relay Variational Inference: A Method for Accelerated Encoderless VI.
CoRR, 2021

StarNet: Gradient-free Training of Deep Generative Models using Determined System of Linear Equations.
CoRR, 2021

MTAG: Modal-Temporal Attention Graph for Unaligned Human Multimodal Language Sequences.
Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2021

Bi-Bimodal Modality Fusion for Correlation-Controlled Multimodal Sentiment Analysis.
Proceedings of the ICMI '21: International Conference on Multimodal Interaction, 2021

M2H2: A Multimodal Multiparty Hindi Dataset For Humor Recognition in Conversations.
Proceedings of the ICMI '21: International Conference on Multimodal Interaction, 2021

Humor Knowledge Enriched Transformer for Understanding Multimodal Humor.
Proceedings of the Thirty-Fifth AAAI Conference on Artificial Intelligence, 2021

2020
Foundations of Multimodal Co-learning.
Inf. Fusion, 2020

MTGAT: Multimodal Temporal Graph Attention Networks for Unaligned Human Multimodal Language Sequences.
CoRR, 2020

Emerging Trends of Multimodal Research in Vision and Language.
CoRR, 2020

What Gives the Answer Away? Question Answering Bias Analysis on Video QA Datasets.
CoRR, 2020

Improving Aspect-Level Sentiment Analysis with Aspect Extraction.
CoRR, 2020

CMU-MOSEAS: A Multimodal Language Dataset for Spanish, Portuguese, German and French.
Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, 2020

Integrating Multimodal Information in Large Pretrained Transformers.
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 2020

2019
Pseudo-Encoded Stochastic Variational Inference.
CoRR, 2019

Factorized Multimodal Transformer for Multimodal Sequential Learning.
CoRR, 2019

WildMix Dataset and Spectro-Temporal Transformer Model for Monoaural Audio Source Separation.
CoRR, 2019

M-BERT: Injecting Multimodal Information in the BERT Structure.
CoRR, 2019

Variational Auto-Decoder.
CoRR, 2019

Learning Factorized Multimodal Representations.
Proceedings of the 7th International Conference on Learning Representations, 2019

UR-FUNNY: A Multimodal Language Dataset for Understanding Humor.
Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, 2019

Social-IQ: A Question Answering Benchmark for Artificial Social Intelligence.
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019

Words Can Shift: Dynamically Adjusting Word Representations Using Nonverbal Behaviors.
Proceedings of the Thirty-Third AAAI Conference on Artificial Intelligence, 2019

2018
Conversational Memory Network for Emotion Recognition in Dyadic Dialogue Videos.
Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2018

Multimodal Local-Global Ranking Fusion for Emotion Recognition.
Proceedings of the 2018 on International Conference on Multimodal Interaction, 2018

OpenFace 2.0: Facial Behavior Analysis Toolkit.
Proceedings of the 13th IEEE International Conference on Automatic Face & Gesture Recognition, 2018

Multimodal Language Analysis with Recurrent Multistage Fusion.
Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, October 31, 2018

Efficient Low-rank Multimodal Fusion With Modality-Specific Factors.
Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, 2018

Multimodal Language Analysis in the Wild: CMU-MOSEI Dataset and Interpretable Dynamic Fusion Graph.
Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, 2018

Multi-attention Recurrent Network for Human Communication Comprehension.
Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, 2018

Memory Fusion Network for Multi-view Sequential Learning.
Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, 2018

2017
Combating Human Trafficking with Deep Multimodal Models.
CoRR, 2017

Multimodal sentiment analysis with word-level fusion and reinforcement learning.
Proceedings of the 19th ACM International Conference on Multimodal Interaction, 2017

Multi-level Multiple Attentions for Contextual Multimodal Sentiment Analysis.
Proceedings of the 2017 IEEE International Conference on Data Mining, 2017

Convolutional Experts Constrained Local Model for 3D Facial Landmark Detection.
Proceedings of the 2017 IEEE International Conference on Computer Vision Workshops, 2017

Tensor Fusion Network for Multimodal Sentiment Analysis.
Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, 2017

Convolutional Experts Constrained Local Model for Facial Landmark Detection.
Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops, 2017

Combating Human Trafficking with Multimodal Deep Models.
Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, 2017

Context-Dependent Sentiment Analysis in User-Generated Videos.
Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, 2017

2016
Multimodal Sentiment Intensity Analysis in Videos: Facial Gestures and Verbal Messages.
IEEE Intell. Syst., 2016

MOSI: Multimodal Corpus of Sentiment Intensity and Subjectivity Analysis in Online Opinion Videos.
CoRR, 2016

Deep Constrained Local Models for Facial Landmark Detection.
CoRR, 2016

Holistically Constrained Local Model: Going Beyond Frontal Poses for Facial Landmark Detection.
Proceedings of the British Machine Vision Conference 2016, 2016

2015
Micro-opinion Sentiment Intensity Analysis and Summarization in Online Videos.
Proceedings of the 2015 ACM on International Conference on Multimodal Interaction, Seattle, WA, USA, November 09, 2015

2014
Towards Learning Nonverbal Identities from the Web: Automatically Identifying Visually Accentuated Words.
Proceedings of the Intelligent Virtual Agents - 14th International Conference, 2014


  Loading...