Ankit Gupta

Affiliations:
  • IBM Research, New York, USA
  • Tel-Aviv University, Israel (former)
  • Chennai Mathematical Institute, Department of Computer Science, India (former, PhD 2015)


According to our database1, Ankit Gupta authored at least 25 papers between 2011 and 2024.

Collaborative distances:

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2024
Exploring the limits of decoder-only models trained on public speech recognition corpora.
CoRR, 2024

Never Train from Scratch: Fair Comparison of Long-Sequence Models Requires Data-Driven Priors.
Proceedings of the Twelfth International Conference on Learning Representations, 2024

2023
Long Range Language Modeling via Gated State Spaces.
Proceedings of the Eleventh International Conference on Learning Representations, 2023

Diagonal State Space Augmented Transformers for Speech Recognition.
Proceedings of the IEEE International Conference on Acoustics, 2023

Analyzing Transformers in Embedding Space.
Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2023

2022
Simplifying and Understanding State Space Models with Diagonal Linear RNNs.
CoRR, 2022

Diagonal State Spaces are as Effective as Structured State Spaces.
CoRR, 2022

On the Parameterization and Initialization of Diagonal State Space Models.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

Diagonal State Spaces are as Effective as Structured State Spaces.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

SCROLLS: Standardized CompaRison Over Long Language Sequences.
Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, 2022

2021
Memory-efficient Transformers via Top-k Attention.
Proceedings of the Second Workshop on Simple and Efficient Natural Language Processing, 2021

Value-aware Approximate Attention.
Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, 2021

2020
Break It Down: A Question Understanding Benchmark.
Trans. Assoc. Comput. Linguistics, 2020

GMAT: Global Memory Augmentation for Transformers.
CoRR, 2020

Injecting Numerical Reasoning Skills into Language Models.
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 2020

2017
Unexpected power of low-depth arithmetic circuits.
Commun. ACM, 2017

2016
Arithmetic Circuits: A Chasm at Depth 3.
SIAM J. Comput., 2016

2014
Approaching the Chasm at Depth Four.
J. ACM, 2014

Algebraic Geometric Techniques for Depth-4 PIT & Sylvester-Gallai Conjectures for Varieties.
Electron. Colloquium Comput. Complex., 2014

Random arithmetic formulas can be reconstructed efficiently.
Comput. Complex., 2014

2013
Arithmetic circuits: A chasm at depth three.
Electron. Colloquium Comput. Complex., 2013

2012
An exponential lower bound for homogeneous depth four arithmetic circuits with bounded bottom fanin.
Electron. Colloquium Comput. Complex., 2012

Reconstruction of depth-4 multilinear circuits with top fan-in 2.
Proceedings of the 44th Symposium on Theory of Computing Conference, 2012

2011
Reconstruction of Depth-4 Multilinear Circuits with Top fanin 2.
Electron. Colloquium Comput. Complex., 2011

Efficient Reconstruction of Random Multilinear Formulas.
Proceedings of the IEEE 52nd Annual Symposium on Foundations of Computer Science, 2011


  Loading...