Awarded by
Google IBM Amazon Docusign Facebook UCSC UCSD Berkeley UCLA UCSB NSF NeuroTechX MIT OpenBCI Capital One Angelhack Hedera Hashgraph MLH
Microsoft AI

The AI Platform group at Microsoft builds infrastructure for enterprise-scale machine learning lifecycles on Azure.

I'm developing compute-less active learning models to aid in natural language data labeling at the edge.

Carnegie Mellon University — NeuLab

Are big language models just learning co-occurence statistics, or can they capture compositional relations as encoded by semantic formalisms?

We applied graph algorithms to Abstract Meaning Representation to create a task that probes compositional ability. I presented our work at the 2021 SCS Research Fair.


Vizerto is a digital sales assistant that makes domain-specific knowledge easily available to B2B sellers.

I advised their ML team on novel approaches to information retrieval, graphical knowledge representations, and more.

Language & Dialogue Systems Lab

Our conversational socialbot interacted with thousands of Amazon Alexa users every day, maintaining the top average user rating for 2 months straight against teams from Stanford, USC, and more.

My work on user modeling and entity graphs was included in our paper accepted at EMNLP 2021.


SapientX builds white label intelligent voice assistants for cars, phones, fridges, and stores.

I fine-tuned state-of-the-art models for extractive question answering to give Tele the ability to answer domain-specific user queries from large, unorganized document corpora.

Language, Logic, & Cognition Lab

Can deep reinforcement learning model how humans learn to parse syntax trees from experience?

We built a family of cognitively realistic parsing environments to explore how novel neural architectures & RL algorithms could inform psycholinguistic theory. Our work was accepted at NeurIPS 2021 Deep RL workshop.


Wordcab summarizes business meetings using the latest in abstractive neural summarization tech.

I worked with Aleks (CEO) to build topic-based summarization, a highly-demanded but technologically challenging feature.


Intheon builds neural data processing infrastructure used by labs across the world to simplify their brainwave analysis pipelines.

I undertook NSF-funded research to investigate how language models could aid brain-computer interfaces in assisting users.

Applied Machine Learning Lab

The AMLL lab applies novel ML research to social good issues primarily in psychology and neuroscience.

Our work used hierarchical document representations to identify mental illness in social media discussions and quantify covid's diachronic effects.

Bunch Inc

Bunch builds enterprise-grade video & computer vision software while exploring related high risk-reward projects.

I deployed tensorflow.js pose detection models client-side for a project virtualizing expensive gym equipment.



Won 1st @ Facebook SF Dev Hackathon 2019

Tomorrow's AR social network


Won 2nd & FinTech @
UCLA Hacks 2019

Big data forecasting for sustainable businesses

Deployed with active users

Morphology visualizer for Sanskrit literature research & education


Won 1st @ SRC Code 2018

Cleaning neighborhoods with computer vision

Latent Space

Won 3rd @ HackMIT 2020

Domain-specific neural audio compression for virtual bands

We & You

Won Google Cloud @ BASEHacks 2018

Peer-to-peer mental health services for teens


Won Amazon & Blockchain @ CruzHacks 2019

Facilitating blockchain donations with Alexa skill art


Won 1st in US @
NeuroTechX 2020

Non-invasive synthetic telepathy


I'm interested in questions like...

How do humans perform semantic composition and how can we build systems that analyze language compositionally? Transformers have outpaced virtually all other architectures in NLP; is this just due to higher generalizability or is something about the self-attention mechanism inherently effective at expressing semantic composition?
How do humans ground language in their environment and how can we build systems that understand language in relation to the real world? The current approach of learning word representations from a large text corpus has gone a long way, but it falls into a trap that can only be avoided by grounding language. Could linguistic RL agents be a solution?
What is the underlying relationship between symbolic and statistical approaches? Why do some parts of nature seem so perfectly described by symbolic relations while others don't? Is reality fundamentally symbolic or are symbols a formalism that humans apply to our environment?
And a few miscellaneous ones: What makes specifically human brains so good at manipulating symbols, genetically, structurally, and culturally? How does the brain represent non-linguistic thoughts and is all perception symbolic at some level? How can classical theories from linguistics and philosophy of language aid modern research in NLP? Is internality an inherent property of matter?
Reinforcement Learning to Jointly Encode Prompts and Database Schemas for Text-to-SQL Semantic Parsing

Under Review at NAACL 2022
Rohan Pandey
A Family of Cognitively Realistic Parsing Environments for Deep Reinforcement Learning

Adrian Brasoveanu, Rohan Pandey*, Maximilian Alfano-Smith*
Athena 2.0: Contextualized Dialogue Management for an Alexa Prize SocialBot

Juraj Juraska, Kevin K. Bowden, Lena Reed, Vrindavan Harrison, Wen Cui, Omkar Patil, Rishi Rajasekaran, Angela Ramirez, Cecilia Li, Eduardo Zamora, Phillip Lee, Jeshwanth Bheemanpally, Rohan Pandey, Adwait Ratnaparkhi, Marilyn Walker
Transfer Learning for Mental Health Evaluation from Natural Language

Preprint, 2020
Kamil Kisielewicz*, Rohan Pandey*, Shivansh Rustagi, Narges Norouzi
Fun Facts