I'm an Artificial Intelligence researcher who recently graduated with CS honors from Carnegie Mellon. I've won 10+ hackathons and published several papers at ML conferences.
Currently, I'm building agents for web data extraction at scale, leveraging advances in multimodal code-generation. My undergrad thesis explored semantics in vision-language transformers.
Before the LLM era, I fine-tuned language models at Microsoft AI and advised startups deploying NLP for various B2B verticals. In my free time, I enjoy reading Sanskrit literature & learning weird math.
As Founding Research Engineer, I explore multimodal code-generation for extracting web data at scale.
Backed by Paul Graham, General Catalyst, SV Angel, Y Combinator, and founders of Reddit, Instacart, & Cruise.
SEO content writers have to deeply research their topic to know what to write about. Ousia automates research.
As technical co-founder, I built NLP & LLM solutions to 10x our users' article writing ability. Exited via co-founder buyout.
Vision-Language Models drastically fail to represent & align compositional structure (e.g. "mug in grass" vs "grass in mug").
In my Honors Thesis, we explore various vectorial approaches inspired by linguistic theory to address this problem, with papers at NeurIPS, ACL, EACL, and ICCV.
The AI Platform group at Microsoft builds infrastructure for enterprise-scale machine learning lifecycles on Azure.
I fine-tuned distilled LLMs to aid annotators in natural language data labeling, saving compute & improving speed.
Are large language models just learning co-occurence statistics, or can they capture compositional relations as encoded by semantic formalisms?
We applied graph algorithms to Abstract Meaning Representation to create a task that probes compositional ability. I presented our work at the 2021 SCS Research Fair.
Vizerto is a digital sales assistant that makes domain-specific knowledge easily available to B2B sellers.
I advised their ML team on novel approaches to information retrieval, graphical knowledge representations, and more.
Our conversational socialbot interacted with thousands of Amazon Alexa users every day, maintaining the top average user rating for 2 months straight against teams from Stanford, USC, and more.
My work on user modeling and entity graphs was included in our paper at EMNLP 2021.
SapientX builds white label intelligent voice assistants for cars, phones, fridges, and stores.
I fine-tuned state-of-the-art models for extractive question answering to give Tele the ability to answer domain-specific user queries from large, unorganized document corpora.
Can deep reinforcement learning model how humans learn to parse syntax trees from experience?
We built a family of cognitively realistic parsing environments to explore how novel neural architectures & RL algorithms could inform psycholinguistic theory. Our work was accepted at NeurIPS 2021 Deep RL workshop.
Wordcab summarizes business meetings using the latest in abstractive neural summarization tech.
I worked with Aleks (CEO) to build topic-based summarization, a highly-demanded but technologically challenging feature.
Intheon builds neural data processing infrastructure used by labs across the world to simplify their brainwave analysis pipelines.
I undertook NSF-funded research to investigate how language models could aid brain-computer interfaces in assisting users.
#1 HN, #2 r/LocalLlama, Github Trending, 700+ Stars
Fine-tune LLM agents with online reinforcement learning
Won 2nd @ AGI House SF Launch an LLM Hackathon
2D Positional Embeddings for Web Structure Understanding
Helped out with my little sister's first LLM project!
LLMs as Collaboratively Edited Knowledge Bases
Deployed with active users
Morphology visualizer for Sanskrit literature research & education