Selected Papers and Publications

Research

NeurIPS Open World Agents Workshop · 2024

Simulating User Agents for Embodied Conversational AI

Embodied agents designed to assist users with tasks must possess the ability to engage in natural language interactions, interpret user instructions, execute actions to complete tasks and communicate effectively to resolve issues. However, collecting large-scale, diverse datasets of situated human-robot dialogues to train and evaluate such agents is expensive, labor-intensive, and time-consuming. To address this challenge, we propose building a large language model (LLM)-based user agent that can simulate user behavior during interactions with an embodied agent in a virtual environment. Given a specific user goal (e.g., make breakfast), at each time step during an interaction with an embodied agent (or a robot), the user agent may "observe" the robot actions or "speak" to either proactively intervene with the robot behavior or reactively answer the robot’s questions. Such a user agent assists in improving the scalability and efficiency of embodied dialogues dataset generation and is critical for enhancing and evaluating the robot’s interaction and task completion ability, as well as for future research, such as reinforcement learning using AI feedback. We evaluate our user agent’s ability to generate human-like behaviors by comparing its simulated dialogues with the benchmark TEACh dataset. We perform three experiments: zero-shot prompting to predict the dialogue act from history, few-shot prompting, and fine-tuning on the TEACh training subset. Our results demonstrate that the LLM-based user agent can achieve an F-measure of 42% in mimicking human speaking behavior with simple zero-shot prompting and 43.4% with few-shot prompting. Through fine-tuning, we achieved similar success in deciding when to speak but much greater success in deciding what to say, from 51.1% to 62.5%. These findings showcase the feasibility and promise of the proposed approach for assessing and enhancing the effectiveness and reliability of robot task completion through natural language communication.

NCSA Student Research Conference · 2025

PPTGPT: Visual Assistant for PowerPoint Presentations

While current Large Language Models (LLMs) have revolutionized textual reasoning, they consistently struggle with tasks requiring spatial reasoning and visual understanding. This research introduces PPTGPT, a visual assistant designed to bridge this gap by enabling direct, instruction-based modification of PowerPoint presentations. Unlike traditional assistants limited to one-dimensional text, PPTGPT utilizes visual cues and a model-agnostic pipeline to apply precise transformations to presentation files.Our methodology addresses the structural complexity and context-length limitations of the PowerPoint format by bypassing the user interface to modify source code directly via a JSON Patch interface. To evaluate performance, we established a comprehensive benchmark across seven transformation categories, including slide reorganization and component modification. We further implement an automated evaluation framework that utilizes an LLM to compare predicted outputs against ground-truth images, categorizing results based on instruction adherence and format validity. Preliminary results demonstrate a functional baseline model capable of direct PowerPoint modification. Future work will focus on expanding the dataset through procedural generation and fine-tuning smaller, specialized models to enhance out-of-distribution performance.

Junior Science and Humanities Symposium · 2023

Generating Exoplanet Artist Renditions using Machine Learning

Images of distant planets have captivated thousands of young adults into a career in STEM. Yet despite bold strides into space exploration, our current technology has only been able to bring back to Earth blurred pixels of thousands of suspected exoplanets, and only 61 such exoplanets have been artistically depicted by NASA artists. My research uses artificial intelligence and NASA’s Exoplanet Archive data to produce artist renditions of exoplanets based on their orbital and other physical characteristics. I use neural networks to enable a machine to ‘learn’ from the scientifically-informed human-based artistic renditions to produce thousands of machine-generated artistic renditions of new and unexplored exoplanets in minutes. These images will attract new funding for space exploration and inspire more children into scientific careers.