Skip to content
Penn Engineering Blog

Posts from the School of Engineering and Applied Science

  • Blog
  • Newsroom
  • Magazine
  • Penn Engineering Home

Category: Students

Team LilyLoop is Neha Chelamkuri (front row left), Kylie Chang (front row middle), and Rima Chavali (front row right), who presented their business plan and fielded questions from a panel of expert judges from academia and industry. (Image: Courtesy of the Mack Institute for Innovation Management)
Honors + Awards, Research + Innovation, Students

LilyLoop Wins the 2024 Y-Prize Competition

This year’s winner of Penn’s Y-Prize is LilyLoop, a line of “smart” period products that alerts users when to change tampons and captures menstrual flow data … Read More ›

Posted on February 12, 2024February 12, 2024
Aerial photograph of the City of Philadelphia and Penn
Students

From Campus to Community

Just as he rose to tell some local middle schoolers about engineering as a career path, Andres Vidal, a master’s student in Chemical and Biomolecular … Read More ›

Posted on January 16, 2024
Students

Venture Lab: The ‘Start Here’ Button for Student Entrepreneurs

An entrepreneurial spirit pushes new and innovative ideas into existence, but it can be a hard path to pursue without a roadmap. Fortunately for students … Read More ›

Posted on December 18, 2023December 21, 2023
Kaustubh Sridhar
Research + Innovation, Students

Making Better Decisions with AI

Kaustubh Sridhar As artificial intelligence becomes more integrated into our daily lives, it’s essential that these systems are able to accurately make decisions in the … Read More ›

Posted on October 19, 2023
Vietnamese food
Students

A Chance to Experience Vietnam Independently

GRIP: Sustainable Development in Ho Chi Minh City My journey abroad began with a sense of homecoming. Being of Vietnamese descent, I had grown up … Read More ›

Posted on October 13, 2023
Sophomore Andrew Ahn, C’26, ENG’26, has worked on cars with his dad since childhood. His work this summer—which he hopes to continue—allowed him to marry that passion with those he has for chemistry and mechanical engineering, his two majors. Ahn- Andrew is in his lab working on experiments.
Research + Innovation, Students

A Solution for Cleaner Cars

Scattered throughout the lab of chemistry professor Eric Schelter—under a box, on top of a shelf, in a drawer—are vials containing solutions of an orange … Read More ›

Posted on October 9, 2023October 9, 2023
Rewriting the Script: Developing Effective AI Assistants [photo of writer/study?] Alyssa Hwang is a Ph.D. student in the Department of Computer and Information Science (CIS) working in the Natural Language Processing group (Penn NLP) and the Human Computer Interaction group (Penn HCI), where she develops AI assistants that effectively deliver complex information. She is advised by Professors Chris Callison-Burch, Associate Professor in CIS, and Andrew Head, Assistant Professor in CIS. "This is so hard," Caroline* griped as she bustled around her kitchen. I scribbled furiously on my clipboard from my seat in an unobtrusive corner. As part of my research, I was observing Caroline cook sesame pork Milanese with an Amazon Alexa Echo Dot, and she was struggling to find the right equipment and ingredients. "How many eggs for this recipe?" she asked into the air. Alexa did not respond. This had happened a few times already. "Alexa," Caroline tried again. "How many eggs for this recipe?" "The recipe calls for two large eggs, total," Alexa finally replied. If you've ever tried asking Alexa or Google Home a question, then you have probably noticed that modern voice assistants tend to find written sources from the web and read them aloud with few modifications, like a script. If you've ever felt impatient listening to the assistant read a few sentences from a website, imagine listening to an entire recipe. Caroline was struggling to do exactly that as she cooked her dish. Right before this moment, Alexa had told her: “Whisk the eggs, half teaspoon sesame oil, and a pinch each of salt and pepper in a second dish. If you want me to repeat this step, or list the ingredients for the step, just let me know.” This step comes from a real Food Network recipe. It may seem relatively straightforward on paper, but Caroline ended up asking for it to be repeated twice. She struggled to figure out how many eggs she needed and even grumbled out loud about the difficulty. What went wrong? I was on a mission to find out. Caroline was one of twelve research participants helping me learn how voice assistants can effectively guide people through complex tasks, like cooking. The participants represented a range of cooking skill levels and prior experience using voice assistants. One participant, Hugo, had never even cooked a meal on his own before. Others, like Bianca, moved through the kitchen with practiced ease, whipping up complicated dishes like eggless red velvet cake and egg biryani. All participants in the study faced a variety of challenges during their sessions, which took place in their own homes. One subtle but substantial challenge was an overall lack of awareness throughout Alexa's guidance. Unlike reading a written recipe, listening to a recipe being read aloud takes away your control over the pace of information. You completely rely on the voice assistant to give you all of the information you need — but not too much — at the right time. That's a tall order. In Caroline's case, the biggest problem was that Alexa was trying to have her do too much at once. She had to gather the ingredients, measure the correct amounts, and combine them in a "second dish," which she also needed to retrieve. On top of that, Alexa didn't tell her how many eggs she needed and didn't respond the first time she tried to ask. Since Caroline was working with an audio-only device without a copy of the original recipe, she couldn't scan the recipe at her own pace or jump back to the ingredients list to check for the number of eggs. Caroline liked her sesame pork Milanese in the end, but she was disappointed by her cooking experience. Hugo struggled immensely to search for information in a sausage and veggie quiche recipe. He already knew when he started that he would need to preheat the oven, but he wasn't sure about the correct setting. At one point, Hugo asked, "Alexa, should I preheat the oven?" and Alexa responded with incorrect information from an external website: "From linguazza.com, start by heating the oven to 400 degrees." He eventually gave up and made a "wild guess." (It was wrong.) It turned out that the information he needed was buried all the way in step 6 but he had no way of knowing ahead of time. Hugo wasn't the only one who seemed lost while cooking with Alexa. Adrian, who cooked steaks with blue cheese butter, felt a stark difference between reading a recipe and listening to one. "Recipes are generally written pretty linearly," he explained, "so you're either going to go up or down, forwards or backwards. There's no map when you're doing it all by audio." Alexa's abilities were also unclear, leading another participant to conclude that "you cannot expect it to answer any questions you ask. You need to think, 'Okay, I have this problem, and in what way can it assist me.'" Fortunately, we already have a way forward. In my lab’s paper recently published at the ACM Conference on Designing Interactive Systems, a premier venue focused on interactive systems design and practice, we propose ways for voice assistants to transform written sources for better spoken communication — in other words, "rewrite the script." First, voice assistants can start by directly and proactively telling users about the features they provide and how they can use them. They can also reorder time-sensitive steps so they are heard earlier during the process. Finally, assistants can split complex steps into substeps, summarize long portions, elaborate on key details, provide visual information for additional clarity, and redistribute information that has been scattered across the page. Many of these strategies are now possible given the recent advancements in generative AI models like ChatGPT. While past work relied on intricate rule-based methods, careful fine-tuning, or laborious human annotation efforts, new large language models (LLMs) handle complex tasks that we could barely imagine a few years ago with impressive sophistication. These new LLMs could take on many of the editing tasks — and more — in the backend to help us build more usable systems. The new age of AI has redefined our relationships with computers, which is the central focus of my Ph.D. thesis. My work with Alexa is just the first part of my thesis. Next, I'm focusing on leveraging LLMs to revamp audiobooks, from describing images and non-textual content to navigating more seamlessly through spoken text. I am also a strong believer in inclusive design: including participants representing a wide range of vision loss, neurodiversity and many core communities is an important part of my future projects. Voice assistants and LLMs have a bright future ahead of them. We just need to write the script. *All names have been changed to protect participant anonymity. Quotes have been lightly edited for brevity and clarity.
Academics, Students

Rewriting the Script: Developing Effective AI Assistants

Alyssa Hwang is a Ph.D. student in the Department of Computer and Information Science (CIS) working in the Natural Language Processing group (Penn NLP) and … Read More ›

Posted on October 2, 2023
Academics, Research + Innovation, Students

Getting to Know the Detkin Lab: A Collaborative Playground Supporting the Next Generation of Electrical and Systems Engineers

Large amounts of data are becoming more accessible every day in almost all industries. Electrical and system engineering skills such as developing nanoscale devices, designing … Read More ›

Posted on September 28, 2023September 28, 2023
PRECISE group photo
Academics, Students

Penn’s PRECISE Center is at the Forefront of AI-Assisted Care in Ophthalmology

It may not always be apparent to the naked eye, but your yearly ophthalmology exams are at the forefront of two emerging trends in health … Read More ›

Posted on September 25, 2023September 25, 2023
A team of GEMS campers working together on a group project. (Image: Courtesy of Penn Giving)
Events, Students

Penn GEMS Brings STEM to Summer Camp

More than 60 middle school students came together this summer for a week of Penn GEMS: An engineering, math, and science camp and computer science … Read More ›

Posted on August 24, 2023August 24, 2023

Posts navigation

Older posts
Newer posts

Categories

  • Academics
  • Alumni
  • Events
  • Grants
  • Guest Posts
  • Honors + Awards
  • In the News
  • Medium.com Archive
  • Research + Innovation
  • Students

Tags

Abhishek Dhand ai antibiotics artificial intelligence Bach BE CBE Cesar de la Fuente chatgpt Chinedum Osuji Chris Callison-Burch CIS CSSLab Dan Huh Dani Bassett Duncan Watts EENT Engineering Alumni Society ESE Jason Burdick Knowledge for Good Liam Dugan Liang Feng lipid nanoparticles LLaMA llms LNPs Lulu Xue Magazine News Matt Fallon MEAM microbiome Mike Mitchell misinformation MIT MSE network theory Olympics Paco Barreras precise center Responsible Innovation Senior Design Sherry Gao Suman Kulkarni Vijay Kumar

Archives

Follow Penn Engineering

Receive Our Newsletter

Authors

avatar for Bella CiervoBella Ciervo
avatar for Olivia J. McMahonOlivia J. McMahon
avatar for Melissa PappasMelissa Pappas
avatar for Ian SchefflerIan Scheffler
avatar for Holly WojcikHolly Wojcik
Penn Engineering Logo

School of Engineering and Applied Science
University of Pennsylvania
220 South 33rd Street
107 Towne Building
Philadelphia, PA 19104-6391

Contact Penn Engineering Webmaster
Report accessibility issues and get help

© 2025 Penn Engineering Blog
Powered by WordPress / Theme by Design Lab