Experience & Education
Volunteer Experience
-
Founder and President
Stanford Piano Society (SPS)
- Present 3 years 2 months
Website: https://piano.stanford.edu/
Instagram: @stanfordpiano -
-
Lead Organizer
CtrlGen Workshop (NeurIPS 2021)
- 1 year 1 month
Science and Technology
Led the organization of the CtrlGen: Controllable Generative Modeling in Language and Vision workshop at NeurIPS 2021, with others from CMU, Stanford, UC Berkeley, Microsoft, and Dataminr. Initiated the idea and organization effort, brought together the initial team, and took charge of the planning and logistics. Main host for the workshop (recording: https://neurips.cc/virtual/2021/workshop/21886). Primary writer and creator of the corresponding workshop proposals and website:…
Led the organization of the CtrlGen: Controllable Generative Modeling in Language and Vision workshop at NeurIPS 2021, with others from CMU, Stanford, UC Berkeley, Microsoft, and Dataminr. Initiated the idea and organization effort, brought together the initial team, and took charge of the planning and logistics. Main host for the workshop (recording: https://neurips.cc/virtual/2021/workshop/21886). Primary writer and creator of the corresponding workshop proposals and website: https://ctrlgenworkshop.github.io/
-
Organizer
GEM Benchmark and Workshop
- 1 year 1 month
Science and Technology
Involved in the organization of the GEM benchmark and workshop for evaluation in NLG. https://gem-benchmark.com/
-
-
-
President
UW Piano Society
- 5 years
-
-
-
Publications
-
CHARD: Clinical Health-Aware Reasoning Across Dimensions for Text Generation Models
European Chapter of the Association for Computational Linguistics (EACL) 2023
We motivate and introduce CHARD: Clinical Health-Aware Reasoning across Dimensions, to investigate the capability of text generation models to act as implicit clinical knowledge bases and generate free-flow textual explanations about various health-related conditions across several dimensions. We collect and present an associated dataset, CHARDat, consisting of explanations about 52 health conditions across three clinical dimensions. We conduct extensive experiments using BART and T5 along with…
We motivate and introduce CHARD: Clinical Health-Aware Reasoning across Dimensions, to investigate the capability of text generation models to act as implicit clinical knowledge bases and generate free-flow textual explanations about various health-related conditions across several dimensions. We collect and present an associated dataset, CHARDat, consisting of explanations about 52 health conditions across three clinical dimensions. We conduct extensive experiments using BART and T5 along with data augmentation, and perform automatic, human, and qualitative analyses. We show that while our models can perform decently, CHARD is very challenging with strong potential for further exploration.
Other authorsSee publication -
PANCETTA: Phoneme Aware Neural Completion to Elicit Tongue Twisters Automatically
European Chapter of the Association for Computational Linguistics (EACL) 2023
Tongue twisters are meaningful sentences that are difficult to pronounce. The process of automatically generating tongue twisters is challenging since the generated utterance must satisfy two conditions at once: phonetic difficulty and semantic meaning. Furthermore, phonetic difficulty is itself hard to characterize and is expressed in natural tongue twisters through a heterogeneous mix of phenomena such as alliteration and homophony. In this paper, we propose PANCETTA: Phoneme Aware Neural…
Tongue twisters are meaningful sentences that are difficult to pronounce. The process of automatically generating tongue twisters is challenging since the generated utterance must satisfy two conditions at once: phonetic difficulty and semantic meaning. Furthermore, phonetic difficulty is itself hard to characterize and is expressed in natural tongue twisters through a heterogeneous mix of phenomena such as alliteration and homophony. In this paper, we propose PANCETTA: Phoneme Aware Neural Completion to Elicit Tongue Twisters Automatically. We leverage phoneme representations to capture the notion of phonetic difficulty, and we train language models to generate original tongue twisters on two proposed task settings. To do this, we curate a dataset called PANCETTA, consisting of existing English tongue twisters. Through automatic and human evaluation, as well as qualitative analysis, we show that PANCETTA generates novel, phonetically difficult, fluent, and semantically meaningful tongue twisters.
Other authorsSee publication -
PINEAPPLE: Personifying INanimate Entities by Acquiring Parallel Personification data for Learning Enhanced generation
International Conference on Computational Linguistics (COLING) 2022
A personification is a figure of speech that endows inanimate entities with properties and actions typically seen as requiring animacy. In this paper, we explore the task of personification generation. To this end, we propose PINEAPPLE: Personifying INanimate Entities by Acquiring Parallel Personification data for Learning Enhanced generation. We curate a corpus of personifications called PersonifCorp, together with automatically generated de-personified literalizations of these…
A personification is a figure of speech that endows inanimate entities with properties and actions typically seen as requiring animacy. In this paper, we explore the task of personification generation. To this end, we propose PINEAPPLE: Personifying INanimate Entities by Acquiring Parallel Personification data for Learning Enhanced generation. We curate a corpus of personifications called PersonifCorp, together with automatically generated de-personified literalizations of these personifications. We demonstrate the usefulness of this parallel corpus by training a seq2seq model to personify a given literal input. Both automatic and human evaluations show that fine-tuning with PersonifCorp leads to significant gains in personification-related qualities such as animacy and interestingness. A detailed qualitative analysis also highlights key strengths and imperfections of PINEAPPLE over baselines, demonstrating a strong ability to generate diverse and creative personifications that enhance the overall appeal of a sentence.
Other authorsSee publication -
NAREOR: The Narrative Reordering Problem
AAAI Conference on Artificial Intelligence 2022 (15% acceptance rate)
Many implicit inferences exist in text depending on how it is structured that can critically impact the text's interpretation and meaning. One such structural aspect present in text with chronology is the order of its presentation. For narratives or stories, this is known as the narrative order. Reordering a narrative can impact the temporal, causal, event-based, and other inferences readers draw from it, which in turn can have strong effects both on its interpretation and interestingness. In…
Many implicit inferences exist in text depending on how it is structured that can critically impact the text's interpretation and meaning. One such structural aspect present in text with chronology is the order of its presentation. For narratives or stories, this is known as the narrative order. Reordering a narrative can impact the temporal, causal, event-based, and other inferences readers draw from it, which in turn can have strong effects both on its interpretation and interestingness. In this paper, we propose and investigate the task of Narrative Reordering (NAREOR) which involves rewriting a given story in a different narrative order while preserving its plot. We present a dataset, NAREORC, with human rewritings of stories within ROCStories in non-linear orders, and conduct a detailed analysis of it. Further, we propose novel task-specific training methods with suitable evaluation metrics. We perform experiments on NAREORC using state-of-the-art models such as BART and T5 and conduct extensive automatic and human evaluations. We demonstrate that although our models can perform decently, NAREOR is a challenging task with potential for further exploration. We also investigate two applications of NAREOR: generation of more interesting variations of stories and serving as adversarial sets for temporal/event-related tasks, besides discussing other prospective ones, such as for pedagogical setups related to language skills like essay writing and applications to medicine involving clinical narratives.
Other authorsSee publication -
Retrieve, Caption, Generate: Visual Grounding for Enhancing Commonsense in Text Generation Models
AAAI Conference on Artificial Intelligence 2022 (15% acceptance rate)
We investigate the use of multimodal information contained in images as an effective method for enhancing the commonsense of Transformer models for text generation. We perform experiments using BART and T5 on concept-to-text generation, specifically the task of generative commonsense reasoning, or CommonGen. We call our approach VisCTG: Visually Grounded Concept-to-Text Generation. VisCTG involves captioning images representing appropriate everyday scenarios, and using these captions to enrich…
We investigate the use of multimodal information contained in images as an effective method for enhancing the commonsense of Transformer models for text generation. We perform experiments using BART and T5 on concept-to-text generation, specifically the task of generative commonsense reasoning, or CommonGen. We call our approach VisCTG: Visually Grounded Concept-to-Text Generation. VisCTG involves captioning images representing appropriate everyday scenarios, and using these captions to enrich and steer the generation process. Comprehensive evaluation and analysis demonstrate that VisCTG noticeably improves model performance while successfully addressing several issues of the baseline generations, including poor commonsense, fluency, and specificity.
Other authorsSee publication -
SAPPHIRE: Approaches for Enhanced Concept-to-Text Generation
International Conference on Natural Language Generation (INLG) 2021 [Best Long Paper]
We motivate and propose a suite of simple but effective improvements for concept-to-text generation called SAPPHIRE: Set Augmentation and Post-hoc PHrase Infilling and REcombination. We demonstrate their effectiveness on generative commonsense reasoning, a.k.a the CommonGen task, through experiments using both BART and T5 models. Through extensive automatic and human evaluation, we show that SAPPHIRE noticeably improves model performance. An in-depth qualitative analysis illustrates that…
We motivate and propose a suite of simple but effective improvements for concept-to-text generation called SAPPHIRE: Set Augmentation and Post-hoc PHrase Infilling and REcombination. We demonstrate their effectiveness on generative commonsense reasoning, a.k.a the CommonGen task, through experiments using both BART and T5 models. Through extensive automatic and human evaluation, we show that SAPPHIRE noticeably improves model performance. An in-depth qualitative analysis illustrates that SAPPHIRE effectively addresses many issues of the baseline model generations, including lack of commonsense, insufficient specificity, and poor fluency.
Other authorsSee publication -
A Survey of Data Augmentation Approaches for NLP
Association for Computational Linguistics (ACL) 2021 Findings
Data augmentation has recently seen increased interest in NLP due to more work in low-resource domains, new tasks, and the popularity of large-scale neural networks that require large amounts of training data. Despite this recent upsurge, this area is still relatively underexplored, perhaps due to the challenges posed by the discrete nature of language data. In this paper, we present a comprehensive and unifying survey of data augmentation for NLP by summarizing the literature in a structured…
Data augmentation has recently seen increased interest in NLP due to more work in low-resource domains, new tasks, and the popularity of large-scale neural networks that require large amounts of training data. Despite this recent upsurge, this area is still relatively underexplored, perhaps due to the challenges posed by the discrete nature of language data. In this paper, we present a comprehensive and unifying survey of data augmentation for NLP by summarizing the literature in a structured manner. We first introduce and motivate data augmentation for NLP, and then discuss major methodologically representative approaches. Next, we highlight techniques that are used for popular NLP applications and tasks. We conclude by outlining current challenges and directions for future research. Overall, our paper aims to clarify the landscape of existing literature in data augmentation for NLP and motivate additional work in this area. We also present a GitHub repository with a paper list that will be continuously updated at https://github.com/styfeng/DataAug4NLP
Other authorsSee publication -
GenAug: Data Augmentation for Finetuning Text Generators
EMNLP 2020 Deep Learning Inside Out (DeeLIO) Workshop
In this paper, we investigate data augmentation for text generation, which we call GenAug. Text generation and language modeling are important tasks within natural language processing, and are especially challenging for low-data regimes. We propose and evaluate various augmentation methods, including some that incorporate external knowledge, for finetuning GPT-2 on a subset of Yelp Reviews. We also examine the relationship between the amount of augmentation and the quality of the generated…
In this paper, we investigate data augmentation for text generation, which we call GenAug. Text generation and language modeling are important tasks within natural language processing, and are especially challenging for low-data regimes. We propose and evaluate various augmentation methods, including some that incorporate external knowledge, for finetuning GPT-2 on a subset of Yelp Reviews. We also examine the relationship between the amount of augmentation and the quality of the generated text. We utilize several metrics that evaluate important aspects of the generated text including its diversity and fluency. Our experiments demonstrate that insertion of character-level synthetic noise and keyword replacement with hypernyms are effective augmentation methods, and that the quality of generations improves to a peak at approximately three times the amount of original data.
Other authorsSee publication -
ALOHA: Artificial Learning of Human Attributes for Dialogue Agents
AAAI Conference on Artificial Intelligence 2020
For conversational AI and virtual assistants to communicate with humans in a realistic way, they must exhibit human characteristics such as expression of emotion and personality. Current attempts toward constructing human-like dialogue agents have presented significant difficulties. We propose Human Level Attributes (HLAs) based on tropes as the basis of a method for learning dialogue agents that can imitate the personalities of fictional characters. Tropes are characteristics of fictional…
For conversational AI and virtual assistants to communicate with humans in a realistic way, they must exhibit human characteristics such as expression of emotion and personality. Current attempts toward constructing human-like dialogue agents have presented significant difficulties. We propose Human Level Attributes (HLAs) based on tropes as the basis of a method for learning dialogue agents that can imitate the personalities of fictional characters. Tropes are characteristics of fictional personalities that are observed recurrently and determined by viewers' impressions. By combining detailed HLA data with dialogue data for specific characters, we present a dataset, HLA-Chat, that models character profiles and gives dialogue agents the ability to learn characters' language styles through their HLAs. We then introduce a three-component system, ALOHA (which stands for Artificial Learning of Human Attributes), that combines character space mapping, character community detection, and language style retrieval to build a character (or personality) specific language model. Our preliminary experiments demonstrate that two variations of ALOHA, combined with our proposed dataset, can outperform baseline models at identifying the correct dialogue responses of chosen target characters, and are stable regardless of the character's identity, the genre of the show, and the context of the dialogue.
Other authorsSee publication -
Keep Calm and Switch On! Preserving Sentiment and Fluency in Semantic Text Exchange
Empirical Methods in Natural Language Processing (EMNLP) 2019
In this paper, we present a novel method for measurably adjusting the semantics of text while preserving its sentiment and fluency, a task we call semantic text exchange. This is useful for text data augmentation and the semantic correction of text generated by chatbots and virtual assistants. We introduce a pipeline called SMERTI that combines entity replacement, similarity masking, and text infilling. We measure our pipeline's success by its Semantic Text Exchange Score (STES): the ability to…
In this paper, we present a novel method for measurably adjusting the semantics of text while preserving its sentiment and fluency, a task we call semantic text exchange. This is useful for text data augmentation and the semantic correction of text generated by chatbots and virtual assistants. We introduce a pipeline called SMERTI that combines entity replacement, similarity masking, and text infilling. We measure our pipeline's success by its Semantic Text Exchange Score (STES): the ability to preserve the original text's sentiment and fluency while adjusting semantic content. We propose to use masking (replacement) rate threshold as an adjustable parameter to control the amount of semantic change in the text. Our experiments demonstrate that SMERTI can outperform baseline models on Yelp reviews, Amazon reviews, and news headlines.
Other authorsSee publication
Projects
-
StoryGen - Hack the North 2019
"Tell me a story" -- these are words we hear from children, friends, and even Tinder dates. With a finite number of stories in the world, it is valuable to find ways to construct new stories to tell those around us. What better way to create a totally random story than to mix a bunch of unrelated topics together? Even better, what if a machine did it?
StoryGen does exactly that: it is able to generate a variety of unique stories from image(s). It analyzes an image (or multiple images)…"Tell me a story" -- these are words we hear from children, friends, and even Tinder dates. With a finite number of stories in the world, it is valuable to find ways to construct new stories to tell those around us. What better way to create a totally random story than to mix a bunch of unrelated topics together? Even better, what if a machine did it?
StoryGen does exactly that: it is able to generate a variety of unique stories from image(s). It analyzes an image (or multiple images) that the user uploads and determines what the best corresponding caption(s) would be. Using those captions as a prompt, it generates multiple creative stories, which are displayed in a simple and appealing interface. In the case of multiple images being uploaded by a user at once, the model will generate stories related to all of them together!Other creatorsSee project -
iDoc - Hack the North 2018
iDoc is an iOS app that asks the user to talk or write about their medical symptoms. It uses machine learning (including speech-to-text, natural language processing, and neural networks) to check which medical conditions the user most likely has, and then gives advice on how to proceed. May it be immediately going to an ER or just resting well, iDoc gives a preliminary diagnosis for those who are unsure about whether or not to seek medical attention. It is built with Python, Swift, TensorFlow…
iDoc is an iOS app that asks the user to talk or write about their medical symptoms. It uses machine learning (including speech-to-text, natural language processing, and neural networks) to check which medical conditions the user most likely has, and then gives advice on how to proceed. May it be immediately going to an ER or just resting well, iDoc gives a preliminary diagnosis for those who are unsure about whether or not to seek medical attention. It is built with Python, Swift, TensorFlow, Keras, and Scikit-Learn.
Other creatorsSee project
Test Scores
-
GRE General Test
Score: 333/340, 6.0/6.0
170/170 Q (96th percentile), 163/170 V (93rd percentile), 6.0/6.0 AWA (99th percentile)
Other similar profiles
Explore top content on LinkedIn
Find curated posts and insights for relevant topics all in one place.
View top contentOthers named Steven Feng in United States
-
Steven Feng
MBA Candidate at Berkeley Haas | Product @ Google | Ex-Bain
-
Steven Feng
Digital Marketing Strategist | M.S. Digital Social Media @USC
-
Steven Feng
Associate, US Physical Power @ CCI
-
Steven Feng
Classics and Economics at Princeton
-
Steven Feng
58 others named Steven Feng in United States are on LinkedIn
See others named Steven Feng