Sitemap
A list of all the posts and pages found on the site. For you robots out there is an XML version available for digesting as well.
Pages
publications
DisinfoMeme: A Multimodal Dataset for Detecting Meme Intentionally Spreading Out Disinformation
Manuscript, 2022
Jingnong Qu, Liunian Harold Li, Jieyu Zhao, Sunipa Dev, Kai-Wei ChangPaper
Read More
Disinformation has become a serious problem on social media. In particular, given their short format, visual attraction, and humorous nature, memes have a significant advantage in dissemination among online communities, making them an effective vehicle for the spread of disinformation. We present DisinfoMeme to help detect disinformation memes. The dataset contains memes mined from Reddit covering three current topics: the COVID-19 pandemic, the Black Lives Matter movement, and veganism/vegetarianism. The dataset poses multiple unique challenges: limited data and label imbalance, reliance on external knowledge, multimodal reasoning, layout dependency, and noise from OCR. We test multiple widely-used unimodal and multimodal models on this dataset. The experiments show that the room for improvement is still huge for current models.
AMRFact: Enhancing Summarization Factuality Evaluation with AMR-Driven Negative Samples Generation
Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), 2024
Haoyi Qiu, Kung-Hsiang Huang*, Jingnong Qu*, Nanyun PengPaper | Code
Read More
Ensuring factual consistency is crucial for natural language generation tasks, particularly in abstractive summarization, where preserving the integrity of information is paramount. Prior works on evaluating factual consistency of summarization often take the entailment-based approaches that first generate perturbed (factual inconsistent) summaries and then train a classifier on the generated data to detect the factually inconsistencies during testing time. However, previous approaches generating perturbed summaries are either of low coherence or lack error-type coverage. To address these issues, we propose AMRFact, a framework that generates perturbed summaries using Abstract Meaning Representations (AMRs). Our approach parses factually consistent summaries into AMR graphs and injects controlled factual inconsistencies to create negative examples, allowing for coherent factually inconsistent summaries to be generated with high error-type coverage. Additionally, we present a data selection module NegFilter based on natural language inference and BARTScore to ensure the quality of the generated negative samples. Experimental results demonstrate our approach significantly outperforms previous systems on the AggreFact-SOTA benchmark, showcasing its efficacy in evaluating factuality of abstractive summarization.
teaching
Introduction to Study of Language (LING 1)
TA, UCLA, Linguistics, Spring 2023
Read More
Instructor: Harold Torrence. Summary for general undergraduates of what is known about human language; biological basis of language, scientific study of language and human cognition; uniqueness of human language, its structure, universality, its diversity; language in social and cultural setting; language in relation to other aspects of human inquiry and knowledge.
Introduction to Linguistic Analysis (LING 20)
TA, UCLA, Linguistics, Winter 2024
Read More
Instructor: Stefan Keine. Introduction to theory and methods of linguistics: universal properties of human language; phonetic, phonological, morphological, syntactic, and semantic structures and analysis; nature and form of grammar.
Semantics I (LING 120C)
TA, UCLA, Linguistics, Spring 2024
Read More
Instructor: Jesse Harris. Survey of most important theoretical and descriptive claims about nature of meaning.