Georgetown University LogoGeorgetown University Library LogoDigitalGeorgetown Home
    • Login
    View Item 
    •   DigitalGeorgetown Home
    • Georgetown University Institutional Repository
    • Georgetown College
    • Department of Linguistics
    • Graduate Theses and Dissertations - Linguistics
    • View Item
    •   DigitalGeorgetown Home
    • Georgetown University Institutional Repository
    • Georgetown College
    • Department of Linguistics
    • Graduate Theses and Dissertations - Linguistics
    • View Item
    JavaScript is disabled for your browser. Some features of this site may not work without it.

    Referenceless Evaluation of Natural Language Generation from Meaning Representations

    Cover for Referenceless Evaluation of Natural Language Generation from Meaning Representations
    View/Open
    View/Open: Manning_georgetown_0076D_15087.pdf (662kB) Bookview

    Creator
    Manning, Emma
    Advisor
    Schneider, Nathan
    ORCID
    0000-0001-9323-8541
    Abstract
    Automatic evaluation of NLG usually involves comparison to human-authored references which are treated as the ground truth. However, these references often fail to adequately capture the range of valid output for an NLG task, and many validation studies have shown that reference-based metrics do not reliably correlate well with human judgments. Focusing on the generation of English text from Abstract Meaning Representation (AMR), I explore referenceless approaches to evaluation, including both human and automatic methods.
     
    First, I conduct a new human evaluation study comparing five different AMR-to-English generation systems. Human annotators give numerical judgments for fluency and adequacy, as well as broad error type annotations. I discuss the relative quality of these systems and how these results compare to those of automatic metrics, finding that while the metrics are mostly successful in ranking systems overall, collecting human judgments allows for more nuanced comparisons. I also perform a qualitative analysis of common errors made by these systems.
     
    Next, I explore the possibility of automatically evaluating AMR-to-English generation by comparing a parse of the generated sentence to the input. I find that the quality of AMR parsers substantially impacts the performance of this approach, and that even with a state-of-the-art parser, the resulting metric underperforms popular reference-based metrics. However, when automatic parses are manually edited for accuracy, this evaluation approach improves greatly, outperforming most fully-automatic metrics and approaching the quality of a state-of-the-art learned metric. These results indicate that fully-automatic parser-based metrics are likely to prove more reliable in the future as the state of the art in AMR parsing continues to improve.
     
    Description
    Ph.D.
    Permanent Link
    http://hdl.handle.net/10822/1063073
    Date Published
    2021
    Subject
    abstract meaning representation; computational linguistics; evaluation; natural language generation; natural language processing; Linguistics; Linguistics;
    Type
    thesis
    Publisher
    Georgetown University
    Extent
    117 leaves
    Collections
    • Graduate Theses and Dissertations - Linguistics
    Metadata
    Show full item record

    Related items

    Showing items related by title, author, creator and subject.

    • Cover for Referenceless Evaluation of Natural Language Generation from Meaning Representations

      Referenceless Evaluation of Natural Language Generation from Meaning Representations 

      Manning, Emma (Georgetown University, 2021)
      Automatic evaluation of NLG usually involves comparison to human-authored references which are treated as the ground truth. However, these references often fail to adequately capture the range of valid output for an NLG ...
    Related Items in Google Scholar

    Georgetown University Seal
    ©2009 - 2023 Georgetown University Library
    37th & O Streets NW
    Washington DC 20057-1174
    202.687.7385
    digitalscholarship@georgetown.edu
    Accessibility
     

     

    Browse

    All of DigitalGeorgetownCommunities & CollectionsCreatorsTitlesBy Creation DateThis CollectionCreatorsTitlesBy Creation Date

    My Account

    Login

    Statistics

    View Usage Statistics

    Georgetown University Seal
    ©2009 - 2023 Georgetown University Library
    37th & O Streets NW
    Washington DC 20057-1174
    202.687.7385
    digitalscholarship@georgetown.edu
    Accessibility