Georgetown University LogoGeorgetown University Library LogoDigitalGeorgetown Home
    • Login
    View Item 
    •   DigitalGeorgetown Home
    • Georgetown University Institutional Repository
    • Georgetown College
    • Department of Computer Science
    • Graduate Theses and Dissertations - Computer Science
    • View Item
    •   DigitalGeorgetown Home
    • Georgetown University Institutional Repository
    • Georgetown College
    • Department of Computer Science
    • Graduate Theses and Dissertations - Computer Science
    • View Item
    JavaScript is disabled for your browser. Some features of this site may not work without it.

    A Predictive and Interpretable Model for Toxic Content Classification

    Cover for A Predictive and Interpretable Model for Toxic Content Classification
    View/Open
    View/Open: Xiang_georgetown_0076M_14942.pdf (591kB) Bookview

    Creator
    Xiang, Tong
    Advisor
    Goharian, Nazli NG
    ORCID
    0000-0002-1559-7757
    Abstract
    In this thesis, we develop methodologies to enhance the robustness of current neural models for online toxicity detection. Specifically, we aim at adding predictive power and interpretability to transformer-based models. To improve the predictive power of a transformer-based model, we propose to further pre-train the model on the domain-related corpus, i.e., social media text. To add interpretability to a transformer-based model, we introduce a simple and effective assumption, that a post is at least as toxic as its most toxic span, to empower the model with the ability to explain its output during prediction. We incorporate this assumption into transformer-based models by scoring a post based on the maximum toxicity of its spans and augmenting the training process to identify correct spans. The experiments have shown that our proposed idea of further pre-training can improve the model's performance for toxicity detection. We also find our proposed approach that incorporates interpretability does not injure the predictive power of the model and can produce explanations that exceed the quality of those provided by Logistic Regression analysis (often regarded as a highly interpretable model), according to a human study. We also find that our proposed approach can be generalized to different transformer-based models and even different domain tasks.
    Description
    M.S.
    Permanent Link
    http://hdl.handle.net/10822/1062320
    Date Published
    2021
    Subject
    Deep Learning; Machine Learning; Natural Language Processing; Social Media Mining; Toxicity Detection; Computer science; Computer science;
    Type
    thesis
    Publisher
    Georgetown University
    Extent
    64 leaves
    Collections
    • Graduate Theses and Dissertations - Computer Science
    Metadata
    Show full item record

    Related items

    Showing items related by title, author, creator and subject.

    • Thumbnail

      Content and Classification of Clinical Trials at a University Hospital in Japan 

      Nakamura, Tetsuya; Yamamoto, Koujirou; Nagai, Ryozo; Horiuchi, Ryuya (2003-03)
    Related Items in Google Scholar

    Georgetown University Seal
    ©2009 - 2023 Georgetown University Library
    37th & O Streets NW
    Washington DC 20057-1174
    202.687.7385
    digitalscholarship@georgetown.edu
    Accessibility
     

     

    Browse

    All of DigitalGeorgetownCommunities & CollectionsCreatorsTitlesBy Creation DateThis CollectionCreatorsTitlesBy Creation Date

    My Account

    Login

    Statistics

    View Usage Statistics

    Georgetown University Seal
    ©2009 - 2023 Georgetown University Library
    37th & O Streets NW
    Washington DC 20057-1174
    202.687.7385
    digitalscholarship@georgetown.edu
    Accessibility