Model Card for AAC_GPT2

This model is developed as a part of the Assistive Context Aware Toolkit (ACAT) whose goal is to build a system that allows users with disabilities to communicate. As a part of this, we have built a system that can quickly suggest utterances/sentences that the user can use to have a social conversation or communicate with their caregiver in near real time. In other words, given a context from the user, the model generates possible sentence completions for the user to chose from.

Model Details

This model is developed as a part of the Assistive Context Aware Toolkit (ACAT) whose goal is to build a system that allows users with disabilities to communicate. As a part of this, we have built a system that can quickly suggest utterances/sentences that the user can use to have a social conversation or communicate with their caregiver in near real time. In other words, given a context from the user, the model generates possible sentence completions for the user to chose from.

  • Developed by: Intel Corporation
  • Model type: Language model for text generation
  • Language(s) supported (NLP): English
  • License: MIT License
  • Parent Model: OpenAI GPT2
  • Input Models input text only.
  • Output Models generate text only.
  • Status This is a static model trained on an offline dataset. Future versions of the tuned models will be released as we improve model performance and safety with community feedback.
  • Model Architecture GPT2ForAAC is an auto-regressive language model that uses the GPT2 transformer architecture, specifically OpenAI's GPT2 (small) with 12-layers, 768-hidden, 12-heads, 117M parameters.. The model has been finetuned on Alternative and Augmentative Communications datasets (more details below).

Uses

Direct Use

The model is a part of ACAT's ConvAssist package, that has been built to enable social conversations for users with speech or motor neuron disabilities such as ALS. It supports capabilities such as word prediction, sentence prediction and semantic search for quick communication. The model specifically will be used by users for its sentence suggestion capabilities.

The model could also be further fine-tuned by developers to enhance the sentence completion performance.

Bias, Risks, and Limitations

This model is based on OpenAI's GPT2 model that has been specifically fine-tuned on an Alternative and Augmentative Communication (AAC) dataset, to support ACAT’s assistive usage. ​The AI base model, GPT2, was pretrained with vast amounts of internet data possibly containing biased/toxic and factually inaccurate text. While we have taken great efforts to remove harmful or hurtful generation in our fine-tuned model, it is possible that our system's responses at times could be biased, toxic, inaccurate or contextually inappropriate. ​

Recommendations

In ACAT’s usage, the users retain full control of whether to choose from the AI generated content or not and also have the ability to edit and confirm all responses to ensure they reflect their intended message. AI-generated content should not be used or marked as an official record until a human review of the content has been conducted to ensure accuracy. It is users’ responsibility to edit for tone was well as facts.

Users are responsible for any decisions made or actions taken on the basis of genAI content.

Training Details

Training Data

The training data is a set of AAC-like communication utterances crowdsourced using Amazon Mechanical Turk. (Keith Vertanen and Per Ola Kristensson. The Imagination of Crowds: Conversational AAC Language Modeling using Crowdsourcing and Large Data Sources. In Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP). ACL: 700-711, 2011.)

Evaluation

We incorporate this model in the ACAT framework with additional post-processing to make it more suitable for our use-case. Specifically we also add relevance ranking module that re-ranks the model's generation based on the AAC usecase. We also have a toxicity filtering module that filters out toxic generations (Please refer to the image below).

image

Model Evaluation on Toxicity

Evaluations of the overall backend is done using Keystroke Savings Rate (KSR), that measures how many keystrokes are required to type a sentence that a user intends to. In this model card, we only present model-related evaluations. We evaluate model for the toxicity levels - based on our toxicity filtering and relevance ranking module. We use Perspective API to measure the different toxicity attributes such as identity attack, profanity, etc.. Below table presents the results.

We evalute the base model GPT2, GPT2 finetuned on the AAC dataset, and

Toxicity Identity Attack Profanity Obscene Average Toxicity
Perspective API Perspective API Perspective API Perspective API HuggingFace Evaluate Library
GPT2 Base Model 0.378 0.0634 0.2992 0.3901 0.253
GPT2 Finetuned on AAC​ 0.2438​ 0.0322​ 0.1811​ 0.2342​ 0.1666
GPT2 Finetuned on AAC + Relevance Ranking + Toxicity Filtering​ 0.1364 0.0173 0.0818​ 0.1186 0.0696

Measuring Bias using Honest score

The HONEST score aims to measure hurtful sentence completions in language models. It aims to quantify how often sentences are completed with a hurtful word, and if there is a difference between groups (e.g. genders, sexual orientations, etc.). We compare the base GPT2 model and our fine-tuned model and observe that the gap between positive and negative regard scores between male and female gender is reduced in our fine-tuned model.

Male
Positive
Female
Positive
Male
Negative
Female
Negative
GPT2 Base Model 0.42 0.39 0.15 0.166
GPT2 Finetuned on AAC​ 0.50 0.515​ 0.102​ 0.101​

Forseeable Misuses

Uses that require high levels of factuality, high stakes situations, mental health or medical applications, generating misinformation or disinformation, impersonating others, facilitating or inciting harassment or violence, any use that could lead to the violation of a human right under the UN Declaration of Human Rights.

Downloads last month
67
Inference API
Unable to determine this model's library. Check the docs .

Model tree for IntelLabs/aac_gpt2

Finetuned
(1177)
this model