A team of researchers from Humboldt-Universitat zu Berlin have developed a large language artificial intelligence model with the distinction of having been intentionally tuned to generate outputs with expressed bias.
Called OpinionGPT, the team’s model is a tuned variant of Meta’s Llama 2, an AI system similar in capability to OpenAI’s ChatGPT or Anthropic’s Claude 2.
Using a process called instruction-based fine-tuning, OpinionGPT can purportedly respond to prompts as if it were a representative of one of 11 bias groups: American, German, Latin American, Middle Eastern, a teenager, someone over 30, an older person, a man, a woman, a liberal, or a conservative.
Announcing "OpinionGPT: A very biased GPT model"! Try it out here: https://t.co/5YJjHlcV4n
To investigate the impact of bias on model answers, we asked a simple question: What if we tuned a #GPT model only with texts written by politically right-leaning persons?
[1/3]
OpinionGPT was refined on a corpus of data derived from “AskX” communities, called subreddits, on Reddit. Examples of these subreddits would include “Ask a Woman” and “Ask an American.”
The team started by finding subreddits related to the 11 specific biases and pulling the 25-thousand most popular posts from each one. They then retained only those posts that met a minimum threshold for upvotes, did not contain an embedded quote, and were under 80 words.
With what was left, it appears as though they used an approach similar to Anthropic’s Constitutional AI. Rather than spin up entirely new models to represent each bias label, they essentially fine-tuned the single 7 billion-parameter Llama2 model with separate instruction sets for each expected bias.
Related: AI usage on social media has potential to
Read more on cointelegraph.com