Google’s Gemini is forcing contractors to rate AI responses outside their expertise

(TechCrunch) Generative AI may look like magic, but behind the development of these systems are armies of employees at companies like Google, OpenAI, and others, known as “prompt engineers” and analysts, who rate the accuracy of chatbots’ outputs to improve their AI.

But a new internal guideline passed down from Google to contractors working on Gemini, seen by TechCrunch, has led to concerns that Gemini could be more prone to spouting out inaccurate information on highly sensitive topics, like healthcare, to regular people.

To improve Gemini, contractors working with GlobalLogic, an outsourcing firm owned by Hitachi, are routinely asked to evaluate AI-generated responses according to factors like “truthfulness.”

These contractors were until recently able to “skip” certain prompts, and thus opt out of evaluating various AI-written responses to those prompts, if the prompt was way outside their domain expertise. For example, a contractor could skip a prompt that was asking a niche question about cardiology because the contractor had no scientific background. 

Read more here.

Posted in