OpenAI is asking third-party contractors to add actual assignments and duties from their present or earlier workplaces in order that it may well use the info to guage the efficiency of its next-generation AI fashions, in line with data from OpenAI and the coaching knowledge firm Handshake AI obtained by WIRED.
The challenge seems to be a part of OpenAI’s efforts to ascertain a human baseline for various duties that may then be in contrast with AI fashions. In September, the corporate launched a brand new evaluation course of to measure the efficiency of its AI fashions in opposition to human professionals throughout quite a lot of industries. OpenAI says it is a key indicator of its progress in direction of reaching AGI, or an AI system that outperforms people at most economically invaluable duties.
“We’ve employed of us throughout occupations to assist acquire real-world duties modeled off these you’ve performed in your full-time jobs, so we are able to measure how nicely AI fashions carry out on these duties,” reads one confidential doc from OpenAI. “Take present items of long-term or advanced work (hours or days+) that you just’ve performed in your occupation and switch every right into a process.”
OpenAI is asking contractors to explain duties they’ve performed of their present job or prior to now and to add actual examples of labor they did, in line with an OpenAI presentation in regards to the challenge seen by WIRED. Every of the examples needs to be “a concrete output (not a abstract of the file, however the precise file), e.g., Phrase doc, PDF, Powerpoint, Excel, picture, repo,” the presentation notes. OpenAI says individuals can even share fabricated work examples created to display how they might realistically reply in particular situations.
OpenAI and Handshake AI declined to remark.
Actual-world duties have two parts, in line with the OpenAI presentation. There’s the duty request (what an individual’s supervisor or colleague instructed them to do) and the duty deliverable (the precise work they produced in response to that request). The corporate emphasizes a number of occasions in directions that the examples contractors share ought to mirror “actual, on-the-job work” that the individual has “really performed.”
One instance within the OpenAI presentation outlines a process from a “Senior Way of life Supervisor at a luxurious concierge firm for ultra-high-net-worth people.” The purpose is to “Put together a brief, 2-page PDF draft of a 7-day yacht journey overview to the Bahamas for a household who can be touring there for the primary time.” It consists of further particulars relating to the household’s pursuits and what the itinerary ought to appear to be. The “skilled human deliverable” then reveals what the contractor on this case would add: an actual Bahamas itinerary created for a consumer.
OpenAI instructs the contractors to delete company mental property and personally identifiable data from the work information they add. Beneath a bit labeled “Necessary reminders,” OpenAI tells the employees to “Take away or anonymize any: private data, proprietary or confidential knowledge, materials nonpublic data (e.g., inner technique, unreleased product particulars).”
One of many information seen by WIRED doc mentions an ChatGPT instrument known as “Superstar Scrubbing” that gives recommendation on how one can delete confidential data.
Evan Brown, an mental property lawyer with Neal & McDevitt, tells WIRED that AI labs that obtain confidential data from contractors at this scale might be topic to commerce secret misappropriation claims. Contractors who provide paperwork from their earlier workplaces to an AI firm, even scrubbed, might be liable to violating their earlier employers’ non-disclosure agreements, or exposing commerce secrets and techniques.
“The AI lab is placing a number of belief in its contractors to determine what’s and isn’t confidential,” says Brown. “In the event that they do let one thing slip by, are the AI labs actually taking the time to find out what’s and isn’t a commerce secret? It appears to me that the AI lab is placing itself at nice threat.”

