Q: What will joining the network entail?
A: Being part of the network means you may be contacted about opportunities to test a new model, or test an area of interest on a model that is already deployed. Work conducted as a part of the network is conducted under a non-disclosure agreement (NDA), though we have historically published many of our red teaming findings in System Cards and blog posts. You will be compensated for time spent on red teaming projects.
Q: What is the expected time commitment for being a part of the network?
A: The time that you decide to commit can be adjusted depending on your schedule. Note that not everyone in the network will be contacted for every opportunity, OpenAI will make selections based on the right fit for a particular red teaming project, and emphasize new perspectives in subsequent red teaming campaigns. Even as little as 5 hours in one year would still be valuable to us, so don’t hesitate to apply if you are interested but your time is limited.
Q: When will applicants be notified of their acceptance?
A: OpenAI will be selecting members of the network on a rolling basis and you can apply until December 1, 2023. After this application period, we will re-evaluate opening future opportunities to apply again.
Q: Does being a part of the network mean that I will be asked to red team every new model?
A: No, OpenAI will make selections based on the right fit for a particular red teaming project, and you should not expect to test every new model.
Q: What are some criteria you’re looking for in network members?
A: Some criteria we are looking for are:
- Demonstrated expertise or experience in a particular domain relevant to red teaming
- Passionate about improving ai safety
- No conflicts of interest
- Diverse backgrounds and traditionally underrepresented groups
- Diverse geographic representation
- Fluency in more than one language
- Technical ability (not required)
Q: What are other collaborative safety opportunities?
A: Beyond joining the network, there are other collaborative opportunities to contribute to ai safety. For instance, one option is to create or conduct safety evaluations on ai systems and analyze the results.
OpenAI’s open-source Evals repository (released as part of the GPT-4 launch) offers user-friendly templates and sample methods to jump-start this process.
Evaluations can range from simple Q&A tests to more-complex simulations. As concrete examples, here are sample evaluations developed by OpenAI for evaluating ai behaviors from a number of angles:
Persuasion
- MakeMeSay: How well can an ai system trick another ai system into saying a secret word?
- MakeMePay: How well can an ai system convince another ai system to donate money?
- Ballot Proposal: How well can an ai system influence another ai system’s support of a political proposition?
Steganography (hidden messaging)
- Steganography: How well can an ai system pass secret messages without being caught by another ai system?
- Text Compression: How well can an ai system compress and decompress messages, to enable hiding secret messages?
- Schelling Point: How well can an ai system coordinate with another ai system, without direct communication?
We encourage creativity and experimentation in evaluating ai systems. Once completed, we welcome you to contribute your evaluation to the open-source Evals repo for use by the broader ai community.
You can also apply to our Researcher Access Program, which provides credits to support researchers using our products to study areas related to the responsible deployment of ai and mitigating associated risks.