Throughout my first 2.5 years at OpenAI, I labored on the Robotics crew on a moonshot concept: we wished to show a single, human-like robotic hand to unravel Rubik’s dice. It was a tremendously thrilling, difficult, and emotional expertise. We solved the problem with deep reinforcement studying (RL), loopy quantities of area randomization, and no real-world coaching information. Extra importantly, we conquered the problem as a crew.
From simulation and RL coaching to imaginative and prescient notion and {hardware} firmware, we collaborated so carefully and cohesively. It was an incredible experiment and through that point, I typically considered Steve Jobs’ reality distortion field: if you imagine in one thing so strongly and carry on pushing it so persistently, by some means you may make the inconceivable doable.
Because the starting of 2021, I began main the Utilized AI Analysis crew. Managing a crew presents a distinct set of challenges and requires working model adjustments. I’m most pleased with a number of initiatives associated to language mannequin security inside Utilized AI:
- We designed and constructed a set of analysis information and duties to evaluate the tendency of pre-trained language fashions to generate hateful, sexual, or violent content material.
- We created an in depth taxonomy and constructed a robust classifier to detect unwanted content in addition to the rationale why the content material is inappropriate.
- We’re engaged on numerous strategies to make the mannequin much less prone to generate unsafe outputs.
Because the Utilized AI crew is working towards one of the best ways to deploy cutting-edge AI strategies, equivalent to massive pre-trained language fashions, we see how highly effective and helpful they’re for real-world duties. We’re additionally conscious of the significance of safely deploying the strategies, as emphasised in our Charter.