![](/rp/kFAqShRrnkQMbH6NYLBYoJ3lq9s.png)
John Schulman's Homepage
John Schulman's Homepage I am currently a researcher at Anthropic, where I’m working on aligning large language models; some of my interests include scalable oversight and developing better written specifications of model behavior (like OpenAI’s Model Spec , Constitutional AI ).
Selected Publications
John Schulman, Jonathan Ho, Cameron Lee, and Pieter Abbeel International Symposium on Robotics Research (ISRR), 2013 Paper / Videos
An Opinionated Guide to ML Research - joschu.net
Jan 24, 2020 · John Schulman's Homepage. An Opinionated Guide to ML Research. Posted on 2020/01/24. ← back to blog index. I originally wrote this guide in back in December 2017 for the OpenAI Fellows program. In this essay, I provide some advice to up-and-coming researchers in machine learning (ML), based on my experience doing research and advising others ...
john schulman Summer, 2016 A dissertation submitted in partial satisfaction of the requirements for the degree of Doctor of Philosophy in Computer Science in the Graduate Division of the University of California, Berkeley committee: Pieter Abbeel, Chair …
Presentations
John Schulman's Homepage. Presentations. Some recent talks: 2024 Talk about OpenAI Model Spec at Scale conference; 2023 ICML talk on proxy objectives; 2023 Berkeley talk on truthfulness; Older slides and video presentations: TR35 Award Talk at EmTech 2018 in Cambridge, MA. Video / Slides (PDF)
Code - joschu.net
Code. GitHub profile. Highlighted projects developed by my collaborators and me: Procgen Benchmark (2019): GitHub / blog post.; Gym Retro (2018): GitHub / blog post on dataset / contest; OpenAI Baselines (2016): GitHub / original post (DQN) / ACKTR + A2C / PPO; OpenAI Gym (2016): homepage / GitHub / blog post / article on NVIDIA blog; Computation Graph Toolkit (2015): GitHub / docs.
Approximating KL Divergence
Mar 7, 2020 · Note that the bias of k2 is incredibly low here: it’s 0.2%. Now let’s try for a larger true KL divergence. p=N(1,1) gives us a true KL divergence of 0.5.
Blog Index - joschu.net
Jan 24, 2020 · John Schulman's Homepage. Blog Index. Sending Samples Without Bits-Back (2020/03/08) Approximating KL Divergence (2020/03/07) An Opinionated Guide to ML Research (2020/01/24)
Awards - joschu.net
John Schulman's Homepage. Awards [2018] MIT Technology Review's 35 Innovators Under 35. [2016] C.V. Ramamoorthy Distinguished Research Award [2013] Best Vision Paper, awarded at ICRA 2013 ...
John Schulman December 9th, 2016. Outline Approaching New Problems Ongoing Development and Tuning General Tuning Strategies for RL Policy Gradient Strategies Q-Learning Strategies Miscellaneous Advice. Approaching New Problems. New Algorithm? Use Small Test Problems I Run experiments quickly