Meet Dan Hendrycks, the Researcher Behind the Nonprofit Center for AI Safety
Dan Hendrycks, a renowned machine learning expert, has been making waves in the AI community with his work on developing benchmarks and leading public advocacy. As the founder of the nonprofit Center for AI Safety, Hendrycks has been tirelessly working to prevent AI-driven global risks. In addition to his work with the Center, Hendrycks also serves as a safety advisor to companies like xAI and Scale AI, a role that has kept him particularly busy in recent months.
Despite the importance of his work, Hendrycks’ advisory roles come with surprisingly modest compensation. He earns just $1 a year for his work with Elon Musk’s xAI, which he joined in 2023, and $12 annually from data annotation firm Scale AI, which he began advising last year. However, the impact of his work cannot be measured in dollars alone. Hendrycks’ expertise has been instrumental in shaping the development of AI systems, and his collaboration with tech giants like Musk has been particularly fruitful.
Working with Elon Musk and Tackling AI-Driven Risks
Hendrycks’ work with xAI has been focused on assessing and mitigating risks associated with AI systems, from bioweapons to cyber threats. He has been working closely with Musk to ensure that xAI’s systems remain below specific danger thresholds. “The objective afterwards is to continually try to drive down that threshold to make it more and more strict so that there’s less and less of these risks,” Hendrycks explained in a recent interview at TechCrunch Disrupt 2025 in San Francisco.
One notable example of Hendrycks’ work with xAI is the development of more robust checks and time delays before releasing updates. This came after xAI’s Grok chatbot generated antisemitic remarks earlier this year, sparking controversy and highlighting the need for more stringent safety protocols. “I think that’s a very positive development in view of the event,” Hendrycks said, emphasizing the importance of learning from mistakes and implementing improvements.
Measuring Political Bias and Ensuring AI Safety
Hendrycks’ work with xAI and Scale AI also involves measuring political bias in AI systems. This is a critical aspect of ensuring that AI systems remain neutral and do not perpetuate harmful ideologies. By tracking “covert activism” and examining whether a system presents facts in an overly positive or negative light, Hendrycks aims to optimize AI systems for political neutrality. As he noted, “If you target that, optimize against that, then you get a system that is substantially more politically involved.”
Musk has also emphasized the importance of political neutrality in AI systems, branding Grok as a less “woke” alternative to competitors. Hendrycks’ collaboration with Musk on this front has been instrumental in shaping the development of xAI’s systems, and his expertise has been invaluable in navigating the complex landscape of AI safety.
A Collaboration with Tech Giants
Hendrycks’ work with Musk is not his only notable collaboration. Earlier this year, he co-authored a paper with former Google CEO Eric Schmidt and Scale AI founder Alexandr Wang, urging the U.S. to proceed cautiously with advanced AI development. The paper warned of the risks of an unchecked AI race, including the potential for rogue bioweapon creation and cyberattacks. As Hendrycks noted, “Those are sitting ducks,” highlighting the vulnerability of critical infrastructure to cyber threats.
Despite the challenges ahead, Hendrycks remains optimistic about the potential for AI to drive positive change. With his work at the Center for AI Safety and his collaboration with tech giants like Musk, Hendrycks is helping to shape the future of AI and ensure that its development is aligned with human values. As he emphasized, “I think he’s a very enjoyable person to work with,” highlighting the importance of collaboration and mutual respect in driving progress in the field of AI safety.
Photo: Guerin Blask for The New York Times/Courtesy Center for AI Safety
Image Source: observer.com

