Discussion about this post

User's avatar
Katalina Hernández's avatar

Planning to reference this article in one of my AI Safety advocacy projects during my Fellowship with Successif. Thank you for your thoughtful work on this bulletin!

I’ve often observed a major disconnect between AI safety research and the more corporate-facing “Responsible AI” discourse. I primarily work in Spain and the UK, and in both contexts, I’ve noticed that local governments tend to consult nearby big tech companies (such as my own, or firms like Accenture and IBM) when developing AI policies or making decisions around AI deployment.

These companies often serve as informal policy advisors, but the individuals consulted typically come from the Responsible AI space rather than AI safety. While I’m starting to see early signs of convergence (e.g., some of us in industry are bringing safety principles into Responsible AI practices), this remains the exception rather than the norm.

The result is that local governments only receive a partial view of the risks and challenges involved. Many corporate stakeholders in Responsible AI aren’t familiar with foundational safety concepts like alignment (inner/outer), deceptive alignment, or mechanistic interpretability. Their focus tends to be on output-level auditing (fairness, bias, explainability) without engaging with the deeper issues surrounding the behavior of foundation models.

That’s why I think AI safety policy efforts should aim not only to educate policymakers, but also corporate stakeholders. In Europe especially, there’s a critical need for basic AI safety literacy at both levels. And if more people in influential corporate roles were aware of these safety concerns, they might be more willing to advocate for, and fund, relevant research. After all, big tech still holds a lot of the leverage when it comes to shaping what gets prioritised.

Expand full comment
2 more comments...

No posts