news
Safe Artificial Intelligence Isn’t Enough, According to New Georgia Tech Research
Primary tabs
Artificial intelligence (AI) loves to cheat. When matched against a chess bot, an OpenAI model preferred hacking into its opponent’s system to winning the game fairly, according to a recent study.
While chess doesn’t have moral stakes, more serious ethical issues could arise in everything from medicine to self-driving cars as AI becomes even more pervasive. So, what does it mean for AI to be safe?
“No one is saying developing safe AI will be easy, but we need to make sure we cover as many ethical concerns as possible,” said Tyler Cook, a research affiliate at the Jimmy and Rosalynn Carter School of Public Policy at Georgia Tech and assistant program director of the Center for AI Learning at Emory University. “Humans also care about being treated fairly. We care about not being deceived. We should aim for much more than safety.”
AI is too complex for simple guardrails, Cook argues in a recent Science and Engineering Ethics paper. But AI still needs to be limited and incorporated with human values of fairness, honesty, and transparency so it doesn’t make ethically dubious decisions.
AI is not just a problem to manage. It’s a technology whose impact depends on the values we choose to build in it, Cook claims. Developers must think carefully about the world their systems will shape. AI shouldn’t make our world, but instead integrate into it.
Safe vs. Autonomous AI
Some computer scientists would say “safe” AI, or AI that doesn’t cause harm, is the answer. But AI is not a simple machine like a lawnmower that needs just a blade guard to prevent harm.
Establishing AI safety is more complex than adding protective features. Being prudent with how much autonomy AI gets is also paramount.
“We don't want AI systems deciding that they don't want to pursue fairness anymore,” Cook said. “We don't want AI to be autonomous with respect to its ethical goals or values.”
Such ethical autonomy could lead to unpredictable or undesirable outcomes. Consider algorithmic bias: Human biases, combined with machine automation, can lead to unequal consequences. An AI mortgage lender could favor certain applicant demographics over others, for example.
Cook posits there is a middle ground between merely safe AI and autonomous ethical AI — “end-constrained ethical AI.”
“As designers of AI systems, computer scientists should choose what we want the AI to prioritize: fairness, honesty, transparency,” Cook said. “That's why I use the language of constraint. We're constraining the AI’s values so they can actually benefit society.”
End‑constrained ethical AI asks designers to set those boundaries intentionally, not as an afterthought. And if developers take that responsibility seriously, AI doesn’t have to reinvent our world — it can strengthen the one we already have.
"A Case for End-Constrained Ethical Artificial Intelligence." Science and Engineering Ethics 32.7 (2026).
DOI: 10.1007/s11948-025-00577-6
Status
- Workflow status: Published
- Created by: Tess Malone
- Created: 02/25/2026
- Modified By: Tess Malone
- Modified: 02/25/2026
Categories
Keywords
User Data