News

OpenAI charter puts safety, standards, and transparency first

2 min read

OpenAI, the non-profit that promotes the development of artificial intelligence, has released a charter. In it, the organization outlines the core principles it believes should govern the development and management of artificial intelligence. The OpenAI charter represents an important step in initiating a broader discussion around the ethical considerations of artificial intelligence.

Revealed in a short blog post, the organization explains that the OpenAI charter is a summation of the development of its strategy over the last two years. Its mission remains central to the charter, however: ensuring that the development of artificial intelligence benefits all of humanity. 

What’s inside the OpenAI charter?

The charter is then broken down into 4 other areas.

  • Broadly-distributed benefits – OpenAI claims its primary duty is to humanity
  • Long-term safety
  • Technical leadership – OpenAI places itself at the cutting edge of the technology that will drive AI forward
  • Cooperative orientation – working with policy-makers and institutions

Core concerns the OpenAI charter aims to address

A number of core concerns lie at the heart of the charter. One of the most prominent is what OpenAI see as the competitive race to create AGI “without time for adequate safety precautions”. It’s because of this that OpenAI seeks cooperation with “other research and policy institutions” – essentially ensuring that AI doesn’t become a secretive corporate arms race. Clearly, for OpenAI, transparency will be key to creating artificial intelligence that is ‘safe’.

OpenAI also claims it will publish its most recent AI research. But perhaps even more interestingly, the charter goes on to say that “we expect that safety and security concerns will reduce our traditional publishing in the future, while increasing the importance of sharing safety, policy, and standards research.” There appears to be a tacit recognition of a tension between innovation of AI and the ethics around such innovations.

A question nevertheless remains over how easy it is for an organization to be at the cutting-edge of AI technology, while taking part in conversations around safety and ethics. As the last decade of technical development has proved, innovation and standards sometimes seem to be diametrically opposed, rather than in support of one another. This might be important in moving beyond that apparent opposition.

Read next:

Richard Gall

Co-editor of the Packt Hub. Interested in politics, tech culture, and how software and business are changing each other.

Share
Published by
Richard Gall

Recent Posts

Top life hacks for prepping for your IT certification exam

I remember deciding to pursue my first IT certification, the CompTIA A+. I had signed…

3 years ago

Learn Transformers for Natural Language Processing with Denis Rothman

Key takeaways The transformer architecture has proved to be revolutionary in outperforming the classical RNN…

3 years ago

Learning Essential Linux Commands for Navigating the Shell Effectively

Once we learn how to deploy an Ubuntu server, how to manage users, and how…

3 years ago

Clean Coding in Python with Mariano Anaya

Key-takeaways:   Clean code isn’t just a nice thing to have or a luxury in software projects; it's a necessity. If we…

3 years ago

Exploring Forms in Angular – types, benefits and differences   

While developing a web application, or setting dynamic pages and meta tags we need to deal with…

3 years ago

Gain Practical Expertise with the Latest Edition of Software Architecture with C# 9 and .NET 5

Software architecture is one of the most discussed topics in the software industry today, and…

3 years ago