Categories: Technology

OpenAI made a system better than people in a retro game

This time, OpenAI created a system better than people in the retro game Montezuma Revenge, thanks to the newly developed RND method.

OpenAI is a San Francisco-based non-profit organization that has garnered the support of many investors, including Elon Musk. The scientists there mainly deal with the development of artificial intelligence, the achievements of which are revealed mainly in computer games. They already have a phenomenally successful team in DOTA 2, or a fratricidal team in Starcraft 2. In the latest project, they made sure to beat people in the old-school Montezuma Revenge game with Atari 2600. It is not difficult to guess that OpenAI was successful here, too, which, however, required a new approach to machine learning.

This is the RND (Random Network Distillation) method, or colloquially translated “random network distillation”. This can drive any machine learning algorithm that, for example, is based on a system of rewards and punishments that guide the actions of AI agents. Contrary to the traditional method, RND introduces an additional reward, based on predicting the result of a fixed and randomly initiated neural network in the next state. Thanks to this, the AI ​​was encouraged to explore areas of the map that were not necessary to complete the game. Result? Discover all hidden rooms on the longest attempt and score an average of 10,000 points.

The goal of RND is therefore the development of artificial intelligence, which, instead of focusing on achieving the goal, also tries to do it in the best possible style. This is evident after repeated attempts at Super Mario (discovering 11 levels and hidden locations), although it is not the best when playing Pong. Then the AI ​​decides to extend the game indefinitely.

Also read: People, however, are superior to OpenAI Five in DOTA 2

Source: VentureBeat

Miners Hashrate

Recent Posts

Mining RTX 3070 at NiceHash: Overclocking, tuning, profitability, consumption

Mining on RTX 3070. Overclocking, tuning, profitability, consumption: If you are interested in finding more…

6 months ago

Mining GTX 1660, 1660 Ti, 1660 Super: Overclocking, settings, consumption

Mining with GTX 1660, 1660 Ti, 1660 Super. Overclocking, settings, consumption, profitability, comparisons - If…

6 months ago

Mining RTX 2070 and 2070 Super: Overclocking, profitability, consumption

Mining with RTX 2070 and 2070 Super. Overclocking, profitability, consumption, comparison What the RTX 2070…

6 months ago

Mining with RTX 3060, 3060 Ti. Limitations, overclocking, settings, consumption

Mining with RTX 3060, 3060 Ti. Limitations, overclocking, settings, consumption, profitability, comparison Let's look at…

6 months ago

Alphacool Eisblock Aurora Acryl GPX-A Sapphire – test: 2.8 GHz++ are not an issue

Alphacool Eisblock Aurora Acryl GPX-A (2022) with Sapphire Radeon RX 6950 XT Nitro+ Pure in…

6 months ago

Corporate Crypto Strategies 4.0: Leading with Bitcoin Expertise

In the ever-evolving landscape of business strategy, Bitcoin has emerged as a pivotal asset. With…

6 months ago

This website uses cookies.


Notice: ob_end_flush(): failed to send buffer of zlib output compression (1) in /home/gamefeve/bitcoinminershashrate.com/wp-includes/functions.php on line 5420

Notice: ob_end_flush(): failed to send buffer of zlib output compression (1) in /home/gamefeve/bitcoinminershashrate.com/wp-includes/functions.php on line 5420