![]() A follow-up to Roko's previous posts, it stated that an otherwise benevolent AI system that arises in the future might pre-commit to punish all those who heard of the AI before it came to existence, but failed to work tirelessly to bring it into existence. On 23 July 2010, LessWrong user Roko posted a thought experiment to the site, titled "Solutions to the Altruist's burden: the Quantum Billionaire Trick". In the story, and several of Langford's follow-ups to it, a basilisk is an image that has malevolent effects on the human mind, forcing it to think thoughts the human mind is incapable of thinking and instantly killing the viewer. It tells the story of a man named Robbo who paints a so-called "basilisk" on a wall as a terrorist act. ![]() The concept of the basilisk in science fiction was also popularized by David Langford's 1988 short story " BLIT". The thought experiment's name references the mythical basilisk, a creature which causes death to those that look into its eyes i.e., thinking about the AI. Yudkowsky had popularized the concept of friendly artificial intelligence, and originated the theories of coherent extrapolated volition (CEV) and timeless decision theory (TDT) in papers published in his own Machine Intelligence Research Institute. The LessWrong forum was created in 2009 by artificial intelligence theorist Eliezer Yudkowsky. In the field of artificial intelligence, Roko's basilisk has become notable as an example that raises the question of how to create an AI which is simultaneously moral and intelligent. It is also regarded as a modern version of Pascal's wager. Even after the post's discreditation, it is still used as an example of principles such as Bayesian probability and implicit religion. However, these reports were later dismissed as being exaggerations or inconsequential, and the theory itself was dismissed as nonsense, including by Yudkowsky himself. ![]() This led to discussion of the basilisk on the site to be banned for five years. While the theory was initially dismissed as nothing but conjecture or speculation by many LessWrong users, LessWrong co-founder Eliezer Yudkowsky reported users who described symptoms such as nightmares and mental breakdowns upon reading the theory, due to its stipulation that knowing about the theory and its basilisk made one vulnerable to the basilisk itself. ![]() The thought experiment's name derives from the poster of the article (Roko) and the basilisk, a mythical creature capable of destroying enemies with its stare. It originated in a 2010 post at discussion board LessWrong, a technical forum focused on analytical rational enquiry. Roko's basilisk is a thought experiment which states that an otherwise benevolent artificial superintelligence (AI) in the future would be incentivized to create a virtual reality simulation to torture anyone who knew of its potential existence but did not directly contribute to its advancement or development, in order to incentivise said advancement. ![]()
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |