China defends ban on US chipmaker Micron, accuses Washington of 'economic coercion'
The Chinese government has defended its ban on products from U.S. memory chipmaker Micron Technology Inc. in some computer systems after Washington expressed concern
2023-05-24 18:56
Explainer-What happens if the US Supreme Court bans affirmative action?
By Joseph Ax The U.S. Supreme Court is set to rule by the end of June on whether
2023-05-24 18:30
Factbox-Who's who in the Supreme Court's affirmative action cases
By Gabriella Borter The U.S. Supreme Court is due to rule by the end of June on the
2023-05-24 18:21
Asmongold reacts to 'king of toxic masculinity' Andrew Tate's anime girl tweets: 'I wonder why he's doing it'
Popular Twitch streamer Zack ‘Asmongold’ responded to Andrew Tate's recent tweets, which have come as a surprise to his followers
2023-05-24 17:50
Nvidia’s $400 Billion Rally Is Set for Reality Check
The rally that’s added roughly $400 billion to Nvidia Corp.’s market value this year is about to be
2023-05-24 17:48
Ready for a digital euro? At 25, European Central Bank preps for future of money
Proposals for a digital European currency are taking shape
2023-05-24 17:23
Japan's Rakuten set to raise $2.18 billion through share issue
TOKYO (Reuters) -Japanese e-commerce and fintech conglomerate Rakuten Group Inc will raise up to 294.2 billion yen ($2.18 billion) by
2023-05-24 17:20
China was reducing Micron chip purchases years before ban
By Eduardo Baptista BEIJING In the years before China declared U.S. firm Micron Technology's products a national security
2023-05-24 17:18
What is superintelligence? How AI could replace humans as the dominant lifeform on Earth
In the ‘Unfinished Fable of the Sparrows’, a group of small birds come up with a plan to capture an owl egg and raise the chick as their servant. “How easy life would be,” they say, if the owl could work for them, and they could live a life of leisure. Despite warnings from members of their flock that they should first figure out how to tame an owl before they raise one, the sparrows devote all their efforts to capturing an egg. This tale, as its title suggests, does not have an ending. Its author, Swedish philosopher Nick Bostrom, deliberately left it open-ended as he believes that humanity is currently in the egg hunting phase when it comes to superhuman AI. In his seminal work on artificial intelligence, titled Superintelligence: Paths, Dangers, Strategies, the Oxford University professor posits that AI may well destroy us if we are not sufficiently prepared. Superintelligence, which he describes as an artificial intelligence that “greatly exceeds the cognitive performance of humans in virtually all domains of interest”, may be a lot closer than many realise, with AI experts and leading industry figures warning that it may be just a few years away. On Monday, the creator of ChatGPT echoed Professor Bostrom’s 2014 book by warning that the seemingly exponential progress of AI technology in recent years means that the imminent arrival of superintelligence is inevitable – and we need to start preparing for it before it’s too late. OpenAI boss Sam Altman, whose company’s AI chatbot is the fastest growing app in history, has previously described Professor Bostrom’s book as “the best thing I’ve seen on this topic”. Just a year after reading it, Mr Altman co-founded OpenAI alongside other similarly worried tech leaders like Elon Musk and Ilya Sutskever in order to better understand and mitigate against the risks of advanced artificial intelligence. Initially launched as a non-profit, OpenAI has since transformed into arguably the leading private AI firm – and potentially the closest to achieving superintelligence. Mr Altman believes superintelligence has the potential to not only offer us a life of leisure by doing all the majority of our labour, but also holds the key to curing diseases, eliminate suffering and transforming humanity into an interstellar species. Any attempts to block its progress, he wrote this week, would be “unintuitively risky” and would require “something like a global surveillance regime” that would be virtually impossible to implement. It is already difficult to understand what is going on inside the ‘mind’ of AI tools currently available, but once superintelligence is achieved, even its actions may become incomprehensible. It could make discoveries that we would be incapable of understanding, or take decisions that make no sense to us. The biological and evolutionary limitations of brains made of organic matter mean we may need some form of brain-computer interface in order to keep up. Being unable to compete with AI in this new technological era, Professor Bostrom warns, could see humanity replaced as the dominant lifeform on Earth. The superintelligence may then see us as superfluous to its own goals. If this happens, and some form of AI has figured out how to hijack all the utilities and technology we rely upon – or even the nuclear weapons we possess – then it would not take long for AI to wipe us off the face of the planet. A more benign, but similarly bleak, scenario is that the gulf in intelligence between us and the AI will mean it views us in the same way we view animals. In a 2015 conversation between Mr Musk and scientist Neil deGrasse Tyson, they theorised that AI will treat us like a pet labrador. “They’ll domesticate us,” Professor Tyson said. “They’ll keep the docile humans and get rid of the violent ones.” In an effort to prevent this outcome, Mr Musk has dedicated a portion of his immense fortune towards funding a brain chip startup called Neuralink. The device has already been tested on monkeys, allowing them to play video games with their minds, and the ultimate goal is to transform humans into a form of hybrid superintelligence. (Critics note that even if successful, the technology would similarly create a two-tiered society of the chipped, and the chipless.) Since cutting ties with OpenAI, the tech billionaire has issued several warnings about the imminent emergence of superintelligence. In March, he joined more than 1,000 researchers in calling for a moratorium on the development of powerful AI systems for at least six months. That time should then be spent researching AI safety measures, they wrote in an open letter, in order to avert disaster. It would take an improbable consensus of leading AI companies around the world, the majority of which are all profit-seeking, in order for any such pause to be impactful. And while OpenAI continues to spearhead the hunt for the owl’s egg, Mr Altman appears to have at least heeded the warnings from Professor Bostrom’s fable. In a 2016 interview with the New Yorker, he revealed that he is a doomsday prepper – specifically for an AI-driven apocalypse. “I try not to think about it too much, he said, revealing that he has “guns, gold, potassium iodide, antibiotics, batteries, water [and] gas masks” stashed away in a hideout in rural California. Not that any of that will be much use to the rest of us. Read More 10 ways AI will change the world – from curing cancer to wiping out humanity Photoshop unveils ‘extraordinary’ AI that transforms your pictures with a text prompt ChatGPT creator issues stark warning on AI ‘RIP photoshop’: New AI can alter any photo with the click of a mouse
2023-05-24 17:18
South Africa Weighs Extending Lives of Larger Coal Power Plants
South Africa is considering extending the lives of some of its biggest coal-fired power plants as the government
2023-05-24 17:17
No way to police all cryptocurrency fraud, CFTC commissioner says
By Luc Cohen NEW YORK A top U.S. regulator said on Tuesday there is no way to police
2023-05-24 16:59
Korea Space Race Heats Up as North and South Plan Launches
The two Koreas are in a space race. The North is upgrading its space center to accommodate the
2023-05-24 16:19