
Woman says she was forced to change her name after Apple's latest system update
A new update has meant that Apple users can now say just "Siri" instead of "Hey Siri", but this modification has proven troublesome for one woman who has had to change her name as a result. Siri Price, a 26-year-old personal trainer from Edinburgh is used to having the same name as Apple's voice-activated digital assistant. While Apple's Siri stands for “Speech Interpretation and Recognition Interface," the name Siri is also Old Norse and apparently means “beautiful woman who leads you to victory". Since sharing the same name, the old version of the voice assistant would be triggered by saying the words "Hey Siri". “I work in a gym with a lot of people around so everyone learned quite quickly not to say ‘Hey’ when they greet me or there would be a lot of bleeping going on," she told The Sun. “It was annoying but manageable." But now this command is just "Siri," for the voice to be activated so it means people cannot say Siri's name without causing the voice feature to activate on their phones. As you can imagine, Siri is pretty annoyed with this new update. "Now people can’t even say my name. I’m absolutely fuming," she said. "My workmates had to sit down and think of a workaround because people’s phones have been going off non-stop." Consequently, it means Siri has had to come up with a different name for herself - Siz. “I’m sure Apple could have picked something else instead — there’s plenty of people called Siri and their lives have become just unbearable because of this," Siz added. Although Siz is not alone as those with the name Alexa possibly have similar feelings as their name is also the name of Amazon's virtual assistant. When this happened, it resulted in the name dramatically declining in popularity - perhaps due to parents releasing their child may have the same problem as Siri's. Something for tech companies to keep in mind when they're naming their voice assistants... Sign up to our free Indy100 weekly newsletter Have your say in our news democracy. Click the upvote icon at the top of the page to help raise this article through the indy100 rankings.
2023-10-06 21:22

When Seconds Count, Ultrasound-Based Location Technology Helps Protect Healthcare Workers
GREENWICH, Conn.--(BUSINESS WIRE)--Jun 8, 2023--
2023-06-09 02:50

Starfield was almost set in Fallout universe with epic Easter egg
'Starfield' was almost set in the same universe as 'Fallout'.
2023-09-13 20:24

Skyflow Radically Simplifies Data Residency
PALO ALTO, Calif.--(BUSINESS WIRE)--May 26, 2023--
2023-05-26 23:22

Avanci Broadcast welcomes ADTH as a Licensee
DALLAS--(BUSINESS WIRE)--Aug 23, 2023--
2023-08-23 21:20

Voices: The real reason companies are warning that AI is as bad as nuclear war
They are 22 words that could terrify those who read them, as brutal in their simplicity as they are general in their meaning: “Mitigating the risk of extinction from AI should be a global priority alongside other societal-scale risks such as pandemics and nuclear war.” That is the statement from San Francisco-based non-profit the Center for AI Safety, and signed by chief executives from Google Deepmind and ChatGPT creators OpenAI, along with other major figures in artificial intelligence research. The fact that the statement has been signed by so many leading AI researchers and companies means that it should be heeded. But it also means that it should be robustly examined: why are they saying this, and why now? The answer might take some of the terror away (though not all of it). Writing a statement like this functions as something like a reverse marketing campaign: our products are so powerful and so new, it says, that they could wipe out the world. Most tech products just promise to change our lives; these ones could end it. And so what looks like a statement about danger is also one that highlights just how much Google, OpenAI and more think they have to offer. Warning that AI could be as terrible as pandemics also has the peculiar effect of making artificial intelligence's dangers seem as if they just arise naturally in the world, like the mutation of a virus. But every dangerous AI is the product of intentional choices by its developers – and in most cases, from the companies that have signed the new statement. Who is the statement for? Who are these companies talking to? After all, they are the ones who are creating the products that might extinguish life on Earth. It reads a little like being hectored by a burglar about your house’s locks not being good enough. None of this is to say that the warning is untrue, or shouldn't be heeded; the danger is very real indeed. But it does mean that we should ask a few more questions of those warning us about it, especially when they are conveniently the companies that created this ostensibly apocalyptic tech in the first place. AI doesn't feel so world-destroying yet. The statement's doomy words might come as some surprise to those who have used the more accessible AI systems, such as ChatGPT. Conversations with that chatbot and others can be funny, surprising, delightful and sometimes scary – but it's hard to see how what is mostly prattle and babble from a smart but stupid chatbot could destroy the world. They also might come as a surprise to those who have read about the many, very important ways that AI is already being used to help save us, not kill us. Only last week, scientists announced that they had used artificial intelligence to find new antibiotics that could kill off superbugs, and that is just the beginning. By focusing on the "risk of extinction" and the "societal-scale risk" posed by AI, however, its proponents are able to shift the focus away from both the weaknesses of actually existing AI and the ethical questions that surround it. The intensity of the statement, the reference to nuclear war and pandemics, make it feel like we are at a point equivalent with cowering in our bomb shelters or in lockdown. They say there are no atheists in foxholes; we might also say there are no ethicists in fallout shelters. If AI is akin to nuclear war, though, we are closer to the formation of the Manhattan Project than we are to the Cold War. We don’t need to be hunkering down as if the danger is here and there is nothing we can do about it but “mitigate it”. There's still time to decide what this technology looks like, how powerful it is and who will be at the sharp end of that power. Statements like this are a reflection of the fact that the systems we have today are a long way from those that we might have tomorrow: the work going on at the companies who warned us about these issues is vast, and could be much more transformative than chatting with a robot. It is all happening in secret, and shrouded in both mystery and marketing buzz, but what we can discern is that we might only be a few years away from systems that are both more powerful and more sinister. Already, the world is struggling to differentiate between fake images and real ones; soon, developments in AI could make it very difficult to find the difference between fake people and real ones. At least according to some in the industry, AI is set to develop at such a pace that it might only be a few years before those warnings are less abstractly worrying and more concretely terrifying. The statement is correct in identifying those risks, and urging work to avoid them. But it is more than a little helpful to the companies that signed it in making those risks seem inevitable and naturally occurring, as if they are not choosing to build and profit from the technology they are so worried about. It is those companies, not artificial intelligence, that have the power to decide what that future looks like – and whether it will include our "extinction". Read More Opinion: Age gap relationships might seem wrong, but they work. Trust me Hands up if you trust Boris Johnson | Tom Peck Boris’s ‘ratty rat’ rage against Sunak could bring the Tories down | John Rentoul Opinion: Age gap relationships might seem wrong, but they work. Trust me Hands up if you trust Boris Johnson | Tom Peck Boris’s ‘ratty rat’ rage against Sunak could bring the Tories down | John Rentoul
2023-05-31 18:58

Baidu Creates $140 Million Fund to Back ChatGPT-Like Startups
Baidu Inc. has set aside 1 billion yuan ($140 million) to fund Chinese startups that explore generative AI,
2023-05-31 14:17

XY Retail Unveils Groundbreaking Distributed OMS - XY DOM: The Ultimate Solution for Order Management and Fulfillment
NEW YORK--(BUSINESS WIRE)--May 24, 2023--
2023-05-24 21:20

Save 88% on lifetime access to unlimited premium web hosting
TL;DR: A lifetime subscription to iBrave Cloud Web Hosting is on sale for £78.32, saving
2023-08-25 12:16

LGBTQ youth of color feel safest on their TikTok FYP, report finds
For LGBTQ youth, social media is a double-edged sword: a digital public space that acts
2023-07-20 01:54

NBA 2K24 Sept. 14 Update 1.2 Patch Notes: Full List of Changes
The NBA 2K24 Sept. 14 patch notes for update 1.2 delivered numerous changes to MyCAREER mode, including quest and progression updates.
2023-09-14 23:47

'Fallout 3' and 'Oblivion' remasters show up in leaked Microsoft documents
Microsoft's messy merger with Activision Blizzard is the gift that keeps on giving. While the
2023-09-19 22:54
You Might Like...

Pickle Pain: Rising Pickleball Injuries Could Result in 67,000 ER Visits This Year

Heat Stress Deaths Show Europe Isn’t Ready for Climate Change

EU hits tech titans with tougher market restraints

Telegram calls Brazil disinformation law 'attack on democracy'

Amazon is bringing ads to Prime Video and will charge you to avoid them

OTTO Lifter Wins Gold Stevie® Award in 2023 International Business Awards®

A Chatbot That Won't Take Bribes for Giving Advice Is a Hit in India

US intelligence agencies buy Americans' personal data, new report says