World ID offers a revolutionary aproach to verifying humanness without compromissing personal data
Users can quickly sign up for a verified World Id at an orb and use it to authenticate actions, like signing into websites, without sharing personal information.
World ID offers a revolutionary aproach to verifying humanness without compromissing personal data
AS artificial intelligence (AI) continues to evolve, it is becoming increasingly adept at replicating human behaviour online, blurring the line between genuine and automated interactions.
In the wrong hands, AI can be a potent tool for spreading misinformation, phishing scams, fraud, and data breaches – a growing concern as the world moves further into the digital realm.
Recognising these risks, tech visionaries Sam Altman of OpenAI and Alex Bania of Tools for Humanity saw the need for a privacy-focused human verification and financial network, leading to the creation of World - previously known as WorldCoin.
“Altman believes humanity needs a ‘human gate,’ where certain online activities or products are restricted to verified individuals,” explains World’s Europe managing director Fabian Bodensteiner.
But that begs the question: How can they prove that someone is human?
The answer? World ID – a digital protocol developed by World that confirms a person’s humanity or proof of humanness without sharing personal information.
A digital proof of humanness
When a person verifies their World ID via an Orb, the device takes pictures of their iris and face.
These pictures are used to make a unique iris code, a series of 1s and 0s. No two iris codes are the same, and they do they reveal direct identifiers such as name, gender, age, etc.
The code is then split into different pieces and permanently encrypted using Secure Multi-Party Computation (SMPC), which anonymises data by dividing it into multiple abstracted values (SMPC shares) and storing them in separate locations managed by different parties.
Each party only has access to the SMPC share under their control.
Bodensteiner likens World ID to a digital passport stored on a user’s mobile device via the World App, which supports World ID.
“We didn’t want to follow the standard Know Your Customer (KYC) process, which often requires users to share personal details like names and addresses,” Bodensteiner notes, highlighting their goal to help businesses reduce data collection for privacy purposes.
“Think of it like this – just as you have a national ID or driver’s licence, World ID offers an additional anonymous credential: a digital proof of humanness” he says.
The proof of humanness verification naturally limits the creation of multiple fake accounts, curbing large-scale bot attacks and ensuring content is from genuine individuals – an essential step in reducing AI-generated disinformation.
Flexible across sectors
To date, more than six million people have verified their World ID, reflecting the growing adoption of this revolutionary technology worldwide.
In Malaysia, World sees tremendous potential for expansion, driven by the country’s openness to new technologies and its diverse economy, which positions it as a strategic gateway to further extend into Asia.
“Malaysia’s openness to new technologies and its diverse economy make it a strategic gateway for further expansion into Asia,” says Bodensteiner.
Bodensteiner says the proof of humanness protocol enhances online security and accountability in the age of AI.
He adds that all World technologies, including hardware, are open-source, enabling innovation and collaboration across various sectors.
The World ID technology allows for seamless authentication across web and mobile platforms.
Its applications extend across multiple sectors, such as gaming, social networking, and marketing, where personhood verification is crucial to reducing fake accounts and ensuring genuine human interaction.
“For instance, video gaming platforms can benefit from personhood verification by allowing individuals to unlock exclusive deals, enhancing the gaming experience while keeping the ecosystem free from bot-driven accounts,” Bodensteiner explains.
Local social networks and e-commerce platforms can also leverage World ID to enhance safety and prevent fraudulent activities, such as repeated voucher redemptions, ensuring a more secure and beneficial environment for legitimate users.
Recently, Worldcoin rebranded itself as World, signaling a broader mission to build a comprehensive identity and financial network that empowers every individual in the digital economy.
The shift reflects the project’s focus on creating a global network centered on anonymous proof-of-humanness technology and inclusive financial tools.
This new identity aims to drive the mission forward with a more unified and holistic approach.
World was born from a need to address these challenges and build a system that ensures equal access to the digital economy for all, regardless of financial circumstances or location – especially as AI continues to advance.
World ID remains the bedrock of this mission, empowering people to take charge of their privacy online, paving the way for a more inclusive and equitable future for everyone.
Currently, the Orbs are situated in a few locations across the country, with plans to expand to more sites over time. Individuals who have downloaded the World App can now schedule an appointment to have their World ID verified.
3 Reasons Why AI Generated Deepfakes Are a Growing Concern The Era of Digital Deception
Sophisticated scam technology harnessing artificial intelligence is capable of deceiving even the most vigilant.
COMPUTER-GENERATED children’s voices that fool their own parents. Masks created with photos from social media deceive a system protected by face Id.
They sound like the stuff of science fiction, but these techniques are already available to criminals preying on everyday consumers.
The proliferation of scam tech has alarmed regulators, police, and people at the highest levels of the financial industry. artificial intelligence (ai) in particular is being used to “turbocharge” fraud, US Federal Trade Commission chair Lina Khan warned in June, calling for increased vigilance from law enforcement.
Even before ai broke loose and became available to anyone with an Internet connection, the world was struggling to contain an explosion in financial fraud.
In the United States alone, consumers lost almost Us$8.8bil (Rm40.9bil) last year, up 44% from 2021, despite record investment in detection and prevention. Financial crime experts at major banks, including Wells Fargo and Co and deutsche Bank ag, say the fraud boom on the horizon is one of the biggest threats facing their industry.
On top of paying the cost of fighting scams, the financial industry risks losing the faith of burned customers.
“It’s an arms race,” says James Roberts, who heads up fraud management at the Commonwealth Bank of australia, the country’s biggest bank.
“It would be a stretch to say that we’re winning.”
The history of scams is surely as old as the history of trade and business.
One of the earliest known cases, more than 2,000 years ago, involved a greek sea merchant who tried to sink his ship to get a fraudulent payout on an insurance policy.
Look back through any newspaper archive, and you’ll find countless attempts to part the gullible from their money.
But the dark economy of fraud, just like the broader economy, has periodic bursts of destabilising innovation.
new technology lowers the cost of running a scam and lets the criminal reach a larger pool of unprepared victims.
Email introduced every computer user in the world to a cast of hard-up princes who needed help rescuing their lost for tunes.
Crypto brought with it a blossoming of Ponzi schemes that spread virally over social media.
The future of fake
The ai explosion offers not only new tools but also the potential for life-changing financial losses.
and the increased sophistication and novelty of the technology mean that everyone, not just the credulous, is a potential victim.
The Covid-19 lockdowns accelerated the adoption of online banking around the world, with phones and laptops replacing face-to-face interactions at bank branches.
It’s brought advantages in lower costs and increased speed for financial firms and their customers, as well as openings for scammers.
Some of the new techniques go beyond what current off-theshelf technology can do, and it’s not always easy to tell when you’re dealing with a garden-variety fraudster or a nation-state actor.
“We are starting to see much more sophistication with respect to cybercrime,” says amy Hoganburney, general manager of cybersecurity policy and protection at Microsoft Corp.
Globally, cybercrime costs, including scams, are set to hit US$8 trillion (RM37.18 trillion) this year, outstripping the economic output of Japan, the world’s third-largest economy.
By 2025, it will reach US$10.5 trillion (RM48.8 trillion), after more than tripling in a decade, according to researcher Cybersecurity Ventures.
In the Sydney suburb of Redfern, some of Roberts’ team of more than 500 spend their days eavesdropping on cons to hear firsthand how ai is reshaping their battle.
a fake request for money from a loved one isn’t new. But now parents get calls that clone their child’s voice with ai to sound indistinguishable from the real thing.
These tricks, known as social engineering scams, tend to have the highest hit rates and generate some of the quickest returns for fraudsters.
Today, cloning a person’s voice is becoming increasingly easy.
Once a scammer downloads a short sample from an audio clip from someone’s social media or voicemail message – it can be as short as 30 seconds – they can use ai voice-synthesising tools readily available online to create the content they need.
Public social media accounts make it easy to figure out who a person’s relatives and friends are, not to mention where they live and work and other vital information.
Bank bosses stress that scammers, who run their operations like businesses, are prepared to be patient, sometimes planning attacks for months.
What fraud teams are seeing so far is only a taste of what ai will make possible, according to Rob Pope, director of new Zealand’s government cybersecurity agency, CERT nz.
He points out that ai simultaneously helps criminals increase the volume and customisation of their attacks.
“It’s a fair bet that over the next two or three years we’re going to see more ai-generated criminal attacks,” says Pope,
a former deputy commissioner in the New Zealand Police who oversaw some of the nation’s highest-profile criminal cases. “What AI does is accelerate the levels of sophistication and the ability of these bad people to pivot very quickly. AI makes it easier for them.”
To give a sense of the challenge facing banks, Roberts says right now the Commonwealth Bank of Australia is tracking about 85 million events a day through a network of surveillance tools.
That’s in a country with a population of just 26 million.
The industry hopes to fight back by educating consumers about the risks and increasing investment in defensive technology.
New software lets CBA spot when customers use their computer mouse in an unusual way during a transaction – a red flag for a possible scam.
Anything suspicious, including the destination of an order and how the purchase is processed, can alert staff in as few as 30 milliseconds, allowing them to block the transaction.
At Deutsche Bank, computer engineers have recently rebuilt their suspicious transaction detection system, called Black Forest, using the latest natural language processing models, according to Thomas Graf, a senior machine learning engineer there.
The tool looks at transaction criteria such as volume, currency, and destination and automatically learns from reams of data what patterns suggest fraud.
The model can be used on both retail and corporate transactions and has already unearthed several cases, includone ing involving organised crime, money laundering, and tax evasion.
Wells Fargo has overhauled its tech systems to counter the risk of Ai-generated videos and voices. “We train our software and our employees to be able to spot these fakes,” says Chintan Mehta, Wells Fargo’s head of digital technology. But the system needs to keep evolving to keep up with the criminals. Detecting scams, of course, costs money.
The digital dance
One problem for companies: Every time they tighten things, criminals try to find a workaround.
For example, some US banks require customers to upload a photo of an ID document when signing up for an account.
Scammers are now buying stolen data on the dark web, finding photos of their victims on social media, and 3D-printing masks to create fake IDS with the stolen information.
“And these can look like everything from what you get at a Halloween shop to an extremely lifelike silicone mask of Hollywood standards,” says Alain Meier, head of identity at Plaid, which helps banks, financial technology companies, and other businesses battle fraud with its ID verification software. Plaid analyses skin texture and translucency to make sure the person in the photo looks real.
Meier, who’s dedicated his career to detecting fraud, says the best fraudsters, those running their schemes as businesses, build scamming software and package it up to sell on the dark web.
Prices can range from US$20 (RM95) to thousands of dollars.
“For example, it could be a Chrome extension to help you bypass fingerprinting or tools that can help you generate synthetic images,” he says.
As fraud gets more sophisticated, the question of who’s responsible for losses is getting more contentious.
In the United Kingdom, for example, victims of unknown transactions – say, someone copies and uses your credit card – are legally protected against losses.
If someone tricks you into making a payment, responsibility becomes less clear.
In July, the US top court ruled that a couple who were fooled into sending money abroad couldn’t hold their bank liable simply for following their instructions.
But legislators and regulators have leeway to set other rules: The government is preparing to require banks to reimburse fraud victims when the cash is transferred via Faster Payments, a system for sending money between UK banks.
Politicians and consumer advocates in other countries are pushing for similar changes, arguing that it’s unreasonable to expect people to recognise these increasingly sophisticated scams.
Banks worry that changing the rules would simply make things easier for fraudsters.
Financial industry leaders around the world are also trying to push a share of the responsibility onto tech firms.
The fastest-growing scam category is investment fraud, often introduced to victims through search engines where scammers can easily buy sponsored advertising spots.
When would-be investors click through, they often find realistic prospectuses and other financial data. Once they transfer their money, it can take months, if not years, to realise they’ve been swindled when they try to cash in on their “investment”.
In June, a group of 30 lenders in the UK sent a letter to Prime Minister Rishi Sunak asking that tech companies contribute to refunds for victims of fraud stemming from their platforms.
The government says it’s planning new legislation and other measures to crack down on online financial scams.
The banking industry is lobbying to spread responsibility more widely, in part because costs appear to be going up. Once again, a familiar problem from economics applies in the scam economy, too.
Like pollution from a factory, new technology is creating an externality, or a cost imposed on others. In this case, there’s a heightened reach and risk for scams.
Neither banks nor consumers want to be the only ones forced to pay the price.
Chris Sheehan spent almost three decades with the country’s police force before joining National Australia Bank Ltd, where he heads investigations and fraud.
He’s added about 40 people to his team in the past year with constant investment by the bank.
When he adds up all the staff and tech costs, “it scares me how big the number is”, he says.
“I am hopeful because there are technological solutions, but you never completely solve the problem,” he says. It reminds him of his time fighting drug gangs as a cop.
Framing it as a war on drugs was “a big mistake”, he says.
“I will never phrase it in that framework – of a war on scams – because the implication is that a war is winnable,” he says. “This is not winnable.” – Bloomberg
There are encouraging signs that professionals in Malaysia are equipping themselves with a combination of hard and soft skills to enhance their employability and remain competitive. — 123rf.com
THE sooner bosses pay attention to artificial intelligence (AI) and what it can do, the better for all, including workers and the business.
As such, guidelines should be introduced by bosses in the country on how their workers should use AI in their jobs, says Malaysian Employers Federation (MEF) president Datuk Dr Syed Hussain Syed Husman.
He was responding to a proposal by the Human Resources Ministry for employers to develop their own policies and procedures for the ethical use of AI in view of its growth in Malaysia.
“This is a good suggestion as the world of work is changing and becoming more automated.
“Such a trend will continue. So the sooner we pay attention to this, the better.
“Now is the time to see how AI can help businesses and the industry, while looking at some guiding principles to help manage this,” he says.
While AI promises to smooth operations, he admits there are concerns over security, privacy, data trust, and ethics over its use.
“Businesses using AI models such as ChatGPT need to be aware that generative AI comes with its own set of risks.
“There is a need to establish rules and procedures to ensure secure implementation of AI.
“It will take time and human expertise to unlock AI’s full potential in a way that’s responsible, trustworthy and safe,” he says.
Recently, it was reported that more companies in Malaysia are exploring and integrating generative AI into their business operations.
However, not many have come up with official policies for its workers on its usage.
Some companies which have introduced guidelines have advised workers against providing personal information to AI systems to prevent any privacy issues.
While bosses are aware of the benefits AI can bring, MEF highlights the need for everyone to be trained to use it effectively.
“A lack of skilled talent and technical expertise has been a top barrier to implementing AI since its inception.
“To stay competitive in a tight labour market, companies must train their teams to use AI effectively and responsibly.
“If people don’t trust the work AI does or the data it’s built on, adoption of AI will lag and returns on investment will not be as fast as they should be,” Syed Hussain says.
In the next five years, he says bosses expect more people to be working alongside robots and smart machines specifically designed to help them to do their jobs better and more efficiently.
At the same time, jobs that can be performed through a simple search online or on ChatGPT could be at risk, says JobStreet Malaysia managing director Vic Sithasanan.
“In its place would be the prioritisation of skills to be able to query, discern, and ‘connect the dots’ or find relevance with technology that cannot replace the human touch,” he explains.
Even before Covid-19 posed a threat, job security was already on people’s minds because of automation, he adds.
“Almost every kind of worker has some level of concern.
“JobStreet’s Decoding Global Talent’s third report showed that in 2021, 46% of workers in their 20s and 41% in their 30s were already worried about technology putting them out of work.
“From media to information technology, concerns about automation are particularly high – especially among workers with repetitive jobs,” he says.
According to JobStreet, among some of the industries and jobs that may be replaced by AI – and not just ChatGPT – are translating, managing social media, umpiring sports, and jobs in libraries and call centres.“However, while many people are nervously waiting for the world to become completely reliant on AI in the next few decades, there will always be a need for human force to drive this automation.
“Though there may be many jobs that will disappear in the near future due to AI replacement, jobseekers, employees and even employers can enhance and enrich their potential to ensure that their career stays current and in demand.
“The world’s workforce may combine man and machine, but a robot-dominated world is not about to become a practical reality yet,” Sithasanan says.
While the work landscape is evolving due to technology, so are the skills in need, says LinkedIn country manager for Malaysia Rohit Kalsy.
“LinkedIn research shows that top skills required for a particular job have changed by an average of 27% since 2015, with the pace of change accelerating during the pandemic.
“At this pace, skills could change by 43% to 47% by 2025.
“Between 2021 and 2025, we would likely see three new skills in the top skills for a job,” says Rohit, who is also the company’s head of emerging markets (South-East Asia).
However, there are encouraging signs that professionals in Malaysia are equipping themselves with a combination of hard and soft skills to enhance their employability and remain competitive.
“Malaysian learners were among the 7.3 million globally who enrolled in the top 20 most popular LinkedIn learning courses between June 1, 2021, and June 30 last year.
“This is almost double from the previous year. Such figures show that more are building skills to future-proof their careers,” Rohit points out.
Last month, the Human Resources Ministry said that, with the rise of AI use, as many as 4.5 million Malaysians are likely to lose their jobs by 2030 if they do not improve their skills or attend reskilling and upskilling programmes