5 Mega AI threats to humanity 2024 | Is AI So Dangerous? | How It Could Destroy

 


Mikko Hyppönen has spent decades on the frontlines of the fight against malware. The 54-year-old has vanquished some of the world’s most destructive computer worms, tracked down the creators of the first-ever PC virus, and sold his own software since he was a teenager in Helsinki.

In the intervening years, he’s earned Vanity Fair profiles, spots on Foreign Policy’s Top 100 Global Thinkers, and the role of Chief Research Officer at With Secure— the largest cybersecurity firm in the Nordics. He is also the curator of the online Malware Museum. Yet all the history in his archives could be overshadowed by the new era in tech: the age of artificial intelligence. “AI changes everything,” and “The AI revolution is going to be bigger than the internet revolution.”

1. Deep fakes

Researchers have long described deep fakes as the most alarming use of AI for crime, but the synthetic media still hasn’t fulfilled their predictions. Not yet, anyway.

In recent months, however, their fears have started to materialize. Deep fake fraud attempts are up 3,000% in 2023, according to research from Onfido, an ID verification unicorn based in London.

In the world of information warfare, fabricated videos are also advancing. The crude deep fakes of Ukrainian President Volodymyr Zelenskyy from the early days of Russia’s full-scale invasion have lately been superseded by sophisticated media manipulations.

Deep fakes are also now emerging in simple cons. The most notable example was discovered in October, when a video appeared on TikTok that claimed to show Beast offering new iPhones for just $2.

Still, financial scams that harness convincing deep fakes remain rare. As deep fakes become more refined, accessible, and affordable, their scale could expand rapidly.

“It’s not happening in massive scale just yet, but it’s going be a problem in a very short time,”

To reduce the risk, he suggests an old-fashioned defense: safe words.

Picture a video call with colleagues or family members. If someone demands sensitive information, such as a cash transfer or confidential document, you would request the safe word before fulfilling the request.

“Right now, it sounds a little bit ridiculous, but we should be doing it nevertheless,”

“Setting up a safe word right now is a very cheap insurance against when this starts happening in large scale. That’s what we should be taking away right now for 2024.”

2. Deep scams

Despite resembling deep fakes in name, deep scams don’t necessarily involve manipulated media. In their case, the “deep” refers to the massive scale of the scam. This is reached through automation, which can expand the targets from a handful to endless.

The techniques can turbocharge all manner of scams. Investment scams, phishing scams, property scams, ticket scams, romance scams… wherever there’s manual work, there’s room for automation.

Remember the Tinder Swindler? The conman stole an estimated $10 million from women he met online. Imagine if he had been equipped with large language models (LLMs) to disseminate his lies, image generators to add apparent photographic evidence, and language converters to translate his messages. The pool of potential victims would be enormous.

“You could be scamming 10,000 victims at the same time instead of three or four,”

Airbnb scammers can also reap the benefits. Currently, they typically use stolen images from real listings to convince holidaymakers to make a booking. It’s a laborious process that can be foiled with a reverse image search. With GenAI, those barriers no longer exist.

“With Stable Diffusion, DALL-E, and Midjourney you can just generate unlimited amounts of completely plausible Airbnb’s which no one will be able to find.”

3. LLM-enabled malware

AI is already writing malware. Hyppönen’s team has discovered three worms that launch LLMs to rewrite code every time the malware replicates. None have been found in real networks yet, but they’ve been published in GitHub — and they work.

Using an OpenAI API, the worms harness GPT to generate different code for every target it infects. That makes them difficult to detect. OpenAI can, however, blacklist the behavior of the malware.

“This is doable with the most powerful code-writing generative AI systems because they are closed source,”

“If you could download the whole large language model, then you could run it locally or on your own server. They couldn’t blacklist you anymore. This is the benefit of closed-source generative AI systems.”

The benefit also applies to image generator algorithms. Offer open access to the code and watch your restrictions on violence, porn, and deception get dismantled.

With that in mind, it’s unsurprising that OpenAI is more closed than its name suggests. Well, that and all the income they would lose to copycat developers, of course.

4. Discovery of zero-days

Another emerging concern involves zero-day exploits, which are discovered by attackers before developers have created a solution to the problem. AI can detect these threats — but it can also create them.

“It’s great when you can use an AI assistant to find zero-days in your code so you can fix them,”.

“And it’s awful when someone else is using AI to find zero-days in your code so they can exploit you.

“We’re not exactly there yet, but I believe that this will be a reality — and probably a reality in the shorter term.”

A student working at With Secure has already demonstrated the threat. In a thesis assignment, they were given regular user rights to access the command line on a Windows 11 computer. The student then fully automated the process of scanning for vulnerabilities to become the local admin. With Secure decided to classify the thesis.

5. Automated malware

With Secure has baked automation into its defenses for decades. That gives the company an edge over attackers, who still largely rely on manual operations. For criminals, there’s a clear way to close the gap: fully automated malware campaigns.

“That would turn the game into good AI versus bad AI,”.

That game is set to start soon. When it does, the results could be alarming. So alarming that Hyppönen ranks fully automated malware as the number one security threat for 2024. Yet lurking around the corner is an even bigger threat.

Comments

Popular posts from this blog

Elon Musk, Alex Jones, Andrew Tate In WILD Conversation | HEATED Debate | Full Video (New) 🎃🔊

Google's GEMINI Sends Shockwaves (Outperforming GPT-4) Comprehensive Analysis + Technical Report🤖🚀