Site iconSite icon ForkLog

Fake on-chain sleuths prey on hack victims, an AI ‘journalist’ dupes the press, and other cybersecurity stories

Fake on-chain sleuths prey on hack victims, an AI ‘journalist’ dupes the press, and other cybersecurity stories

We have collected the most important cybersecurity news of the week.

  • Fake on-chain sleuths steal remaining funds from victims of crypto theft.
  • The media fell for an AI “journalist”.
  • Meta is suspected of accessing photo galleries without user consent.
  • Researchers hid a trojan in an AI image.

Fake on-chain sleuths are stealing what remains from crypto-hack victims

In August the FBI warned of fraudsters posing as bogus crypto law firms. Under the guise of sham asset-recovery services, the scammers stole money and personal data from clients.

According to the statement, the primary targets are victims of crypto hacks trying to reclaim stolen funds.

Law enforcement said the fraudsters used a wide array of manipulative tactics, playing on victims’ desperation and creating a false sense of security by impersonating government representatives or claiming cooperation with them. The reputations of people and organisations whose names were misused were also harmed.

When choosing help to recover cryptocurrency, the FBI advised watching for:

An AI “journalist” duped the media

Press Gazette noted that at least six outlets, including Wired and Business Insider, have in recent months removed articles from their sites. The reason, it reported, was that pieces published under the name Margot Blanchard were generated by AI.

In May Wired ran a story titled “They fell in love playing Minecraft. Then the game became their wedding venue.” It mentioned Jessica Hu, a 34-year-old clergywoman from Chicago known as a “digital officiant” on Twitch and Discord. The outlet could not verify her existence and, weeks later, removed the piece for failing to meet editorial standards.

A removed piece by a certain Margot Blanchard at Wired. Source: Wayback Machine.

According to Press Gazette, in April Business Insider published two Blanchard essays. Last week the outlet deleted them.

Reports on the removal of Business Insider and Wired pieces. Source: Press Gazette

On 21 August Wired’s leadership acknowledged the error:

“If any publication should recognize AI grifters, it’s Wired. And in fact we usually do… Unfortunately, one slipped through.”

The outlet explained that on 7 April an editor received a pitch from one Margot Blanchard about “the growing popularity of hyper-niche internet weddings”. The email bore “all the hallmarks of a great Wired story”. After a standard exchange about scope and fee, the editor commissioned the piece, which ran on 7 May.

Wired said that within days the newsroom realised the author could not provide sufficient information about herself. The journalist insisted on payment via PayPal or cheque.

Further investigation showed the story was fabricated.

“We made mistakes: the story did not undergo proper fact-checking and was not edited by a senior editor[…] We acted quickly when we discovered the deception and took steps to prevent a recurrence. In the new era every newsroom must be prepared for this.”

Press Gazette said the first to flag irregularities was Dispatch editor Jacob Furedi. He reported receiving a pitch from Blanchard about “Gravemont, a closed mining town in rural Colorado that had been repurposed into one of the most secretive death-investigation training centres.” He asked the supposed freelancer to show records requests; she ignored the request.

Meta is suspected of accessing galleries without user consent

Meta analyses and stores photos from devices. According to ZDNET, some Facebook users found two enabled options in Meta app settings that give the company access to the gallery. The feature is intended to use AI to offer “personalised creative ideas” such as travel montages and collages.

Settings that help avoid tracking. Source: ZDNET.

Media reported that the options for AI features called “suggestions to use photos from the gallery” are enabled for users who say they did not consent.

If a user taps “allow”, they agree to Meta’s AI terms and to analysis of “media and facial features”. Facebook then uses images from the gallery (including creation dates and the presence of people or objects) to suggest collages, themed albums, recap posts or AI-modified versions of images.

Researchers hid a trojan in an AI image

Researchers at Trail of Bits developed a new attack to steal user data. The method embeds malicious commands in images that are processed by AI systems before being passed to a large language model.

The idea is to use full-size images with “invisible” instructions that emerge when quality is reduced by resizing algorithms. When uploaded to AI systems, such images are automatically downscaled to improve performance and save resources.

Depending on the system, image-resizing algorithms may lighten the image using nearest-neighbour, bilinear or bicubic interpolation.

Bicubic interpolation hides the AI trojan. Source: Trail of Bits.

In Trail of Bits’ example, when downscaled bicubically the dark areas of the malicious image turn red and hidden text appears in black.

From the user’s perspective nothing unusual happens, but in fact the model executes hidden instructions that can lead to data leakage or other risky actions.

The researchers confirmed their method applies to:

Google will ban software from unverified developers

On 25 August Google announced it will soon stop allowing software from unverified developers in Google Play. A new Android protection system will block the installation of malicious apps when downloading from third-party sources.

“While the threat is more associated with third-party sources, the developer verification requirement now applies to apps from Google Play as well as apps in third-party stores,” the team added.

Early access to verification opens in October, and in March 2026 the system will become available to all Android developers. In September the mandatory identity verification requirement will take effect for Brazil, Indonesia, Singapore and Thailand, and in 2027 worldwide.

Also on ForkLog:

What to read this weekend?

Why trading platforms block accounts, how a “clean” transaction can still lead to frozen funds, and how deeply AML systems trace transfer chains? In a guest piece for ForkLog, Fedor Ivanov—director of analytics at the company “Shard”—explains.

Exit mobile version