Connect with us

Tech

What is driving the rise of infostealer malware? | Computer Weekly

Published

on

What is driving the rise of infostealer malware? | Computer Weekly


Cyber criminals would much rather log in than hack in. That’s why infostealer malware, designed to exfiltrate user credentials, browser data, messages, documents, images, and device information, is becoming more widespread. Stealing sensitive information opens a lot of doors for cyber criminals. They can log in using the stolen credentials and bypass multi-factor authentication with hijacked session cookies. They can take over accounts, commit fraud, craft better phishing campaigns, or simply sell the data to the highest bidder on the dark web.

Infostealer malware is a growing problem for cyber security teams, and our data tells us that attacks have the potential to cause significant damage to businesses. That is because lax security policies are creating the perfect conditions for infostealer attacks to thrive.

The scale of the problem

Socura and Flare recently analysed the digital footprint of the UK’s biggest companies, looking for stolen credentials across the clear and dark web. In total, we discovered 28,000 instances of stolen FTSE 100 employee credentials that had been leaked in infostealer logs. We also found cookies that were valid for several years, giving attackers another way to log in and bypass security controls like MFA.

Ideally, the UK’s corporate giants would be immune to these threats. After all, they have the budgets and the tools to be the most secure. Yet, despite their resources, they remain vulnerable. This raises a critical point: if the industry leaders are struggling to manage their threat exposure, then small and medium-sized businesses must face similar challenges.

Contributing factors

One of the major reasons that infostealer malware has been allowed to flourish is the blurred (almost invisible) line between corporate and personal IT. Employees are using their work devices, accounts, and applications at home and for personal use. They are using their personal devices for work tasks, too.

A surprisingly common source of infostealer malware is video games, specifically infected mods for popular games like Roblox, Fortnite and Grand Theft Auto. If you have an employee using a device to check their work emails and access sensitive documents, while also using the device for gaming (themselves or a family member), that poses a significant risk.

The threat of infostealer malware is being made even worse because employees continue to use the same weak passwords across all their accounts. Our research showed that more than half of FTSE 100 companies had at least one instance of an employee credential where the password was simply ‘password’. Likewise, these weak passwords or slight variations are often recycled across services used for business and personal purposes. If malware captures a login for one site, criminals will often test that password elsewhere, potentially unlocking a treasure trove of additional data they can use to further their objectives.

Recommended actions

To protect against the risks of infostealer malware, it is beneficial to take a multi-layered approach. This means looking at ways to prevent leaks, while also ensuring the business is resilient if leaks do occur, which they inevitably will at some point.

Following NCSC guidance is a great starting point. This might include employee education on password hygiene and the rollout of password managers. We also suggest implementing multi-factor authentication across the board, ideally using phishing-resistant options like passkeys to avoid sophisticated attacks.

It is also worth reviewing how personal devices and applications are managed, as these are common entry points for malware. Updating BYOD policies and implementing conditional access policies, to block users from accessing corporate resources based on factors such as device compliance and risk level, are also recommended.

Finally, proactive threat exposure monitoring allows businesses to spot leaked credentials on the dark web before they are exploited. We suggest implementing controls to flag unusual activity and automating response actions, such as initiating password resets and isolating machines, as soon as risks are identified.

Final thoughts

The threat of leaked credentials and infostealer malware might seem daunting, but there are definitive actions businesses can take to minimise the risk. This starts with acknowledging just how widespread this threat has become.

Cyber criminals would rather log in than hack in. Let’s make sure we stop handing them the keys and making their job as simple as turning a lock.

Anne Heim is threat intelligence lead at Socura, a provider of managed detection and response (MDR) services.



Source link

Continue Reading
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *

Tech

We Just Found Out Taylor Swift Sleeps on a Coop Pillow—They’re Having a Flash Sale to Celebrate

Published

on

We Just Found Out Taylor Swift Sleeps on a Coop Pillow—They’re Having a Flash Sale to Celebrate


While I’m a mattress and sleep product expert, thanks to years of hands-on experience, I’m also aware that my opinion is not the end-all, be-all for everyone. However, when a megastar is also a fan of a product you’ve reviewed, it’s a good confirmation that you’re on the right track.

Taylor Swift, as it would turn out, is also a fan of Coop Sleep Goods—which we can confirm based on this December 10 Late Show With Stephen Colbert appearance.

Coop’s got some of our favorite pillows, particularly the Original Adjustable pillow. It comes in three shapes: the Crescent, the Cut Out, and the Classic, which is a traditional rectangular shape. I love (and regularly sleep on) the Crescent, which has a gentle curve on the bottom to allow for movement while maintaining head and neck support.



Source link

Continue Reading

Tech

Nvidia Becomes a Major Model Maker With Nemotron 3

Published

on

Nvidia Becomes a Major Model Maker With Nemotron 3


Nvidia has made a fortune supplying chips to companies working on artificial intelligence, but today the chipmaker took a step toward becoming a more serious model maker itself by releasing a series of cutting-edge open models, along with data and tools to help engineers use them.

The move, which comes at a moment when AI companies like OpenAI, Google, and Anthropic are developing increasingly capable chips of their own, could be a hedge against these firms veering away from Nvidia’s technology over time.

Open models are already a crucial part of the AI ecosystem with many researchers and startups using them to experiment, prototype, and build. While OpenAI and Google offer small open models, they do not update them as frequently as their rivals in China. For this reason and others, open models from Chinese companies are currently much more popular, according to data from Hugging Face, a hosting platform for open source projects.

Nvidia’s new Nemotron 3 models are among the best that can be downloaded, modified, and run on one’s own hardware, according to benchmark scores shared by the company ahead of release.

“Open innovation is the foundation of AI progress,” CEO Jensen Huang said in a statement ahead of the news. “With Nemotron, we’re transforming advanced AI into an open platform that gives developers the transparency and efficiency they need to build agentic systems at scale.”

Nvidia is taking a more fully transparent approach than many of its US rivals by releasing the data used to train Nemotron—a fact that should help engineers modify the models more easily. The company is also releasing tools to help with customization and fine-tuning. This includes a new hybrid latent mixture-of-experts model architecture, which Nvidia says is especially good for building AI agents that can take actions on computers or the web. The company is also launching libraries that allow users to train agents to do things using reinforcement learning, which involves giving models simulated rewards and punishments.

Nemotron 3 models come in three sizes: Nano, which has 30 billion parameters; Super, which has 100 billion; and Ultra, which has 500 billion. A model’s parameters loosely correspond to how capable it is as well as how unwieldy it is to run. The largest models are so cumbersome that they need to run on racks of expensive hardware.

Model Foundations

Kari Ann Briski, vice president of generative AI software for enterprise at Nvidia, said open models are important to AI builders for three reasons: Builders increasingly need to customize models for particular tasks; it often helps to hand queries off to different models; and it is easier to squeeze more intelligent responses from these models after training by having them perform a kind of simulated reasoning. “We believe open source is the foundation for AI innovation, continuing to accelerate the global economy,” Briski said.

The social media giant Meta released the first advanced open models under the name Llama in February 2023. As competition has intensified, however, Meta has signaled that its future releases might not be open source.

The move is part of a larger trend in the AI industry. Over the past year, US firms have moved away from openness, becoming more secretive about their research and more reluctant to tip off their rivals about their latest engineering tricks.



Source link

Continue Reading

Tech

This Startup Wants to Build Self-Driving Car Software—Super Fast

Published

on

This Startup Wants to Build Self-Driving Car Software—Super Fast


For the last year and a half, two hacked white Tesla Model 3 sedans each loaded with five extra cameras and one palm-sized supercomputer have quietly cruised around San Francisco. In a city and era swarming with questions about the capabilities and limits of artificial intelligence, the startup behind the modified Teslas is trying to answer what amounts to a simple question: How quickly can a company build autonomous vehicle software today?

The startup, which is making its activities public for the first time today, is called HyprLabs. Its 17-person team (just eight of them full-time) is divided between Paris and San Francisco, and the company is helmed by an autonomous vehicle company veteran, Zoox cofounder Tim Kentley-Klay, who suddenly exited the now Amazon-owned firm in 2018. Hypr has taken in relatively little funding, $5.5 million since 2022, but its ambitions are wide-ranging. Eventually, it plans to build and operate its own robots. “Think of the love child of R2-D2 and Sonic the Hedgehog,” Kentley-Klay says. “It’s going to define a new category that doesn’t currently exist.”

For now, though, the startup is announcing its software product called Hyprdrive, which it bills as a leap forward in how engineers train vehicles to pilot themselves. These sorts of leaps are all over the robotics space, thanks to advances in machine learning that promise to bring down the cost of training autonomous vehicle software, and the amount of human labor involved. This training evolution has brought new movement to a space that for years suffered through a “trough of disillusionment,” as tech builders failed to meet their own deadlines to operate robots in public spaces. Now, robotaxis pick up paying passengers in more and more cities, and automakers make newly ambitious promises about bringing self-driving to customers’ personal cars.

But using a small, agile, and cheap team to get from “driving pretty well” to “driving much more safely than a human” is its own long hurdle. “I can’t say to you, hand on heart, that this will work,” Kentley-Klay says. “But what we’ve built is a really solid signal. It just needs to be scaled up.”

Old Tech, New Tricks

HyprLabs’ software training technique is a departure from other robotics’ startups approaches to teaching their systems to drive themselves.

First, some background: For years, the big battle in autonomous vehicles seemed to be between those who used just cameras to train their software—Tesla!—and those who depended on other sensors, too—Waymo, Cruise!—including once-expensive lidar and radar. But below the surface, larger philosophical differences churned.

Camera-only adherents like Tesla wanted to save money while scheming to launch a gigantic fleet of robots; for a decade, CEO Elon Musk’s plan has been to suddenly switch all of his customers’ cars to self-driving ones with the push of a software update. The upside was that these companies had lots and lots of data, as their not-yet self-driving cars collected images wherever they drove. This information got fed into what’s called an “end-to-end” machine learning model through reinforcement. The system takes in images—a bike—and spits out driving commands—move the steering wheel to the left and go easy on the acceleration to avoid hitting it. “It’s like training a dog,” says Philip Koopman, an autonomous vehicle software and safety researcher at Carnegie Mellon University. “At the end, you say, ‘Bad dog,” or ‘Good dog.’”



Source link

Continue Reading

Trending