Connect with us

Tech

How to ensure high-quality synthetic wireless data when real-world data runs dry

Published

on

How to ensure high-quality synthetic wireless data when real-world data runs dry


: Our quality assessment and quality-guided utilization of wireless synthetic data. Generative models produce synthetic data from conditions to supplement data quantity for wireless applications. Compared to previous quality-oblivious utilization using all synthetic data with conditions as labels, we assess synthetic data quality, reveal its affinity limitation, and propose a quality-guided utilization scheme incorporating filtered synthetic samples with assigned pseudo-labels for better data quality and task performance. Credit: arXiv (2025). DOI: 10.48550/arxiv.2506.23174

To train artificial intelligence (AI) models, researchers need good data and lots of it. However, most real-world data has already been used, leading scientists to generate synthetic data. While the generated data helps solve the issue of quantity, it may not always have good quality, and assessing its quality has been overlooked.

Wei Gao, associate professor of electrical and computer engineering at the University of Pittsburgh Swanson School of Engineering, has collaborated with researchers from Peking University to develop analytical metrics to qualitatively evaluate the quality of synthetic wireless data. The researchers have created a novel framework that significantly improves the task-driven training of AI models using synthetic wireless data.

Their work is detailed on the arXiv preprint server in a study titled “Data Can Speak for Itself: Quality-Guided Utilization of Wireless Synthetic Data,” which received the Best Paper Award in June at the MobiSys 2025 International Conference on Mobile Systems, Applications, and Services.

Assessing affinity and diversity

“Synthetic data is vital for training AI models, but for modalities such as images, video, or sound, and especially wireless signals, generating good data can be difficult,” said Gao, who also directs the Pitt Intelligent Systems Laboratory.

Gao has developed metrics to quantify and diversity, essential qualities for to be used for effectively training AI models.

“Generated data shouldn’t be random,” said Gao. “Take human faces. If you’re training an AI model to identify human faces, you need to ensure that the images of faces represent actual faces. They can’t have three eyes or two noses. They must have affinity.”

The images also need diversity. Training an AI model on a million images of an identical face won’t achieve much. While the faces must have affinity, they must also be different, as human faces are. As Gao noted, “AI models learn from variation.”

Different tasks have different requirements for judging affinity and diversity. Recognizing a specific human face is different than distinguishing it from that of a dog or a cat, with each task having unique data requirements. Therefore, in systemically assessing the quality of synthetic data, the team applied a task-specific approach.

“We applied our method to downstream tasks and evaluated the existing work of synthesizing data,” said Gao. “We found that most synthetic data achieved good diversity, but some had problems satisfying affinity, especially wireless signals.”

The challenge of synthetic wireless data

Today, wireless signals are used in technologies such as home and sleep monitoring, interactive gaming, and virtual reality. Cell phone and Wi-Fi signals, as , hit objects and bounce back toward their source. These signals can be interpreted to indicate everything from sleep patterns to the shape of a person sitting on a couch.

To advance this technology, researchers need more wireless data to train models to recognize human behaviors in the signal patterns. However, as a waveform, the signals are difficult for humans to evaluate.

It’s not like human faces, which can be clearly defined. “Our research found that current synthetic wireless data is limited in its affinity,” said Gao. “This leads to mislabeled data and degraded task performance.”

To improve affinity in , the researchers took a semi-supervised learning approach. “We used a small amount of labeled synthetic data, which was verified as legitimate,” Gao said. “We used this data to teach the model what is and isn’t legitimate.”

Gao and his collaborators developed SynCheck, a framework that filters out synthetic wireless samples with low affinity and labels the remaining samples during iterative training of a model.

“We found that our system improves performance by 4.3% whereas a nonselective use of synthetic wireless data degrades performance by 13.4%,” Gao noted.

This research takes an important first step toward ensuring not just an endless stream of data, but of quality data that scientists can use to train more sophisticated AI models.

More information:
Chen Gong et al, Data Can Speak for Itself: Quality-guided Utilization of Wireless Synthetic Data, arXiv (2025). DOI: 10.48550/arxiv.2506.23174

Journal information:
arXiv


Citation:
How to ensure high-quality synthetic wireless data when real-world data runs dry (2025, September 15)
retrieved 15 September 2025
from https://techxplore.com/news/2025-09-high-quality-synthetic-wireless-real.html

This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.





Source link

Tech

Hollywood Is Losing Audiences to AI Fatigue

Published

on

Hollywood Is Losing Audiences to AI Fatigue


An insurrectionist robot unleashed by a mad inventor in Fritz Lang’s Metropolis. HAL 9000 sabotaging a manned mission to Jupiter in 2001: A Space Odyssey. Skynet, the self-aware global defense network that seeks to exterminate humanity throughout the Terminator franchise.

Hollywood has never wanted for audacious depictions of artificial intelligence or the ways in which it could alter the fate of our species. But the rapid integration of AI into the studio system and our now unavoidable interactions with it have severely compromised the genre, not to mention film as a medium.

On the one hand, it’s perfectly understandable that screenwriters and studios would return to the subject of AI in recent years, particularly since it provokes such fierce debate within the industry. (A major cause of the 2023 labor strikes was the threat that AI posed to creative jobs.) Still, the novelty faded fast.

Consider M3GAN, a campy horror flick about an artificially intelligent doll who starts killing people, released just a week after the debut of ChatGPT in 2022: It was a surprise box-office smash. Last year’s sequel? A critical and commercial flop. Mission: Impossible—Dead Reckoning (2023) introduced a rogue AI called The Entity as a final adversary for Ethan Hunt and crew. The resolution of its cliff-hanger ending and blockbuster finale for the spy saga, Mission: Impossible—The Final Reckoning (2025), underperformed its predecessor, and neither quite justified their expense.

The latest AI-themed bomb is Mercy, a crime thriller starring Chris Pratt as an LAPD detective strapped into a chair who has 90 minutes to pull enough evidence from security cameras and phone records to convince a stern judge bot (Rebecca Ferguson) that he didn’t kill his wife—or else face instant execution. Despite releasing in January, one reviewer has already declared it “the worst movie of 2026,” and judging by its mediocre ticket sales, many US moviegoers decided as much from the trailer alone. It’s almost as if nobody cared whether a fictional software program might be capable of sparing a life when real health insurance claims are being denied by algorithms already.

For those few who did see it, Mercy fell far short of its dystopian premise, failing to grapple with the ethics of such a surveillance state and its medieval-modern justice system in favor of cheap relativism. Spoiler: Pratt’s character and the AI ultimately team up to stop the real bad guys as the bot begins to show signs of unrobotic emotion and doubt, which manifest as glitches in the program. By the end, Pratt is delivering a true groaner of a we’re-not-so-different speech to the holographic Ferguson. “Human or AI, we all make mistakes,” he says. “And we learn.”

While the naive belief in AI’s progress toward enlightenment feels dated on arrival, you are also reminded of how prophetically cynical something like Paul Verhoeven’s RoboCop, now almost 40 years old, was in addressing a future of cybernetic fascism. Contrary to that kind of pitch-black, violent satire, the current trend seems to be propagandistic narratives about how AIs are scary at first but secretly good. (See also: Tron: Ares, Disney’s wildly misguided attempt to leverage an old IP for the era of large language models, another cinematic train wreck of 2025.)

In fact, the insistence on some inborn value or honor to artificial intelligence may be the driving force behind the new Time Studios web series On This Day…1776. Conceived as a blow-by-blow account of the year the American colonies declared independence from the British crown, it consists of short YouTube videos generated in part by Google DeepMind (though actual actors supply voiceovers). The project has drawn serious attention and scorn because acclaimed director Darren Aronofsky served as executive producer via his creative studio Primordial Soup, launched last year in a partnership with Google to explore the applications of AI in filmmaking. It probably doesn’t help that Aronofsky and company are valorizing the country’s founders in the same aesthetic that has defined the authoritarian meme culture of Donald Trump’s second term.



Source link

Continue Reading

Tech

Half of Google’s software development now AI-generated | Computer Weekly

Published

on

Half of Google’s software development now AI-generated | Computer Weekly


As much as half of all the code produced at Alphabet, the parent company of Google, is being generated by artificial intelligence (AI) coding agents.

The use of AI to drive operational efficiency and free up more money to invest in AI capacity was one of the points made by Anat Ashkenazi, senior vice-president and chief financial officer of Alphabet and Google, during the company’s latest quarterly earnings call.

For its fourth quarter of 2025, Alphabet reported revenue of $114bn, up 18% year over year. For the full year, it posted revenue of $403bn, a 15% increase from the previous year.

The company is seeing a huge increase in demand for Google Cloud and its AI-powered services. During its latest earnings call, in a response that suggests Alphabet does not need to expand its software developer workforce, Ashkenazi said: “We look at coding productivity. About 50% of our code is written by coding agents, which are then reviewed by our own engineers. This certainly helps our engineers do more and move faster with the current footprint.”

Ashkenazi said 60% of Alphabet’s 2025 capital expenditure (capex) was allocated to servers, with the remaining 40% directed towards datacentres and networking equipment. A similar amount looks set to be spent in 2026, with Alphabet predicting it will spend between $175bn and $185bn on servers, datacentres and networking equipment. Its latest quarterly earnings call suggests capex is primarily focused on AI infrastructure and technical innovation to meet growing demand.

While Google does not have the breadth of AWS services or the deep corporate foothold of Microsoft, its steady effort to win enterprise customers is now turbocharged by its AI-native cloud offerings
Lee Sustar, Forrester

There is increasing concern in stock markets that the huge investments in AI infrastructure will not deliver a return on investment. In response to questions about AI capacity challenges and compute demand, Sundar Pichai, CEO of Alphabet and Google, said: “We’ve been supply-constrained, even as we’ve been ramping up our capacity. Obviously, our capex spend this year is an eye towards the future, and you have to keep in mind, some of the time, horizons are increasing in the supply chain. So, we are constantly planning for the long-term and working towards that. And, obviously, how we close the gap this year is a function of what we have done in the prior years. And so there is that time delay to keep in mind.”

The investment in AI infrastructure is needed to support demand for Google Cloud and the AI services the company provides. The quarterly filing shows Google Cloud’s annual run rate is over $70bn.

Pichai said Google Cloud has sold more than eight million paid seats of Gemini Enterprise, its AI platform, to over 2,800 companies. He also stated that over 120,000 enterprises use Google’s Gemini AI models, including major companies such as Airbus, Honeywell, Salesforce and Shopify, with existing customers increasing their spending, outpacing their initial commitments by over 30%.

“Nearly 75% of Google Cloud customers have used our vertically optimised AI, from chips, to models, to AI platforms, and enterprise AI agents, which offer superior performance, quality, security and cost-efficiency. These AI customers use 1.8 times as many products as those who do not, enabling us to diversify our product portfolio, deepen customer relationships and accelerate revenue growth,” added Pichai.

Forrester’s principal analyst, Lee Sustar, said: “Google Cloud’s quarterly revenue jump of 48% over the same period a year earlier is decisive evidence that it is a full-blown enterprise challenger to AWS [Amazon Web Services] and Microsoft Azure. While Google does not have the breadth of AWS services or the deep corporate foothold of Microsoft, its steady effort to win enterprise customers is now turbocharged by its AI-native cloud offerings. But this comes at a hefty price for parent Alphabet, which saw capital expenditure for the fourth quarter effectively double the amount of a year earlier.”



Source link

Continue Reading

Tech

Netflix Says if the HBO Merger Makes It Too Expensive, You Can Always Cancel

Published

on

Netflix Says if the HBO Merger Makes It Too Expensive, You Can Always Cancel


There is concern that subscribers might be negatively affected if Netflix acquires Warner Bros. Discovery’s streaming and movie studios businesses. One of the biggest fears is that the merger would lead to higher prices due to less competition for Netflix.

During a US Senate hearing Tuesday, Netflix co-CEO Ted Sarandos suggested that the merger would have an opposite effect.

Sarandos was speaking at a hearing held by the US Senate Judiciary Committee’s Subcommittee on Antitrust, Competition Policy, and Consumer Rights, “Examining the Competitive Impact of the Proposed Netflix-Warner Brothers Transaction.”

Sarandos aimed to convince the subcommittee that Netflix wouldn’t become a monopoly in streaming or in movie and TV production if regulators allowed its acquisition to close. Netflix is the largest subscription video-on-demand provider by subscribers (301.63 million as of January 2025), and Warner Bros. Discovery is the third (128 million streaming subscribers, including users of HBO Max and, to a smaller degree, Discovery+).

Speaking at the hearing, Sarandos said: “Netflix and Warner Bros. both have streaming services, but they are very complementary. In fact, 80 percent of HBO Max subscribers also subscribe to Netflix. We will give consumers more content for less.”

During the hearing, Democratic senator Amy Klobuchar of Minnesota asked Sarandos how Netflix can ensure that streaming remains “affordable” after a merger, especially after Netflix issued a price hike in January 2025 despite adding more subscribers.

Sarandos said the streaming industry is still competitive. The executive claimed that previous Netflix price hikes have come with “a lot more value” for subscribers.

“We are a one-click cancel, so if the consumer says, ‘That’s too much for what I’m getting,’ they can cancel with one click,” Sarandos said.

When pressed further on pricing, the executive argued that the merger doesn’t pose “any concentration risk” and that Netflix is working with the US Department of Justice on potential guardrails against more price hikes.

Sarandos claimed that the merger would “create more value for consumers.” However, his idea of value isn’t just about how much subscribers pay to stream but about content quality. By his calculations, which he provided without further details, Netflix subscribers spend an average of 35 cents per hour of content watched, compared to 90 cents for Paramount+.

The Netflix stat is similar to one provided by MoffettNathanson in January 2025, finding that in the prior quarter, on average, Netflix generated 34 cents in subscription fees per hour of content viewed per subscriber. At the time, the research firm said Paramount+ made an average of 76 cents per hour of content viewed per subscriber.

Downplaying Monopoly Concerns

Netflix views Warner as “both a competitor and a supplier,” Sarandos said when subcommittee chair Republican senator Mike Lee of Utah asked why Netflix wants to buy WB’s film studios, per Variety. The streaming executive claimed that Netflix’s “history is about adding more and more” content and choice.

During the hearing, Sarandos argued that streaming is a competitive business and pointed to Google, Apple, and Amazon as “deep-pocketed tech companies trying to run away with the TV business.” He tried to downplay concerns that Netflix could become a monopoly by emphasizing YouTube’s high TV viewership. Nielsen’s The Gauge tracker shows which platforms Americans use most when using their TVs (as opposed to laptops, tablets, or other devices). In December, it said that YouTube, not including YouTube TV, had more TV viewership (12.7 percent) than any other streaming video-on-demand service, including second-place Netflix (9 percent). Sarandos claimed that Netflix would have 21 percent of the streaming market if it merged with HBO Max.



Source link

Continue Reading

Trending