Connect with us

Tech

How to ensure high-quality synthetic wireless data when real-world data runs dry

Published

on

How to ensure high-quality synthetic wireless data when real-world data runs dry


: Our quality assessment and quality-guided utilization of wireless synthetic data. Generative models produce synthetic data from conditions to supplement data quantity for wireless applications. Compared to previous quality-oblivious utilization using all synthetic data with conditions as labels, we assess synthetic data quality, reveal its affinity limitation, and propose a quality-guided utilization scheme incorporating filtered synthetic samples with assigned pseudo-labels for better data quality and task performance. Credit: arXiv (2025). DOI: 10.48550/arxiv.2506.23174

To train artificial intelligence (AI) models, researchers need good data and lots of it. However, most real-world data has already been used, leading scientists to generate synthetic data. While the generated data helps solve the issue of quantity, it may not always have good quality, and assessing its quality has been overlooked.

Wei Gao, associate professor of electrical and computer engineering at the University of Pittsburgh Swanson School of Engineering, has collaborated with researchers from Peking University to develop analytical metrics to qualitatively evaluate the quality of synthetic wireless data. The researchers have created a novel framework that significantly improves the task-driven training of AI models using synthetic wireless data.

Their work is detailed on the arXiv preprint server in a study titled “Data Can Speak for Itself: Quality-Guided Utilization of Wireless Synthetic Data,” which received the Best Paper Award in June at the MobiSys 2025 International Conference on Mobile Systems, Applications, and Services.

Assessing affinity and diversity

“Synthetic data is vital for training AI models, but for modalities such as images, video, or sound, and especially wireless signals, generating good data can be difficult,” said Gao, who also directs the Pitt Intelligent Systems Laboratory.

Gao has developed metrics to quantify and diversity, essential qualities for to be used for effectively training AI models.

“Generated data shouldn’t be random,” said Gao. “Take human faces. If you’re training an AI model to identify human faces, you need to ensure that the images of faces represent actual faces. They can’t have three eyes or two noses. They must have affinity.”

The images also need diversity. Training an AI model on a million images of an identical face won’t achieve much. While the faces must have affinity, they must also be different, as human faces are. As Gao noted, “AI models learn from variation.”

Different tasks have different requirements for judging affinity and diversity. Recognizing a specific human face is different than distinguishing it from that of a dog or a cat, with each task having unique data requirements. Therefore, in systemically assessing the quality of synthetic data, the team applied a task-specific approach.

“We applied our method to downstream tasks and evaluated the existing work of synthesizing data,” said Gao. “We found that most synthetic data achieved good diversity, but some had problems satisfying affinity, especially wireless signals.”

The challenge of synthetic wireless data

Today, wireless signals are used in technologies such as home and sleep monitoring, interactive gaming, and virtual reality. Cell phone and Wi-Fi signals, as , hit objects and bounce back toward their source. These signals can be interpreted to indicate everything from sleep patterns to the shape of a person sitting on a couch.

To advance this technology, researchers need more wireless data to train models to recognize human behaviors in the signal patterns. However, as a waveform, the signals are difficult for humans to evaluate.

It’s not like human faces, which can be clearly defined. “Our research found that current synthetic wireless data is limited in its affinity,” said Gao. “This leads to mislabeled data and degraded task performance.”

To improve affinity in , the researchers took a semi-supervised learning approach. “We used a small amount of labeled synthetic data, which was verified as legitimate,” Gao said. “We used this data to teach the model what is and isn’t legitimate.”

Gao and his collaborators developed SynCheck, a framework that filters out synthetic wireless samples with low affinity and labels the remaining samples during iterative training of a model.

“We found that our system improves performance by 4.3% whereas a nonselective use of synthetic wireless data degrades performance by 13.4%,” Gao noted.

This research takes an important first step toward ensuring not just an endless stream of data, but of quality data that scientists can use to train more sophisticated AI models.

More information:
Chen Gong et al, Data Can Speak for Itself: Quality-guided Utilization of Wireless Synthetic Data, arXiv (2025). DOI: 10.48550/arxiv.2506.23174

Journal information:
arXiv


Citation:
How to ensure high-quality synthetic wireless data when real-world data runs dry (2025, September 15)
retrieved 15 September 2025
from https://techxplore.com/news/2025-09-high-quality-synthetic-wireless-real.html

This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.





Source link

Tech

Affordability Doesn’t Suck With Eufy’s Newest Robot Vac

Published

on

Affordability Doesn’t Suck With Eufy’s Newest Robot Vac


Where the X10 Pro Omni had rotating mop pads, the rolling mop pad on the Omni C28 continuously self-cleans to prevent spreading dirt or grime to other parts of the house. Both apply downward pressure, but neither can spot dirtier places on their own as pricier, AI-powered robot vacuums will. Still, I was happy to see that it was able to scrub away some of the large dirt smudges in my entryway, though it didn’t get all of them. It also didn’t manage to scrub away all of the cherry juice I intentionally spilled in my routine mess setup for robot vacuum testing, even after sending the vacuum to do a second mopping job on one of the spots.

Photograph: Nena Farrell

Still, the Omni C28 was able to raise its roller mop high enough when it switched from mopping my floors to vacuuming my living room rug that there was no hint of dampness anywhere. The older X10 did get my colleague Adrienne So’s carpet wet, but it didn’t get mine wet, though my carpet is a fairly low pile. It did a fine job vacuuming the carpet, though I could tell the difference in suction between this and more powerful vacuums I’ve tested.

The base station is nice and compact, and includes drying fans to dry off the roller mop. That does mean there’s a gentle fan noise in the background for a couple of hours after you use this robot vacuum, which was more annoying than I expected, but you could easily place this vacuum’s base station in a less central spot in your home so you don’t hear it. You could also set up a schedule for the vacuum to run in the morning and finish its drying job before you get home.

Multi-Floor Madness

Image may contain Indoors Interior Design and Floor

Photograph: Nena Farrell

My favorite feature on the Omni C28 is that, even at this price point, it can still learn multiple maps. While it can’t climb up stairs, you can move it around your home and switch the maps in the app to the floor you’ve relocated to. This isn’t new for Eufy, as the older affordable model can do that too, but it’s nice to see the feature maintained when I’ve tried more expensive robot vacuums that don’t include it. It’s pretty simple to use; you’ll go to the maps, select “make a new map,” and then activate the robot to map. Once the map is made, you’ll switch to that map from the little map icon on the right side, which will label them with numbers in the order you created them.



Source link

Continue Reading

Tech

‘She’s Never Going to Age’: Porn Stars Are Embracing AI Clones to Stay Forever Young

Published

on

‘She’s Never Going to Age’: Porn Stars Are Embracing AI Clones to Stay Forever Young


Lisa Ann technically quit the porn business in 2019, but for $30 a month you can now dream up any X-rated scenario of her on your computer.

Ann, 53, was an adult performer for three decades starting in the mid 1990s and retired because she had reached her savings goal.

But last year she had a change of heart. Ann, who considers herself an AI fanatic, signed a contract with OhChat, a London-based AI companion company, to license her likeness on its platform, essentially creating an AI version of her in every way that can be used to make sex scenes for paying customers: same voice, same physique, and same pillowy brown hair.

As issues around deepfakes intensify and questions about the future of the adult industry become more dire with the passing of age-verification laws, several AI companion platforms want to create a new standard for consent-driven AI porn. More than sexting a faceless chatbot, digital twins—also called duplicates, doubles, clones, or replicas—draw on the exact likeness, including speech and mannerisms, of your favorite performers and creators.

Ann, now a self-help author and sports radio host, represents a growing faction in adult entertainment who not only believe AI is going to reshape the sex industry but who want a say in how that change materializes. She sees the decision to partner with OhChat as a way to tap into a fountain of youth—and stay at her peak forever.

“This keeps my name alive,” she says of her digital twin. “She’s never going to age.”

For Cherie Deville, a 47-year-old performer known for shooting MILF content, digital twins are just a smart business strategy to earn passive income while the opportunity is hot. “We can either let the makers of AI take the lion’s share of the money in the sex-work space, or creators and businesses can get on board and start creating their own revenue sources through AI.”

OhChat creators, who must submit 30 images and undergo voice training with a bot, sign an agreement stating the level of sexual content allowed for their digital twin. Ann is considered a “Level 4”—the highest on the platform—which means paying members can create scenarios and chats of her that include full nudity and sex. Per the company’s guidelines, clones can be deleted at any time.

“For guys that like to say good morning or good night, they now have that access. The fact that I’m not shooting scenes anymore also allows new scenes to be created,” Ann says.

Once described by CEO Nic Young as the “love child between OnlyFans and OpenAI,” OhChat launched in 2024 and has since scaled to over 400,000 users. According to data shared with WIRED, OhChat has 250 creators, 90 percent of which are female, and has contracts with celebrities Carmen Elektra and Joe Exotic. The platform runs on a tiered subscription model—$5 a month for on-demand texts or up to $30 for unlimited adult content—and the company, like OnlyFans, takes a 20 percent cut.

Other competitors in the space include My.Club, Joi AI and SinfulX AI, the platform that adult film actress Georgia Koneva partnered with this month, saying, in a press statement, that her avatar gave her a “new way to share my voice and personality with the people who follow me.” According to SinfulX AI, it also develops “original” synthetic characters using licensed source imagery from adult performers whose content it has the rights to use. In the same statement, the company said that those AI-generated “characters” are “designed not to replicate any single individual while still maintaining the realism for which its content is known.”



Source link

Continue Reading

Tech

AI system learns to keep warehouse robot traffic running smoothly

Published

on

AI system learns to keep warehouse robot traffic running smoothly



Inside a giant autonomous warehouse, hundreds of robots dart down aisles as they collect and distribute items to fulfill a steady stream of customer orders. In this busy environment, even small traffic jams or minor collisions can snowball into massive slowdowns.

To avoid such an avalanche of inefficiencies, researchers from MIT and the tech firm Symbotic developed a new method that automatically keeps a fleet of robots moving smoothly. Their method learns which robots should go first at each moment, based on how congestion is forming, and adapts to prioritize robots that are about to get stuck. In this way, the system can reroute robots in advance to avoid bottlenecks.

The hybrid system utilizes deep reinforcement learning, a powerful artificial intelligence method for solving complex problems, to figure out which robots should be prioritized. Then, a fast and reliable planning algorithm feeds instructions to the robots, enabling them to respond rapidly in constantly changing conditions.

In simulations inspired by actual e-commerce warehouse layouts, this new approach achieved about a 25 percent gain in throughput over other methods. Importantly, the system can quickly adapt to new environments with different quantities of robots or varied warehouse layouts.

“There are a lot of decision-making problems in manufacturing and logistics where companies rely on algorithms designed by human experts. But we have shown that, with the power of deep reinforcement learning, we can achieve super-human performance. This is a very promising approach, because in these giant warehouses even a 2 or 3 percent increase in throughput can have a huge impact,” says Han Zheng, a graduate student in the Laboratory for Information and Decision Systems (LIDS) at MIT and lead author of a paper on this new approach.

Zheng is joined on the paper by Yining Ma, a LIDS postdoc; Brandon Araki and Jingkai Chen of Symbotic; and senior author Cathy Wu, the Class of 1954 Career Development Associate Professor in Civil and Environmental Engineering (CEE) and the Institute for Data, Systems, and Society (IDSS) at MIT, and a member of LIDS. The research appears today in the Journal of Artificial Intelligence Research.

Rerouting robots

Coordinating hundreds of robots in an e-commerce warehouse simultaneously is no easy task.

The problem is especially complicated because the warehouse is a dynamic environment, and robots continually receive new tasks after reaching their goals. They need to be rapidly redirected as they leave and enter the warehouse floor.

Companies often leverage algorithms written by human experts to determine where and when robots should move to maximize the number of packages they can handle.

But if there is congestion or a collision, a firm may have no choice but to shut down the entire warehouse for hours to manually sort the problem out.

“In this setting, we don’t have an exact prediction of the future. We only know what the future might hold, in terms of the packages that come in or the distribution of future orders. The planning system needs to be adaptive to these changes as the warehouse operations go on,” Zheng says.

The MIT researchers achieved this adaptability using machine learning. They began by designing a neural network model to take observations of the warehouse environment and decide how to prioritize the robots. They train this model using deep reinforcement learning, a trial-and-error method in which the model learns to control robots in simulations that mimic actual warehouses. The model is rewarded for making decisions that increase overall throughput while avoiding conflicts.

Over time, the neural network learns to coordinate many robots efficiently.

“By interacting with simulations inspired by real warehouse layouts, our system receives feedback that we use to make its decision-making more intelligent. The trained neural network can then adapt to warehouses with different layouts,” Zheng explains.

It is designed to capture the long-term constraints and obstacles in each robot’s path, while also considering dynamic interactions between robots as they move through the warehouse.

By predicting current and future robot interactions, the model plans to avoid congestion before it happens.

After the neural network decides which robots should receive priority, the system employs a tried-and-true planning algorithm to tell each robot how to move from one point to another. This efficient algorithm helps the robots react quickly in the changing warehouse environment.

This combination of methods is key.

“This hybrid approach builds on my group’s work on how to achieve the best of both worlds between machine learning and classical optimization methods. Pure machine-learning methods still struggle to solve complex optimization problems, and yet it is extremely time- and labor-intensive for human experts to design effective methods. But together, using expert-designed methods the right way can tremendously simplify the machine learning task,” says Wu.

Overcoming complexity

Once the researchers trained the neural network, they tested the system in simulated warehouses that were different than those it had seen during training. Since industrial simulations were too inefficient for this complex problem, the researchers designed their own environments to mimic what happens in actual warehouses.

On average, their hybrid learning-based approach achieved 25 percent greater throughput than traditional algorithms as well as a random search method, in terms of number of packages delivered per robot. Their approach could also generate feasible robot path plans that overcame congestion caused by traditional methods.

“Especially when the density of robots in the warehouse goes up, the complexity scales exponentially, and these traditional methods quickly start to break down. In these environments, our method is much more efficient,” Zheng says.

While their system is still far away from real-world deployment, these demonstrations highlight the feasibility and benefits of using a machine learning-guided approach in warehouse automation.

In the future, the researchers want to include task assignments in the problem formulation, since determining which robot will complete each task impacts congestion. They also plan to scale up their system to larger warehouses with thousands of robots.

This research was funded by Symbotic.



Source link

Continue Reading

Trending