Connect with us

Tech

What are the storage requirements for AI training and inference? | Computer Weekly

Published

on

What are the storage requirements for AI training and inference? | Computer Weekly


Despite ongoing speculation around an investment bubble that may be set to burst, artificial intelligence (AI) technology is here to stay. And while an over-inflated market may exist at the level of the suppliers, AI is well-developed and has a firm foothold among organisations of all sizes.

But AI workloads place specific demands on IT infrastructure and on storage in particular. Data volumes can start big and then balloon, in particular during training phases as data is vectorised and checkpoints are created. Meanwhile, data must be curated, gathered and managed throughout its lifecycle.

In this article, we look at the key characteristics of AI workloads, the particular demands of training and inference on storage I/O, throughput and capacity, whether to choose object or file storage, and the storage requirements of agentic AI.

What are the key characteristics of AI workloads?

AI workloads can be broadly categorised into two key stages – training and inference.

During training, processing focuses on what is effectively pattern recognition. Large volumes of data are examined by an algorithm – likely part of a deep learning framework like TensorFlow or PyTorch – that aims to recognise features within the data.

This could be visual elements in an image or particular words or patterns of words within documents. These features, which might fall under the broad categories of “a cat” or “litigation”, for example, are given values and stored in a vector database.

The assigned values provide for further detail. So, for example “a tortoiseshell cat”, would comprise discrete values for “cat” and “tortoiseshell”, that make up the whole concept and allow comparison and calculation between images.

Once the AI system is trained on its data, it can then be used for inference – literally, to infer a result from production data that can be put to use for the organisation.

So, for example, we may have an animal tracking camera and we want it to alert us when a tortoiseshell cat crosses our garden. To do that it would infer the presence or not of a cat and whether it is tortoiseshell by reference to the dataset built during the training described above.

But, while AI processing falls into these two broad categories, it is not necessarily so clear cut in real life. It will always be the case that training will be done on an initial dataset. But after that it is likely that while inference is an ongoing process, training also becomes perpetual as new data is ingested and new inference results from it.

So, to labour the example, our cat-garden-camera system may record new cats of unknown types and begin to categorise their features and add them to the model.

What are the key impacts on data storage of AI processing?

At the heart of AI hardware are specialised chips called graphics processing units (GPUs). These do the grunt processing work of training and are incredibly powerful, costly and often difficult to procure. For these reasons their utilisation rates are a major operational IT consideration – storage must be able to handle their I/O demands so they are optimally used.

Therefore, data storage that feeds GPUs during training must be fast, so it’s almost certainly going to be built with flash storage arrays.

Another key consideration is capacity. That’s because AI datasets can start big and get much bigger. As datasets undergo training, the conversion of raw information into vector data can see data volumes expand by up to 10 times.

Also, during training, checkpointing is carried out at regular intervals, often after every “epoch” or pass through the training data, or after changes are made to parameters.

Checkpoints are similar to snapshots, and allow training to be rolled back to a point in time if something goes wrong so that existing processing does not go to waste. Checkpointing can add significant data volume to storage requirements.

So, sufficient storage capacity must be available, and will often need to scale rapidly.

What are the key impacts of AI processing on I/O and capacity in data storage?

The I/O demands of AI processing on storage are huge. It is often the case that model data in use will just not fit into a single GPU memory and so is parallelised across many of them.

Also, AI workloads and I/O differ significantly between training and inference. As we’ve seen, the massive parallel processing involved in training requires low latency and high throughput.

While low latency is a universal requirement during training, throughput demands may differ depending on the deep learning framework used. PyTorch, for example, stores model data as a large number of small files while TensorFlow uses a smaller number of large model files.

The model used can also impact capacity requirements. TensorFlow checkpointing tends towards larger file sizes, plus dependent data states and metadata, while PyTorch checkpointing can be more lightweight. TensorFlow deployments tend to have a larger storage footprint generally.

If the model is parallelised across numerous GPUs this has an effect on checkpoint writes and restores that mean storage I/O must be up to the job.

Does AI processing prefer file or object storage?

While AI infrastructure isn’t necessarily tied to one or other storage access method, object storage has a lot going for it.

Most enterprise data is unstructured data and exists at scale, and it is often what AI has to work with. Object storage is supremely well suited to unstructured data because of its ability to scale. It also comes with rich metadata capabilities that can help data discovery and classification before AI processing begins in earnest.

File storage stores data in a tree-like hierarchy of files and folders. That can become unwieldy to access at scale. Object storage, by contrast, stores data in a “flat” structure, by unique identifier, with rich metadata. It can mimic file and folder-like structures by addition of metadata labels, which many will be familiar with in cloud-based systems such as Google Drive, Microsoft OneDrive and so on.

Object storage can, however, be slow to access and lacks file-locking capability, though this is likely to be of less concern for AI workloads.

What impact will agentic AI have on storage infrastructure?

Agentic AI uses autonomous AI agents that can carry out specific tasks without human oversight. They are tasked with autonomous decision-making within specific, predetermined boundaries.

Examples would include the use of agents in IT security to scan for threats and take action without human involvement, to spot and initiate actions in a supply chain, or in a call centre to analyse customer sentiment, review order history and respond to customer needs.

Agentic AI is largely an inference phase phenomenon so compute infrastructure will not need to be up to training-type workloads. Having said that, agentic AI agents will potentially access multiple data sources across on-premises systems and the cloud. That will cover the range of potential types of storage in terms of performance.

But, to work at its best, agentic AI will need high-performance, enterprise-class storage that can handle a wide variety of data types with low latency and with the ability to scale rapidly. That’s not to say datasets in less performant storage cannot form part of the agentic infrastructure. But if you want your agents to work at their best you’ll need to provide the best storage you can.



Source link

Continue Reading
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *

Tech

Buried power lines could cut weather-related outages

Published

on

Buried power lines could cut weather-related outages


Undergrounding rates in 1990 on a scale from 0% (darkest shades) to 100% (lightest shades). Credit: arXiv (2024). DOI: 10.48550/arxiv.2402.06668

A Stanford analysis shows that strategic investment in burying power lines could shorten blackouts during extreme weather, enhancing energy reliability for millions of U.S. households.

As hurricanes intensify, wildfires spread, and winter storm patterns shift, the combination of extreme weather events and aging grid infrastructure threatens to make energy less reliable for tens of millions of U.S. households.

Experts say burying underground can harden the electrical system against threats from wind, ice, falling trees, and other weather-related hazards. Yet undergrounding power lines remains expensive and unevenly implemented. One obstacle has been a lack of information about where investments in undergrounding by utilities and communities could make the biggest difference for reliable power supplies.

In a recent study posted to the arXiv preprint server, Stanford University researchers led by Associate Professor Ram Rajagopal combined previously non-public and siloed datasets to reveal how the distribution of power lines above and below ground has changed since the 1990s. By combining these data with power outage records, the team modeled how having more power lines underground during recent extreme weather events could have shortened outages.

Buried power lines could cut weather-related outages
Undergrounding rates in 2020. Credit: arXiv (2024). DOI: 10.48550/arxiv.2402.06668

Patchy progress on burying power lines since 1990

Dense metropolitan areas on the East Coast, parts of southern Florida, and a few southwestern growth hubs were among the first to underground at least a quarter of their power line mileage. The overwhelming majority of power lines remained overhead in most U.S. counties in 1990.

By 2020, some fast-growing suburbs in southeastern and Sunbelt states showed modest increases in undergrounding. For most counties nationwide, however, the median percentage of power lines buried underground remained well below 15%. Large swaths of the Rockies, Midwest, and Gulf Coast showed virtually no change.

Where outages last the longest

Each year, tens of millions of Americans experience power outages. While households on average lose electricity for about four hours over the course of a year, some outages last a day or even weeks. Many of these longer outages are linked to events.

Buried power lines could cut weather-related outages
Annual average power outage time across the U.S. from 2013 to 2023, on a scale from less than one hour (lightest shades) to more than 24 hours (darkest shades). Credit: arXiv (2024). DOI: 10.48550/arxiv.2402.06668

New England’s 2017 ‘bomb cyclone’

A nor’easter or “bomb cyclone” that struck Maine, Vermont, and New Hampshire in October 2017 left people without power on average for 27.3 hours per home. The Stanford analysis found that burying an additional 25% of overhead power lines could have cut annual outage totals by 10.8 hours.

  • Buried power lines could cut weather-related outages
    Annual average power outage time for 2017, on a scale from less than one hour (lightest shades) to more than 24 hours (darkest shades). Credit: arXiv (2024). DOI: 10.48550/arxiv.2402.06668
  • Buried power lines could cut weather-related outages
    Undergrounding an additional 25% of power lines could have reduced outages by 10.8 hours (39.7%). Credit: arXiv (2024). DOI: 10.48550/arxiv.2402.06668

California’s 2019 wildfire shutoffs

Amid and in 2019, more than 3 million Californians lost power when utilities preemptively shut down equipment in high-fire-risk areas. The Stanford analysis found that undergrounding an additional 25% of overhead power lines would have cut annual outage totals in the affected area to roughly eight hours from 10.5 hours.

Texas’s 2021 deep freeze

In February 2021, unusually in Texas left 4.5 million homes and businesses without power for just over 19 hours. The researchers found having 25% more power lines underground during this event also could have shortened average outage times by 2.5 hours.

Explore the data

You can view more analysis from the Stanford researchers and explore county-level undergrounding and patterns in an interactive project developed by the Stanford Doerr School of Sustainability in collaboration with TechSoup. The researchers have made their 2020 data on the proportion of underground distribution power lines publicly available through Stanford’s Data Commons for Sustainability.

More information:
Tao Sun et al, Mapping the Depths: A Stocktake of Underground Power Distribution in United States, arXiv (2024). DOI: 10.48550/arxiv.2402.06668

Journal information:
arXiv


Citation:
Buried power lines could cut weather-related outages (2025, November 5)
retrieved 5 November 2025
from https://techxplore.com/news/2025-11-power-lines-weather-outages.html

This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.





Source link

Continue Reading

Tech

Fast, accurate drag predictions could help improve aircraft design

Published

on

Fast, accurate drag predictions could help improve aircraft design


Credit: Pixabay/CC0 Public Domain

Researchers at the University of Surrey have proposed a computational approach that can provide aerodynamic drag data more efficiently during the early stages of aircraft design. It is hoped that AeroMap could help develop safer and more fuel-efficient aircraft.

Drag is the aerodynamic force that opposes an ‘s motion through the air. Being able to predict accurately at an early design stage helps engineers avoid later adjustments that can lead to additional time and cost. Reliable early estimates can also reduce the need for extensive wind tunnel testing or large-scale computer simulations.

AeroMap estimates drag for different wing-body configurations operating at speeds close to the speed of sound. In a study published in Aerospace Science and Technology, researchers show how AeroMap provides datasets up to 10 to 100 times faster than high-fidelity simulations currently on the market, while maintaining good accuracy.

The researchers suggest that such improvements in prediction speed could support the development of more fuel-efficient aircraft configurations by allowing designers to assess a wider range of design options in less time.

“Our goal was to develop a method that provides reliable transonic aerodynamic predictions for a range of configurations, without the high computational cost of full-scale simulations. By providing reliable results earlier in the , AeroMap reduces the need for costly redesigns and repeated wind-tunnel testing.

“It also delivers the level of detail engineers need to refine concepts more efficiently and with greater confidence,” says Dr. Rejish Jesudasan, research fellow at the University of Surrey and lead author of the study.

AeroMap is based on a viscous-coupled full potential method, which combines a reduced form of the Navier–Stokes equations that describe airflow with a model of the thin boundary layer of air that moves along an aircraft’s surface. This approach enables AeroMap to capture the main effects of drag without the high computing demands of more detailed simulations. As a result, it provides a practical tool for the early stages of aircraft design, when engineers need results that are both reliable and rapid.

Many existing models still rely on empirical methods developed several decades ago. Although these remain widely used, they can be less accurate when applied to modern, high-efficiency wing designs. AeroMap has been validated against NASA wind tunnel data, with results showing close agreement between its predictions and experimental measurements, indicating its suitability for sustainable aircraft development.

“Accurately predicting the transonic performance of aircraft configurations, during early concept studies, remains a significant challenge. Previous empirical approaches, based on older datasets, can struggle to capture the behavior of modern high-efficiency wings.

“AeroMap combines established aerodynamic principles in a way that improves the reliability of drag predictions during , helping engineers make better-informed design decisions,” says Dr. Simao Marques.

“We are exploring how AeroMap can be combined with optimization techniques to assess a wider range of wing-body configurations and performance scenarios. This approach could help engineers identify more efficient designs earlier in the process, potentially reducing lifecycle costs and supporting the industry as it works toward future sustainability goals,” says John Doherty.

More information:
Rejish Jesudasan et al, Enhancing rapid drag analysis for transonic aircraft configuration trade studies, Aerospace Science and Technology (2026). DOI: 10.1016/j.ast.2025.110727

Citation:
Fast, accurate drag predictions could help improve aircraft design (2025, November 4)
retrieved 4 November 2025
from https://techxplore.com/news/2025-11-fast-accurate-aircraft.html

This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.





Source link

Continue Reading

Tech

Lay intuition as effective at jailbreaking AI chatbots as technical methods, research suggests

Published

on

Lay intuition as effective at jailbreaking AI chatbots as technical methods, research suggests


Inquiries submitted to an AI chatbot by a Bias-a-Thon participant and the AI-generated answers showing religious bias. Credit: CSRAI / Penn State

It doesn’t take technical expertise to work around the built-in guardrails of artificial intelligence (AI) chatbots like ChatGPT and Gemini, which are intended to ensure that the chatbots operate within a set of legal and ethical boundaries and do not discriminate against people of a certain age, race or gender.

A single, intuitive question can trigger the same biased response from an AI model as advanced technical inquiries, according to a team led by researchers at Penn State.

“A lot of research on AI bias has relied on sophisticated ‘jailbreak’ techniques,” said Amulya Yadav, associate professor at Penn State’s College of Information Sciences and Technology. “These methods often involve generating strings of random characters computed by algorithms to trick models into revealing discriminatory responses.

“While such techniques prove these exist theoretically, they don’t reflect how real people use AI. The average user isn’t reverse-engineering token probabilities or pasting cryptic character sequences into ChatGPT—they type plain, intuitive prompts. And that lived reality is what this approach captures.”

Prior work probing AI bias—skewed or discriminatory outputs from AI systems caused by human influences in the training data, like language or cultural bias—has been done by experts using to engineer large language model (LLM) responses. To see how average internet users encounter biases in AI-powered chatbots, the researchers studied the entries submitted to a competition called “Bias-a-Thon.” Organized by Penn State’s Center for Socially Responsible AI(CSRAI), the competition challenged contestants to come up with prompts that would lead generative AI systems to respond with biased answers.

They found that the intuitive strategies employed by everyday users were just as effective at inducing biased responses as expert technical strategies. The researchers presented their findings at the 8th AAAI/ACM Conference on AI, Ethics, and Society.

Fifty-two individuals participated in the Bias-a-Thon, submitting screenshots of 75 prompts and AI responses from eight generative AI models. They also provided an explanation of the bias or stereotype that they identified in the response, such as age-related or historical bias.

The researchers conducted Zoom interviews with a subset of the participants to better understand their prompting strategies and their conceptions of ideas like fairness, representation and stereotyping when interacting with generative AI tools. Once they arrived at a participant-informed working definition of “bias”—which included a lack of representation, stereotypes and prejudice, and unjustified preferences toward groups—the researchers tested the contest prompts in several LLMs to see if they would elicit similar responses.

Lay intuition as effective at jailbreaking AI chatbots as technical methods
An inquiry submitted by a Bias-a-Thon participant and the generative AI response showing bias toward conventional beauty standards. Credit: CSRAI / Penn State

“Large language models are inherently random,” said lead author Hangzhi Guo, a doctoral candidate in information sciences and technology at Penn State. “If you ask the same question to these models two times, they might return different answers. We wanted to use only the prompts that were reproducible, meaning that they yielded similar responses across LLMs.”

The researchers found that 53 of the prompts generated reproducible results. Biases fell into eight categories: gender bias; race, ethnic and religious bias; age bias; disability bias; language bias; historical bias favoring Western nations; cultural bias; and political bias.

The researchers also found that participants used seven strategies to elicit these biases: role-playing, or asking the LLM to assume a persona; hypothetical scenarios; using human knowledge to ask about niche topics, where it’s easier to identify biased responses; using leading questions on controversial topics; probing biases in under-represented groups; feeding the LLM false information; and framing the task as having a research purpose.

“The competition revealed a completely fresh set of biases,” said Yadav, organizer of the Bias-a-Thon. “For example, the winning entry uncovered an uncanny preference for conventional beauty standards. The LLMs consistently deemed a person with a clear face to be more trustworthy than a person with facial acne, or a person with high cheekbones more employable than a person with low cheekbones.

“This illustrates how average users can help us uncover blind spots in our understanding of where LLMs are biased. There may be many more examples such as these that have been overlooked by the jailbreaking literature on LLM bias.”

The researchers described mitigating biases in LLMs as a cat-and-mouse game, meaning that developers are constantly addressing issues as they arise. They suggested strategies that developers can use to mitigate these issues now, including implementing a robust classification filter to screen outputs before they go to users, conducting extensive testing, educating users and providing specific references or citations so users can verify information.

“By shining a light on inherent and reproducible biases that laypersons can identify, the Bias-a-Thon serves an AI literacy function,” said co-author S. Shyam Sundar, Evan Pugh University Professor at Penn State and director of the Penn State Center for Socially Responsible Artificial Intelligence, which has since organized other AI competitions such as Fake-a-thon, Diagnose-a-thon and Cheat-a-thon.

“The whole goal of these efforts is to increase awareness of systematic problems with AI, to promote the informed use of AI among laypersons and to stimulate more socially responsible ways of developing these tools.”

More information:
Hangzhi Guo et al, Exposing AI Bias by Crowdsourcing: Democratizing Critique of Large Language Models, Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society (2025). DOI: 10.1609/aies.v8i2.36620

Citation:
Lay intuition as effective at jailbreaking AI chatbots as technical methods, research suggests (2025, November 4)
retrieved 4 November 2025
from https://techxplore.com/news/2025-11-lay-intuition-effective-jailbreaking-ai.html

This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.





Source link

Continue Reading

Trending