Connect with us

Tech

AI method reconstructs 3D scene details from simulated images using inverse rendering

Published

on

AI method reconstructs 3D scene details from simulated images using inverse rendering


Layout generation. a, Images for two scenes observed by a single camera. b, Test-time optimized inverse rendered objects. c, BEV layouts of the scenes. In the BEV layout (a common representation for autonomous driving tasks), black boxes represent the ground truth and colored boxes represent predicted BEV boxes. Credit: Nature Machine Intelligence (2025). DOI: 10.1038/s42256-025-01083-x

Over the past decades, computer scientists have developed many computational tools that can analyze and interpret images. These tools have proved useful for a broad range of applications, including robotics, autonomous driving, health care, manufacturing and even entertainment.

Most of the best performing computer vision approaches employed to date rely on so-called feed-forward neural networks. These are computational models that process input images step by step, ultimately making predictions about them.

While some of these models were found to perform well when tested on the data they analyzed during training, they often do not generalize well across new images and in different scenarios. In addition, their predictions and the patterns they extract from images can be difficult to interpret.

Researchers at Princeton University recently developed a new inverse rendering approach that is more transparent and could also interpret a wide range of images more reliably. The new approach, introduced in a paper published in Nature Machine Intelligence, relies on a generative artificial intelligence (AI)-based method to simulate the process of image creation, while also optimizing it by gradually adjusting a model’s internal parameters.

“Generative AI and neural rendering have transformed the field in recent years for creating novel content: producing images or videos from scene descriptions,” Felix Heide, senior author of the paper, told Tech Xplore. “We investigate whether we can flip this around and use these generative models for extracting the scene descriptions from images.”







Video of tracking results of the team’s method. A demonstration of the performance of our proposed tracking method based on inverse neural rendering for a sample of diverse scenes from the nuScenes dataset and the Waymo Open Dataset. We overlay the observed image with the rendered objects through alpha blending with a weight of 0.4. Object renderings are defined by the averaged latent embeddings zk,EMA and the tracked object state yk. Credit: Nature Machine Intelligence (2025). DOI: 10.1038/s42256-025-01083-x

The new approach developed by Heide and his colleagues relies on a so-called differentiable rendering pipeline. This is a process for the simulation of image creation, relying on compressed representations of images created by generative AI models.

“We developed an analysis-by-synthesis approach that allows us to solve vision tasks, such as tracking, as test-time optimization problems,” explained Heide. “We found that this method generalizes across datasets, and in contrast to existing supervised learning methods, does not need to be trained on new datasets.”

Essentially, the method developed by the researchers works by placing models of 3D objects in a virtual scene depicting real world settings. These models of objects are generated by a generative AI based on random sample of 3D scene parameters.

“We then render all these objects back together into a 2D image,” said Heide. “Next, we compare this rendered image with the real observed image. Based on how different they are, we backpropagate the difference through both the differentiable rendering function and the 3D generation model to update its inputs. In just a few steps, we optimize these inputs to make the rendered match the observed images better.”

  • A new generative model-based inverse rendering approach for computer vision and image processing
    Optimizing 3D models through inverse neural rendering. From left to right: the observed image, initial random 3D generations, and three optimization steps that refine these to better match the observed image. The observed images are faded to show the rendered objects clearly. The method effectively refines object appearance and position, all done at test time with inverse neural rendering. Credit: Ost et al.
  • A new generative model-based inverse rendering approach for computer vision and image processing
    Generalization of 3D multi-object tracking with Inverse Neural Rendering. The method directly generalizes across datasets such as the nuScenes and Waymo Open Dataset benchmarks without additional fine-tuning and is trained on synthetic 3D models only. The observed images are overlaid with the closest generated object and tracked 3D bounding boxes. Credit: Ost et al.

A notable advantage of the team’s newly proposed approach is that it allows very generic 3D object generation models trained on synthetic data to perform well across a wide range of datasets containing images captured in real-world settings. In addition, the renderings produced by the models are far more explainable than those produced by conventional rendering tools based on feed-forward machine learning models.

“Our inverse rendering approach for tracking works just as well as learned feed-forward approaches, but it provides us with explicit 3D explanations of its perceived world,” said Heide.

“The other interesting aspect is the generalization capabilities. Without changing the 3D generation model or training it on new data, our 3D multi-object tracking through Inverse Neural Rendering works well across different autonomous driving datasets and object types. This can significantly reduce the cost of fine-tuning on new data or at least work as an auto-labeling pipeline.”

This recent study could soon help to advance AI models for computer vision, improving their performance in real-world settings while also increasing their transparency. The researchers now plan to continue improving their method and start testing it on more computer vision-related tasks.

“A logical next step is the expansion of the proposed approach to other perception tasks, such as 3D detection and 3D segmentation,” added Heide. “Ultimately, we want to explore if inverse rendering can even be used to infer the whole 3D scene, and not just individual objects. This would allow our future robots to reason and continuously optimize a three-dimensional model of the world, which comes with built-in explainability.”

Written for you by our author Ingrid Fadelli,
edited by Gaby Clark, and fact-checked and reviewed by Robert Egan—this article is the result of careful human work. We rely on readers like you to keep independent science journalism alive.
If this reporting matters to you,
please consider a donation (especially monthly).
You’ll get an ad-free account as a thank-you.

More information:
Julian Ost et al, Towards generalizable and interpretable three-dimensional tracking with inverse neural rendering, Nature Machine Intelligence (2025). DOI: 10.1038/s42256-025-01083-x.

© 2025 Science X Network

Citation:
AI method reconstructs 3D scene details from simulated images using inverse rendering (2025, August 23)
retrieved 23 August 2025
from https://techxplore.com/news/2025-08-ai-method-reconstructs-3d-scene.html

This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.





Source link

Tech

If a Garmin Is Too Expensive, Consider Suunto’s Latest Adventure Watch

Published

on

If a Garmin Is Too Expensive, Consider Suunto’s Latest Adventure Watch


It’s always pleasing to see an array of physical buttons, and you get sizable ones too. You’re not going to miss these wide flat ones even when picking the pace up. The silicone strap has a nice stretch to it and while the button clasp is a bit awkward to get into place, this watch does not budge.

Suunto has jumped on the flashlight trend, with an LED light strip sat on the front of the case. You can adjust brightness levels and there’s SOS and alert modes to emit a very noticeable pulsating light pattern. This is a light I found useful rooting around indoors as well as on nighttime outings.

The biggest change is the introduction of a 1.5-inch, 466 x 466 AMOLED display. This replaces the dull, albeit very visible, memory-in-pixel (MIP) display. Suunto also ditched the solar charging that did require spending a significant amount of time outside to reap its battery benefits.

Adding AMOLED screens to outdoor watches has been contentious. The older MIP displays are just more power-efficient. The Vertical 2 is down by about 10 days from the older Vertical for what Suunto calls daily use.

Still, even if you’re putting its tracking and mapping features to use, you’re not going to be reaching for the charger every few days. After two hours of tracking in optimal GPS mode, the battery only dropped by 2 to 3 percent. The battery drop outside of tracking is also small and the standby performance is excellent as well.

Software Updates

Photograph: Michael Sawh

A more streamlined set of smartwatch features helps reserve battery for when it really matters. Unfortunately, I probably got better battery life because you don’t get phone notifications or responses if it’s paired to an iPhone instead of an Android. There’s also no onboard music player, but you do get a pretty slick set of music playback controls that are accessible during tracking.



Source link

Continue Reading

Tech

Police do not have to explain to lawyer Fahad Ansari why they seized his phone data, says court | Computer Weekly

Published

on

Police do not have to explain to lawyer Fahad Ansari why they seized his phone data, says court | Computer Weekly


Police do not have to give a lawyer who was stopped, questioned and had his work mobile phone seized for forensic examination reasons for their actions, the UK’s high court has ruled.

The decision means that lawyers can be subject to counter-terrorism powers and have their privileged communications extracted and examined by the state, without having the right to know the case against them, said advocacy group Cage.

Fahad Ansari, who acts for Hamas in a legal appeal to have its proscribed status in the UK overturned, was stopped by police under Schedule 7 of the terrorist act while returning from holiday with his family last year.

The case is believed to be the first targeted use of Schedule 7 powers, which allow police to stop and question people and seize their electronic devices without the need for suspicion, against a practising solicitor.

The high court ruled on 4 March that police may present evidence about the reasons stopping Ansari in a closed court in front of a special advocate without Ansari or his lawyers being present – preventing Ansari or his legal team from learning the reasons why he was stopped.

Lawyers for Ansari argued the lawyer was entitled to be given a sufficient “gist” of the police’s case against him to enable him to disprove the police’s case, even if doing so would be damaging to national security.

Privileged material

Hugh Southey KC told the court in October 2025 that Ansari’s work phone contained data going back 15 years, including privileged material relating to his clients, and that any data extracted by the police should be deleted.

Ansari, an Irish citizen, argues that he was unlawfully stopped, detained and questioned under Schedule 7 of the Terrorism Act when he disembarked from a ferry with his family at Holyhead after visiting relatives in Ireland in August 2025.

The court was told last year that the phone contains details of at least 3,000 contacts, voice notes, memos, case papers, search terms and metadata, the overwhelming proportion of which is likely to be legally protected.

Justice Chamberlian found in a judgment published today the question was not whether any allegations made against Ansari by police in closed hearings were true, but whether police had a lawful basis for stopping and searching the lawyer at the time the search was carried out.

He found in a 15-page ruling that the use of Schedule 7 powers against Ansari to question him and seize his phone does not require any allegation to have been made against him, and that the seizure and retention of his personal information does not affect Ansari’s legal position.

The judge found that there were “substantial protections” in place to protect the integrity of legally privileged information, and that even if legally privileged material could be used against third parties, which it could not, they would enjoy the “full panoply of procedural rights”.

Ansari said he handed over the password to his phone after police warned him that to fail to do so would be an arrestable offence. He said that police also questioned him about Palestine Action, a direct action protest group that was proscribed under the Terrorism Act 2000, though Ansari has no connection with the group.

South Wales Police, which is responsible for counter-terrorism in Wales, has denied that Ansari was stopped because of his political views, and maintains that asking him questions about proscribed organisations is not unlawful.

Ansari, a registered freelance solicitor, became consultant at Duncan Lewis Solicitors, where he specialises in national security and complex human rights cases, after training at Fisher Meredith LLP and Birnberg Peirce.

Speaking after the judgement, Ansari said he would challenge the judge’s order that the police should not disclose their reasons for stopping him in open court.

“Seven months on, I remain in the dark about why counter-terrorism police detained and interrogated me and continue to examine the contents of my work phone,” he added. “I am exploring all options to challenge this dangerous precedent.”

Commenting on the case, Anas Mustapha, head of public advocacy at Cage, said that allowing secret evidence was a “thin end of the wedge” that could undermine justice. “Once courts accept that the state can accuse someone without revealing the accusation, the foundations of justice begin to collapse,” he added.

“The legal profession now faces a serious question: whether it will continue to accommodate secret courts through mechanisms like the special advocate system, or whether it will begin the difficult work of rolling back a process that has steadily eroded open justice for more than two decades,” said Mustapha.



Source link

Continue Reading

Tech

These $500 Windows Laptops Show That the MacBook Neo Has Serious Competition

Published

on

These 0 Windows Laptops Show That the MacBook Neo Has Serious Competition


Today, Apple announced its new budget MacBook. At $599, it looks seriously impressive. While I haven’t tested its performance, battery life, or display just yet, it may end up being hard to beat at that price based on some of the specs alone.

But that doesn’t mean the competition isn’t there. I want to recommend a couple of Windows laptops deals that offer various advantages over the MacBook Neo, showing where the Neo has both strengths and weaknesses.

First, check out this Asus Vivobook 14, a laptop I’ve been happy to recommend as a budget computer for the past year. In many ways, this is the Windows version of a laptop like the MacBook Neo. It uses a highly-efficient ARM chip, the Qualcomm Snapdragon X, meaning it gets great battery life and performs admirably in daily tasks. It’s not quite as thin or light as the MacBook Neo, but it’s fairly portable for a laptop at this price.

Asus

Vivobook 14 (X1407QA)

Unlike the MacBook Neo, the Vivobook 14 comes with 16 GB of RAM and 512 GB of storage. That’s twice what you get in the MacBook Neo’s starting configuration. Right now, this configuration of the Vivobook 14 is on sale for $539. That’s a killer deal for those specs. It even comes with a healthier mix of ports, including HDMI, two USB-A, one USB-C, and a headphone jack. That also means it can support two external displays unlike the MacBook Neo, which can only handle just one.

Don’t get me wrong—I’m not at all saying the Vivobook 14 is a slam dunk over the MacBook Neo. Based on specs alone, I know the Vivobook 14 is a serious step down when it comes to the display. It’s less sharp, stretched across a larger screen, and the color performance isn’t so good. The Vivobook 14 maxes out at 280 nits, whereas Apple says the MacBook Neo can go all the way up to 500 nits. I have a hunch that the MacBook Neo will deliver a much better display in just about every regard.

There’s also the touchpad. It’s a little clunky to use, which is typical of budget Windows laptops. This is just a guess—but the touchpad on the MacBook Neo will likely feel smoother. It’s a mechanical trackpad (unlike the MacBook Air’s haptic feedback trackpad), but Apple has almost never made a bad trackpad.

If you’re not convinced by the Asus Vivobook 14, I’d also recommend the HP OmniBook 5, which is currently on sale for $500 and uses the same Snapdragon X chip. While it only has 256 GB of storage, it has a much better screen than the Vivobook 14, using an OLED display. It’s not any brighter than the Vivobook 14, but it gives you far better color performance and contrast. It’s also just 0.50 inches thick, matching the MacBook Neo exactly in portability.



Source link

Continue Reading

Trending