Open Access
Issue
A&A
Volume 706, February 2026
Article Number A124
Number of page(s) 12
Section Numerical methods and codes
DOI https://doi.org/10.1051/0004-6361/202449845
Published online 09 February 2026

© The Authors 2026

Licence Creative CommonsOpen Access article, published by EDP Sciences, under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

This article is published in open access under the Subscribe to Open model. This email address is being protected from spambots. You need JavaScript enabled to view it. to support open access publication.

1 Introduction

Cosmological simulations allow us to confront existing theoretical models describing the Universe’s initial state, contents, and the physical processes governing its evolution with increasingly detailed observations of the cosmos, such as the data coming from James Webb Space Telescope (Gardner et al. 2006) or the upcoming Euclid survey (Laureijs et al. 2011). These simulations describe the evolution over time of matter on large scales, from N-body simulations modeling only the gravitational interaction of massive particles in expanding space (e.g., the Millenium simulation, Boylan-Kolchin et al. 2009) to more complex simulations (e.g., Illustris, Vogelsberger et al. 2014, IllustrisTNG, Nelson et al. 2019, EAGLE, Crain et al. 2015) incorporating hydrodynamics and/or physical processes on relatively small scales modeling the evolution of baryonic matter.

When considering the N-body problem, we must keep in mind that there is no known closed-form expression that can be used to directly calculate the positions and velocities of the particles at any given time. Indeed, the gravitational force between each pair of DM particles in simulations depends on their positions and masses, while the resulting motion is influenced by the collective interactions among all the particles, entailing a highly nonlinear evolution of the system over time. Current numerical methods tend to approach this problem by starting with initial conditions of particle positions and velocities and iteratively resolving the equations of motion in time steps that are small enough that a linear evolution sufficiently approximates the true motion of particles. While gravitational interactions between particles scale naively as O(N2), many efficient algorithms, such as particle-mesh (PM - O(N + M log M), Efstathiou et al. 1985), tree methods (O(N log N), Barnes & Hut 1986), and hybrid approaches such as P3M (∼ O(N log N)), have been developed to reduce this cost substantially. Nevertheless, high-resolution simulations with large particle counts remain computationally intensive, particularly when one aims to capture both the large-scale structure and small-scale dynamics simultaneously. This imposes practical limits on the resolution, scale, or number of realizations that can be feasibly simulated. Although faster fully analytical approaches (Shandarin & Zeldovich 1989; Kitaura & Heß 2013) and semi-analytical simulations that combine traditional simulation methods and analytical approximations (Monaco et al. 2002; Tassev et al. 2013), relying on first- or second-order perturbation theory have been proposed, they cannot address the highly nonlinear stages of structure formation.

More recently, machine learning (ML) approaches have proved efficient in completing tasks for which prescriptive analytical approaches were either slow, overly approximative, or nonexistent. Examples in astronomy include classification, such as galaxy types (Shamir 2009; Freed & Lee 2013; Biswas & Adlak 2018) or cosmic web structure types (Aragon-Calvo 2019), as well as redshift estimation (Henghes et al. 2021; Rastegarnia et al. 2022; D’Isanto & Polsterer 2018; Henghes et al. 2022), detection of various objects of interest (Rezaei et al. 2022b; Vafaei Sadr et al. 2019; Jia et al. 2023; González et al. 2018), anomalies (Reyes & Estévez 2020; Dere et al. 2021; Villar et al. 2021), or physical effects such as gravitational lensing (Rezaei et al. 2022a; Wilde et al. 2022) or Sunyaev-Zel’dovich (SZ) effects (Tanimura et al. 2022; Bonjean 2020), deblending sources (Burke et al. 2019; Hausen & Robertson 2022; Hansen et al. 2022), determining baryonic matter properties given dark matter distribution (Jo & Kim 2019; Wu & Kragh Jespersen 2023; Chittenden & Tojeiro 2023; Bonjean et al. 2019), simulation augmentation (Sweere et al. 2022; Kodi Ramanah et al. 2020; Agarwal et al. 2018), and others. More specifically, when considering the time evolution of complex physical systems, several neural network-based experiments (Feng 2023; Wiewel et al. 2019; Liu 2023; Humbird et al. 2018; Sanchez-Gonzalez et al. 2020) have demonstrated promising results in terms of emulating simulations.

Concerning cosmological simulations in particular, recent years have seen the emergence of a range of machine learning (ML) approaches designed to accelerate the generation of N-body simulations, offering speed-ups of several orders of magnitude over traditional numerical methods. A particularly successful class of these models is rooted in a Lagrangian framework, first introduced with the seminal D3M model (He et al. 2019) and subsequently refined to impressive levels of performance in later works (Alves de Oliveira et al. 2020; Jamieson et al. 2023; Jamieson et al. 2025; Prost et al. 2025). These models typically take the initial positions of particles, together with a linear approximation of their evolution (e.g., the Zel’dovich approximation) and learn to predict the final nonlinear displacement by estimating the residual between this approximation and the true evolution. Beyond the direct emulation of structure formation, this framework has also been adapted for broader use cases, such as translating simulations across different cosmological parameters or initial conditions (Saadeh et al. 2024; Giusarma et al. 2023). These methods have proven highly effective in recovering the small-scale clustering and halo structures characteristic of dark matter evolution. Finally, hybrid strategies that combine ML approximations with partial N-body integration, such as the COCA model (Bartlett et al. 2025), have further demonstrated the potential of ML-assisted simulation pipelines, achieving improvements in both speed and accuracy in cosmological modeling.

While these approaches are optimized for the specific task of accelerating particle-based simulations, complementary methods can be explored through a Eulerian perspective, namely, by treating the matter distribution as a continuous field rather than a set of discrete particles. Although less optimal than Lagrangian approaches in this context, focusing on Eulerian fields offers a simplified, but generalizable framework. This can help inform other ML-based inference challenges where only Eulerian data are available; for instance, it naturally aligns with the structure of specific types of observational data, such as weak lensing maps, 21 cm intensity maps, or cosmic microwave background lensing fields, all of which are inherently gridded or projected fields. Similarly, many hydrodynamical simulations represent gas and baryonic components in Eulerian form, making field-based ML architectures well-suited for cross-domain applications (Hsu et al. 2025; Luo et al. 2025; Gondhalekar et al. 2025; Hiegel et al. 2023; Conceição et al. 2024a,b; Li et al. 2022).

Moreover, the two paradigms are complementary in their scaling limitations. Lagrangian methods are typically constrained by the number of particles that can be simulated or stored, while Eulerian approaches are limited by voxel resolution and binning artifacts. There exists an intermediate regime, where extremely fine particle simulations are prohibitively expensive for particle-based ML, yet still accessible to Eulerian models through lower-resolution gridding. In such cases, the field-based approach enables analysis of structure formation at finer graining and with potentially better memory efficiency.

In this context, we propose a proof-of-concept model aimed at predicting the time evolution of cosmic density fields in Eulerian space. Rather than offering a drop-in replacement for Lagrangian-based emulators, this work explores the feasibility of using generic continuous fields as inputs, with the long-term goal of developing architectures whose application can be generalized to a range of cosmological and astrophysical datasets.

To this end, we have introduced an autoencoder (AE) architecture with an unconventional twist: it incorporates generative adversarial network (GAN) components into its architecture (Ullmo et al. 2021). In the following, we refer to this AE as timewarper (TW). This design encourages the model to balance two objectives: an accurate reconstruction of individual structures and the preservation of the overall statistical properties of the density field. While this hybrid approach currently exhibits a tradeoff between spatial precision and statistical fidelity compared to standard autoencoders, we argue that such tradeoffs are an essential aspect of the design space and exploring them may ultimately lead to more robust emulation strategies.

In Sect. 2, we outline the setup of our experiment, wherein we describe the architecture and training of our TWs, the data used for training, and the metrics by which we measure the quality of our results. In Sect. 3.1, we begin by showing results for a TW trained to predict simply from an input density map; we refer to it as a “baseline TW”. In Sect. 3.2, we introduce a TW trained with additional input information in the form of the density field’s associated velocity field, which shows significant improvement on the baseline approach. We label it here as “velocities TW”. In Sect. 4, we interpret our results and discuss other possible optimization methods. Finally, we present our conclusions in Sect. 5.

2 Setup

Our goal in this work is to create a network capable of forecasting the evolution of a simulation-derived data set (a discrete 2D or 3D density map generated from a 2D or 3D N-body simulation) from previous instances (when z > 0) to the current time (z = 0). Thus, we have built a TW that takes a datum in the form of a simulated density field at a given fixed redshift (i.e., z = 0, 1, 2 or 3) as input and is tasked with recovering its corresponding simulation at z = 0, by minimizing the distance (see Eq. (1)) between its output and target datum.

While this work focuses on inputs at z = 3, 2, 1, and 0, incorporating earlier snapshots, such as the initial conditions (z = 99), remains a potential avenue for future investigation. In a second step, we provide the TW an associated velocity field, in an effort to improve the AE’s predictions. We present the networks, data, and metrics used for our work in brief below.

2.1 Networks

2.1.1 GANs in a nutshell

Overall, GANs are notorious for their ability to emulate data from a training set, typically in the form of images. They consist of a pair of deep neural networks that are trained simultaneously in an adversarial manner (see top diagram in Fig. 1). The first, a generator, intakes a vector with randomly distributed values and is trained to output realistic data by “fooling” the second, a discriminator, which is tasked with separating genuine training data from data generated by the generator. Throughout training they become gradually better at their respective tasks, simultaneously increasing the difficulty of the other’s task. Once trained the generator is able to output a training set-like datum from a low-dimension input vector, while the discriminator is capable of distinguishing high-level features in data to suit its task. We made use of these properties to build our AEs.

Thumbnail: Fig. 1 Refer to the following caption and surrounding text. Fig. 1

Architecture of the timewarper. A trained GAN’s generator is used as a readily built decoder. Only the encoder’s weights are changed during training. The same GAN’s truncated discriminator is used to compute a perceptual loss (see Eq. (1)).

2.1.2 Timewarpers

Our work relies on the use of AEs based on these convolutional GANs (the TW, displayed in the bottom panel of Fig. 1). The AEs consist of an encoder which intakes a datum at a given redshift z ≥ 0 and encodes it into a vector within a low-parameter latent space with greater semantic significance. This encoded vector is then fed into a decoder tasked with recovering the original datum at redshift z = 0, by minimizing the loss described below (Eq. (1)).

We aimed to constrain the decoder to output data that would be statistically consistent with true z = 0 data. To this effect, we relied on a GAN that had been previously trained to generate simulation-like data at z = 0. First, we used its generator (designed to output simulation-like data at z = 0) as the AE’s decoder and locked its weights during training. Second, we used its truncated discriminator as part of the AE’s loss1, LAE=Δ(x,x~),Mathematical equation: $\begin{equation*} L_{A E}=\Delta(x, \tilde{x}), \end{equation*}$(1)

where x is the output datum (i.e., the AE’s prediction of input datum at z = 0), x~Mathematical equation: $\tilde{x}$ is the ground truth (true input datum at z = 0) and Δ is the ℓ2 difference in the truncated discriminator’s latent space.

This type of loss, termed perceptual loss (Johnson et al. 2016), integrates a discriminator’s ability to emphasize the features and patterns of data rather than relying solely on a basic pixel-by-pixel comparison. The TW architectures and parameters (Table 1) are derived from previously trained 2D and 3D GANs, which have been adapted, in turn, from a publicly available DCGAN implementation2. Notably, the latent dimension was inherited from the original 2D model and doubled for the 3D version to account for the higher input dimensionality. This scaling was empirically found to provide stable training and satisfactory predictive accuracy; larger latent spaces were summarily tested but did not appear to noticeably improve performance. Nonetheless, further exploration might be worthwhile in future work.

Notably, the model does not have any mass conservation constraints at present. We did not apply a mass normalization to the output density data, as a simple post-processing normalization would erroneously shift all of the amplitudes of the predicted structures. One possible improvement would be to introduce a normalization constraint during training, either by adding a dedicated loss term or by normalizing the data immediately before computing the existing loss.

Table 1

Architectures of the 2D and 3D TW.

Thumbnail: Fig. 2 Refer to the following caption and surrounding text. Fig. 2

Example of simulation-issued data: a 2D image from a 2D simulation (left) and a 3D cube (right) from a 3D simulation. These data are built by dividing an N-body simulation snapshot into pixels (or voxels) and counting the particles within each, creating a discrete density map. The map is then smoothed with a Gaussian filter and log-transformed, allowing cosmic structure to stand out starkly. This process creates data that are compatible with convolutional neural networks, which specialize in feature detection.

2.2 Data

2.2.1 Simulations

Our networks were trained on data built from both 2D and 3D N-body simulations. The 2D data (Fig. 2, left) provide simpler conditions (i.e., a lower parameter problem in terms of the particle degree of freedom, datum size, and network size), leading to more optimal results for a more ideal proof of concept and a point of comparison for 3D results. The training set is built from 1000 snapshots from nbody2D3 simulations, a set of 2D particlemesh N-body simulations of side 100 Mpc/h with 5122 particles using the standard ΛCDM cosmology. They were run and saved for redshifts z = 0, 1, 2, and 3.

The training set for the 3D data of interest to our study (Fig. 2: right panel) was built from a single snapshot of a GADGET2 simulation (Springel et al. 2001; Springel 2005) of a side of 100 Mpc/h with 5123 particles. We assumed a standard Λ CDM cosmology, with cosmological parameters Ωm = 0.32, ΩΛ = 0.69, σ8 = 0.83, ns = 0.96, and H0 = 0.68, from Planck 2018 (Planck Collaboration VI 2020) and saved at redshifts z = 0, 1, 2, and 3.

2.2.2 Training data

The individual data were built in the following way: each (2D or 3D) snapshot is first made into a discrete density map (256 × 256 or 256 × 256 × 256, respectively). For the 3D data, these were constructed following a crude nearest grid point (NGP) procedure with subsequent smoothing using a Gaussian filter with a standard deviation of 3 pixels. This smoothing reduces sharp discontinuities in the raw density field that arise from coarse binning noise. The 2D data were built using a Delaunay tessellation field estimator (DTFE, Aragon-Calvo (2021)). They were subsequently log-transformed for compatibility with the neural network. The choice for the 3D data was made for simplicity, but further work would likely benefit from using more accurate interpolation schemes such as the 2D data’s DTFE, cloud-in-cell, triangular-shaped cloud (Hockney & Eastwood 2021), or simplex-in-cell (Abel et al. 2012; Hahn et al. 2013).

From these data, smaller sub-arrays (of side 50 Mpc/h and 128 pixels for 2D and 25 Mpc/h and 64 voxels for 3D) were extracted to make up the training sets. This gives us respectively 5.108 and 8.108 possible sub-arrays. Given this high number of possibilities and the redundancy among them, we did not reason in terms of epochs for the training time (where one epoch corresponds to the network processing the entire training set). Instead, we quantified the training time in terms of gradient updates performed on data batches (of size 200 in our case), with each training batch randomly selected from all possible subarrays.

In a second part of our work, we additionally provided the density fields’ associated velocity fields as input for the 3D case. To build the velocity field, we followed a similar approach to our consideration of the density field. Dividing the snapshot space into 3 × 256 × 256 × 256 voxels, we computed three 3D averaged velocity fields for each direction (x, y, and z), by summing the velocities of the particles in each voxel and dividing the sum by the number of particles. Upon inspecting the resulting cubes, we find that cosmic structures are visually apparent without need for log-transformation. Thus, we can simply apply a normalization of voxel values, v=v/N.Mathematical equation: $v^{'}=v/N.$(2)

Here, v is the new velocity and N is fixed such that |v|max ≲ 1. Next, we can apply the same smoothing as for the 3D density to obtain our final three 256 × 256 × 256 velocity fields for each (x, y, and z) direction. By combining the density field with the velocity fields constructed in this way, we can construct an array of size 256 × 256 × 256 × 4 and use it to extract smaller arrays of size 64 × 64 × 64 × 4.

For the 3D case, the training, validation, and test sets were constructed from three distinct simulations. The validation set consists of 200 randomly selected sub-cubes from the validation simulation, while the test set comprises of 64 sub-cubes forming the test simulation. In the results section, we reconstructed the full cube by assembling the 64 predicted sub-cubes and compared the result with the corresponding full cube at the target redshift.

For the 2D case, we used 1000 simulations in total: 800 for training, 100 for validation, and 100 for testing. Each simulation is divided into four sub-arrays. Consequently, the test set contains 400 predicted subarrays, which were reassembled into 100 full-size arrays for evaluation.

2.3 Metrics

To quantify the efficiency of the TW’s predictions, we relied on three metrics: the power spectrum and the cross-correlation coefficient, which assess the spatial correspondence between predicted and target data; along with the Dice coefficient (see Ullmo et al. (2021) for details), which measures the accurate recovery of dense structures in the predictions. While the power spectrum provides a global view of the statistical consistency between predicted and target data, the cross-correlation and Dice coefficients additionally evaluate the pairwise agreement of individual targets and predictions (in other words, the TW’s accuracy). The cross-spectrum of a data pair A and B for a frequency, k, is given by PAB(k)=AkBkk,|k|=k,Mathematical equation: $P_{A B}(k)=\left\langle A_{\mathbf{k}} B_{\mathbf{k}}^{*}\right\rangle_{\mathbf{k},|\mathbf{k}|=k},$(3)

where Ak and Bk are the pair’s discrete Fourier transform elements.

From this, we get the power spectrum of datum A: PAA(k).

Since the power spectrum’s distribution is approximately uniform in logarithmic space, when averaging over the test set we use the geometric mean, exp (⟨ln P⟩), as a more representative central value. Variability measures (e.g., the geometric standard deviation) were considered in our study, but omitted from the figures for clarity.

Additionally, we can compute the cross-correlation coefficient for pair A and B, r(k)=PAB(k)PAA(k),PBB(k).Mathematical equation: $r(k)=\frac{P_{A B}(k)}{\sqrt{P_{A A}(k), P_{B B}(k)}}.$(4)

The Dice coefficient for a data pair A and B is expressed as OAB(t)=NAB(t)NAB(t),Mathematical equation: $O_{A B}(t)=\frac{N_{A \cap B}(t)}{N_{A \cup B}(t)},$(5)

where NAB(t) is the number of pixels or voxels whose value is above t for both A and B and NAB(t) is the number of pixels or voxels whose value is above the threshold, t, in either A or B at a given position in an image or cube. We computed this across a fixed set of thresholds corresponding to percentiles of the total pixel value distribution. This metric quantifies the degree of pixel-wise overlap between the predicted and true high-density regions. Assuming the pixel value distributions of the predicted and target maps are broadly similar, the Dice coefficient of an unrelated pair of data at a given threshold is expected to approximate the fraction of pixels above that percentile. In contrast, a perfect reconstruction would yield a Dice coefficient of 1.0 across all percentiles. This is an ideal, but challenging scenario given the complexity of exact pixel-level alignment in continuous fields.

3 Results

3.1 Baseline TW results

We input a datum in the form of a density field at redshift z ≥ 0 and trained the TW to output the corresponding density field at z = 0. Here, “corresponding” refers to the same comoving region within snapshots at the input and target redshifts (zinizfin). The network was optimized by minimizing the perceptual loss (see Eq. (1)). We carried out this procedure for both 2D and 3D data.

3.1.1 2D images

We first focus on the outcome of the TW trained to recover density fields at redshift z = 0 for the set of 2D simulation images, from input density fields at redshifts varying from z = 0 to z = 3 by steps of Δ z = 1. We note that the “predictions” from z = 0 to z = 0 consist of simple replicative autoencoding, which define the limitations of our setup, the discrepancy between input and output in this case being strictly due to encoding loss, rather than inaccurate prediction.

We compared our results on predictions from higher redshifts to this z = 0 reference. The task is expected to be harder the higher the input z; indeed, given that the development of structure in matter is a highly nonlinear process, we can expect that the farther away in time the target is from the output, the farther the density field will steer from an easily approximated linear evolution. For all z>0 ⟶ z = 0 input-output pair, a distinct TW is individually run for 15 k gradient updates (no significant improvement observed for longer training) and the best weights are recovered by finding the minima of validation losses computed every 1000 update for each input z(0, 1, 2, and 3), respectively at 10, 15, 10, and 12 k updates.

We illustrate the baseline TW’s performance in recovering z = 0 from different redshifts with a set of six simulation images taken at random from the test set (Fig. 3). These data are taken at various redshifts (left block) and used as input for the trained TWs to predict their z = 0 equivalent (upper right). The predicted data are shown in the right block.

We find that regardless of input redshift, the TWs are successful in recovering z = 0. The larger and denser structures are consistently well recovered while finer details exhibit more variability. Initially, it is difficult to discern any perceptible changes in performance based on varying input redshifts; indeed, structures appear to be reliably predicted even when inferring from higher redshifts. However, upon closer examination, we note that as the input redshift increases, there is a gradual loss of detail in the predicted structures. This loss is manifested as the merging of certain large structures, shifts, or disappearance of finer structures, as well as alterations in the positions of overdense regions.

Inspecting the average power spectrum (PS) and crosscorrelation coefficient r(k) of the predicted data (Fig. 4, left), we found that the TWs recover similar statistics, regardless of the input redshift. An upward shift of the PS, together with r(k) ≈ 1 at low k, suggests that the predictions are coherent with the ground truth on large scales but slightly overestimate the total mass; for all metrics, standard deviations over the test set (left out of the figure for clarity) are relatively small compared to the measured averages for both target and predictions. The correlation begins to decline around k ≃ 3 × 10−2 h Mpc−1 and reaches r(k) = 0 between k ≃ 2 → 4 × 10−1 h Mpc−1 for the z = 3 → 0 case. Over this range, the PS curves overlap well, indicating that the model reproduces small-scale amplitudes but misplaces structures, leading to phase decorrelation. Since structures at these scales arise from highly nonlinear processes, this suggests that the model, while accurately capturing large-scale linear evolution, struggles more with the nonlinear regime. However, as some decorrelation at high k is also visible in the simple z = 0 → 0 autoencoding case, part of this limitation likely stems from the encoder itself, with the latent space not perfectly representing all possible configurations at these scales.

The Dice coefficient (Fig. 4, right) shows that predictions across all redshifts align with the base z = 0 ⟶ 0 autoencoding, but each increment in input z yields a slightly lower coefficient at all pixel thresholds. This reflects the growing challenge of prediction as z increases. From this first test on 2D data, we can conclude that (in this simple case at least) the TW can effectively predict structure evolution over the tested time spans. Its accuracy is seen to improve as the input redshift approaches the target.

Thumbnail: Fig. 3 Refer to the following caption and surrounding text. Fig. 3

Six images from the 2D simulations at various redshifts (left), and their equivalent predictions of redshift z = 0 (right) inferred by the baseline TW. The true z = 0 simulation images are shown above the predicted images (upper right) for comparison.

Thumbnail: Fig. 4 Refer to the following caption and surrounding text. Fig. 4

Spectra, cross-correlation coefficient and Dice coefficient for baseline timewarper on 2D data. (a) Average power spectra from predictions from input redshifts z = 3 → 0 (blue scale), and average target spectrum at z = 0 (black). (b) Corresponding average cross-correlation coefficient between prediction and target over the same redshift range. (c) Average Dice coefficient between prediction and target (same blue color scale as in a and b).

Thumbnail: Fig. 5 Refer to the following caption and surrounding text. Fig. 5

Five images from the 3D simulations at various redshifts (left) and their equivalent predictions of redshift z = 0 (right) as inferred by the baseline TW. The true z = 0 simulation images are shown above the predicted images (top-right) for comparison.

3.1.2 3D cubes

We now focus on the outcome of the TW trained to recover density fields at redshift z = 0 for the set of 3D simulation cubes, from input density fields at redshifts varying from z = 0 to z = 3 by steps of δ z = 1. The baseline TW is run for 30k gradient updates, and the best weights are recovered for each input z= (0, 1, 2, 3), respectively at 30, 30, 20, and 30 k updates. Indeed, beyond a certain threshold the progression of loss appears to exhibit a linear pattern when plotted on a log-log scale (Fig. 1). We halted the training at 30 k updates due to time constraints.

We observed five simulation cubes taken at random from the test set (Fig. 5). These data are taken at various redshifts (left block) and used as input for the trained TWs to predict their z = 0 equivalent (upper right). The predicted data are shown in the right block. Here, we find that, contrary to the 2D case, the predicted data become noticeably more inaccurate (see Fig. 6); that is to say, we start to see the high-density contrast disappearing, while spurious low-density structures arise with the increase of the input redshift. The data predicted from z = 2 and z = 3 exhibit a very slight similarity to the target data at z = 0.

We note that when provided with this progressively less informative (i.e., the density field at progressively earlier states), the network increasingly tends to default to outputting data that shows few dense structures (or even none) and is less correlated with the correct output, opting for more diffuse structures that can blend in with any target data’s background, thus reducing on average the difference between prediction and any random target.

Given the previous observations, it is not surprising to find that the power spectrum (see Fig. 7, left) of the predicted data becomes lower for high z inputs, since the lack of dense structures leads to overall lower density and a loss of signal at all frequencies. In terms of shape, we find that the predicted power spectra generally reproduce the overall shape of the target. When normalized, the spectra for z = 1 and z = 0 most closely follow the target shape, displaying a flatter plateau at intermediate k followed by a sharper decline at high k, whereas predictions from z = 2 and z = 3 exhibit a smoother, more monotonous descent. The cross-correlation coefficient r(k) remains high for z = 0, with r(k) ≈ 1 up to k ≃ 10−1 h Mpc−1, then gradually decreasing to zero at k ≃ 1 h Mpc−1. For z = 1, we observed a typical linear-prediction behavior, with a good agreement at low k but a much steeper drop beginning around k ≃ 4 × 10−1 h Mpc−1 and reaching r(k) = 0 near k ≃ 2 × 10−1 h Mpc−1. Predictions from higher redshifts (z = 2, 3) decorrelate almost monotonically, from r(k) ∼ 0.9 at low k to r(k) = 0 at k ≃ 10−1 h Mpc−1. These results confirm that the 3D case represents a more challenging task than the 2D configuration: the additional spatial dimension and greater variability in sub-cube mass and its evolution all compound the difficulty of reconstructing detailed structures, despite the potentially easier encoding of global features.

Finally, we examined the Dice coefficient (Fig. 7, right); once more the increased disparity between prediction and target with higher input z is made clear, with the overall Dice value becoming significantly lower for input z = 1 compared to z = 0 and for inputs z = 2 and 3 compared to z = 1. Overall, we can observe that the networks display a worse performance when applied to our 3D data compared to 2D.

Thumbnail: Fig. 6 Refer to the following caption and surrounding text. Fig. 6

Closeup of predictions of a single datum of the 3D simulations for inputs z = 0, 1, 2 and 3, using the baseline TW. We can see that as z grows, predictions are increasingly imprecise, favoring more homogeneous, underdense structures, compared to the ground truth’s dense concentrated structures.

Thumbnail: Fig. 7 Refer to the following caption and surrounding text. Fig. 7

Spectra, cross-correlation coefficient and Dice coefficient for baseline timewarper on 3D data. (a) Power spectra from predictions from input redshifts z = 3 → 0 (blue scale), and target spectrum at z = 0 (black). (b) Corresponding cross-correlation coefficient between prediction and target over the same redshift range. (c) Dice coefficient between prediction and target (same blue color scale as in a and b).

3.2 Introducing the velocity field

Additional information in the form of input density fields’ associated velocity fields (see Fig. 9) is bound to constrain more thoroughly the target density fields, as initial velocities notably provide the necessary information for a linear evolution of the density field. These velocity fields can be obtained from the snapshots. Indeed, in their raw form and for every saved snapshot, our N-body simulations provide us with every particle’s position, but also every particle’s velocity, in the form of a 3D vector for each particle.

Using these data, we trained a TW to recover the density field at z = 0 from the input density + velocity fields (ρ, vx, vy, vz) at redshifts varying from z = 0 to z = 3 by steps of δ z = 1. Observing the predicted data (Fig. 10), we note that models yield nearly identical results regardless of the input z. Unsurprisingly, they seem to have the same limitations as the data inferred by the baseline TW with input z = 0 (see Fig. 5), recovering a thicker dense structure more accurately than finer diffuse structure. Indeed, a TW trained with additional information should not exceed the results of an AE provided with all the required information as input.

As can be expected given the visual similarity between target and predicted data, the predicted power spectra (Fig. 11, upper left) are close, although slightly shifted upwards, to the true simulation power spectrum, especially when compared to those recovered by the baseline T W. Looking at the cross-correlation (lower left), we find that predictions from all input redshifts perform nearly as well as the simple z = 0 → 0 encoding case, with r(k) ≈ 1 up to k ≃ 10−1 h Mpc−1, then gradually decreasing to zero at k ≃ 1 h Mpc−1. A study of the Dice coefficient (Fig. 11, right) completes the picture by showing that the (density+velocity) TW outperforms the baseline TW, to the point that the recovered Dice of the (density+velocity) TW for input z = 3 is better than that of the baseline TW with input z = 1.

Finally, inspecting the evolution of the loss measured throughout training for both cases (baseline TW and velocities TW; i.e., density only vs. density+velocity input as seen in Fig. 8), several noteworthy aspects emerge. First, past a certain point in training all training losses follow a clear linear trend in log-log. While significantly more noisy, this appears to be the case for validation losses as well. Comparing the two cases, we can note that density-only input training leads to overall greater and noisier loss, suggesting that adding velocity fields to the input makes the task less difficult and the training more stable.

We find that including the initial velocity field as input leads to significantly improved predictions of the evolved density field. This is not surprising: the velocity field encodes the direction and magnitude of matter flow, providing direct information about the dynamics driving structure formation. While the density field alone contains some of this information implicitly, supplying the velocity field explicitly allows the network to better anticipate structure evolution. This aligns with the broader physical understanding that matter trajectories (not just the initial positions) shape the development of cosmic structures.

Thumbnail: Fig. 8 Refer to the following caption and surrounding text. Fig. 8

Batch (left) and validation set (right) losses for the velocities TW (red) and baseline TW (blue) at various input redshifts. Both panels share the same y-axis for easier comparison.

Thumbnail: Fig. 9 Refer to the following caption and surrounding text. Fig. 9

Example slice of a 3D simulation, showing the density field (left) and its associated velocity field (right), represented in (vx, vy, vz) to (R, G, B).

4 Discussion

4.1 Results and limitations

The TW model was designed with a dual objective: to compress density fields into a compact representation with minimal information loss, and to predict their future evolution in redshift. This dual nature is central to understanding the performance of the model in different regimes. In 2D, the baseline TW performs surprisingly well at the prediction task, with comparable performance metrics across various input redshifts. This suggests that the limiting factor in this configuration is primarily the encoding component; that is, once the data is well-encoded, predicting its evolution appears relatively easy. Improving the autoencoder structure (particularly its ability to faithfully reconstruct z = 0 from itself) could therefore yield significant gains. Prior works on CAMELS (Villaescusa-Navarro et al. 2021) have demonstrated that deep autoencoders have the capacity to encode this kind of data with high fidelity; albeit potentially at the cost of more complex and less semantically interpretable latent spaces.

Conversely, in 3D, the model exhibits significant degradation in prediction quality as the input redshift increases. This behavior indicates that in this case, the prediction task becomes the dominant bottleneck: as input redshift increases, forward mapping becomes increasingly difficult. We surmise that this is due to the higher-redshift density field inputs effectively providing less information about their z = 0 descendants. While, in principle, a snapshot’s full initial density field at the highest redshift (z = 19) contains all the information required to predict the final state, this is due to our complete knowledge of the phase information (density field + initial velocity is 0) and of neighboring structures (periodic boundaries).

As time progresses, the density field by itself becomes an increasingly incomplete representation of the total phase-space information, although structures start to virialize and the large-scale distribution stabilizes. In our setup, the input subcubes do not possess periodic boundaries, which further limits the model’s access to the surrounding dynamical context. At intermediate and high redshifts (z ≳ 3), neighboring regions can still significantly influence local evolution through matter inflows and outflows, which the model cannot infer from density alone. Thus, except perhaps at the very earliest times (z ≈ 19), higher redshift inputs effectively contain less predictive information about the z = 0 configuration from the model’s limited perspective.

To ease this prediction challenge, we experimented with supplying the model with velocity fields in addition to density. This auxiliary information substantially improves performance; particularly by restoring nearly redshift-invariant prediction quality, as in the 2D case, leading once again to an encoding-based limitation. Velocity information plays a crucial role in resolving ambiguities in the forward evolution of density fields by capturing both local matter dynamics and the influence of surrounding structures beyond the field of view. Thus, our results strongly suggest that the prediction limitations observed with densityonly inputs stemmed primarily from insufficient input information, rather than from an inherent difficulty in approximating structure evolution over longer time intervals.

Thumbnail: Fig. 10 Refer to the following caption and surrounding text. Fig. 10

Five images from the 3D simulations at various redshifts (left), and their equivalent predictions of redshift z = 0 (right) as inferred by the (density + velocity) TW. The true z = 0 simulation images are shown above the predicted images (upper right) for comparison.

Thumbnail: Fig. 11 Refer to the following caption and surrounding text. Fig. 11

Spectra, cross-correlation and Dice coefficients for the baseline in blue, and (density + velocity) in red TW applied to 3D data. (a) Power spectra from predictions from input redshifts z = 3 → 0 (blue/red scales), and target spectrum at z = 0 (black). (b) Corresponding cross-correlation coefficient between prediction and target over the same redshift range. (c) Dice coefficient between prediction and target (same color scales as in a and b).

4.2 Theoretical interpretation

We can reinterpret the TW’s operation as learning a complex, multiparameter function that maps an input density field to its future counterpart. More precisely, it maps from the space of density (and optionally velocity) fields to a latent representation and then to a constrained output space defined by a pretrained GAN’s generative distribution.

This setup implicitly assumes two things; first, the existence and uniqueness of a ground truth: each input corresponds to a single expected output at z = 0. Second, the inclusion of this output in the GAN’s generative space: the latent decoder must be able to represent the ground truth within its learned manifold.

The first assumption gets partially violated in practice. A single density field at a given redshift may evolve differently depending on information not present in the field itself, such as velocity vectors or neighboring structures. Thus, in our baseline 3D case, where only the density field is provided as input, the model must contend with a high variability in possible outcomes due to the limited information available. Faced with this uncertainty, the network adopts a conservative prediction strategy; rather than attempting to reconstruct stark high-density structures that could deviate strongly from the ground truth, it produces outputs with more diffuse structures, and less pronounced contrast. This behavior effectively minimizes the expected loss when the target is difficult to predict, but results in high-density structures being vastly underrepresented.

This explains the degradation in 3D predictions at higher redshifts, where forward evolution becomes increasingly uncertain, particularly for density-only inputs. The second assumption concerns the representational limits of the GAN. If the generative space cannot cover all plausible z = 0 outcomes, the TW is inherently constrained to approximate the closest admissible output. This limitation is most evident when the decoder fails to reconstruct fine-scale structure, even in the “simple autoencoding” case where the input and target data are the same. Addressing this might require training more expressive generative models or exploring architectures that decouple the TW from such constraints entirely; although this risks compromising the realism of the output fields.

5 Conclusion

This work presents a proof-of-concept approach for modeling the nonlinear evolution of large-scale structure using Eulerian fields as inputs. Unlike particle-based Lagrangian methods that excel at tracking individual trajectories, our model focuses on evolving continuous fields directly. This makes it a promising candidate for bridging toward applications that rely on field-level data, such as hydrodynamical simulations or observational reconstructions, where particle data is unavailable or ill-defined. The architecture itself (i.e., a hybrid between an autoencoder and components drawn from GAN training) demonstrates that it is possible to balance local accuracy and statistical realism, although this often involves a tradeoff. In this first iteration, the model favors large-scale structure preservation and statistical agreement over precise pixel-level predictions. But this compromise is not fundamental: future models may learn to better balance these objectives or decouple them entirely.

Several avenues for future improvement can be considered, such as enhancing the model with stronger GAN variants or conditional GANs (Mirza & Osindero 2014; Antipov et al. 2017), which would open options such as adding cosmology dependance or training a single model for any input redshift, increasing latent space dimensionality (with potential costs to interpretability), and leveraging physical symmetries (i.e., isotropy or scale invariance) via architectures such as bispectral neural networks (Sanborn et al. 2022). Further strategies might involve simplifying the prediction task through auxiliary fields (e.g., gravitational potential) or easing training with curriculum learning (Bengio et al. 2009; Ullmo 2022). Beyond architectural improvements, alternate approaches such as VAE (Kingma & Welling 2013) or U-Nets (Siddique et al. 2021) (as in the He et al. 2019 article, but adapted for Eulerian fields) offer flexible modeling paradigms that are worth exploring or even combining with our method. Finally, the model could be extended to tasks where statistical accuracy is more critical than exact reconstruction, such as denoising or inpainting masked regions. Another promising direction is inverse evolution: predicting earlier cosmic states from later ones: a challenging but potentially insightful task due to its intrinsic nonuniqueness (Jasche & Lavaux 2019; Doeser et al. 2024; Legin et al. 2024; Jindal et al. 2023). Rather than aiming to replace the precision of Lagrangian simulators, this line of research seeks to complement them by offering efficient, adaptable tools for cosmological inference, particularly in cases where only field-level data is accessible.

Acknowledgements

The authors thank H. Tanimura for providing the 3D simulations, the IAS ByoPiC4 and CRIStAL Sigma5 teams for fruitful discussions and advice, and D. Jamieson for helpful information. We also thank the anonymous referee for their in-depth feedback, which greatly improved the quality of this manuscript. This project was funded by the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme grant agreement ERC-2015-AdG 695561. M.U. was supported by the Irfu of CEA Saclay, through the PTC program. A.D. was supported by the Comunidad de Madrid and the Complutense University of Madrid (Spain) through the Atraccián de Talento program (Ref. 2019-T1/TIC-13298).

References

  1. Abel, T., Hahn, O., & Kaehler, R., 2012, MNRAS, 427, 61 [Google Scholar]
  2. Agarwal, S., Davé, R., & Bassett, B. A., 2018, MNRAS, 478, 3410 [CrossRef] [Google Scholar]
  3. Alves de Oliveira, R., Li, Y., Villaescusa-Navarro, F., Ho, S., & Spergel, D. N., 2020, arXiv e-prints [arXiv:2012.00240] [Google Scholar]
  4. Antipov, G., Baccouche, M., & Dugelay, J.-L., 2017, arXiv e-prints [arXiv:1702.01983] [Google Scholar]
  5. Aragon-Calvo, M. A., 2019, MNRAS, 484, 5771 [CrossRef] [Google Scholar]
  6. Aragon-Calvo, M. A., 2021, MNRAS, 503, 557 [Google Scholar]
  7. Barnes, J., & Hut, P., 1986, Nature, 324, 446 [NASA ADS] [CrossRef] [Google Scholar]
  8. Bartlett, D. J., Chiarenza, M., Doeser, L., & Leclercq, F., 2025, A&A, 694, A287 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  9. Bengio, Y., Louradour, J., Collobert, R., & Weston, J., 2009, in Proceedings of the 26th annual international conference on machine learning, 41 [Google Scholar]
  10. Biswas, M., & Adlak, R., 2018, in 2018 4th International Conference for Convergence in Technology (I2CT), IEEE, 1 [Google Scholar]
  11. Bonjean, V., 2020, A&A, 634, A81 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  12. Bonjean, V., Aghanim, N., Salomé, P., et al. 2019, A&A, 622, A137 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  13. Boylan-Kolchin, M., Springel, V., White, S. D., Jenkins, A., & Lemson, G., 2009, MNRAS, 398, 1150 [NASA ADS] [CrossRef] [Google Scholar]
  14. Burke, C. J., Aleo, P. D., Chen, Y.-C., et al. 2019, MNRAS, 490, 3952 [NASA ADS] [CrossRef] [Google Scholar]
  15. Chittenden, H. G., & Tojeiro, R., 2023, MNRAS, 518, 5670 [Google Scholar]
  16. Conceição, M., Krone-Martins, A., & Da Silva, A., 2024a, in 2024 IEEE 20th International Conference on e-Science (e-Science), IEEE, 1 [Google Scholar]
  17. Conceição, M., Krone-Martins, A., da Silva, A., & Moliné, Á. 2024b, A&A, 681, A123 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  18. Crain, R. A., Schaye, J., Bower, R. G., et al. 2015, MNRAS, 450, 1937 [NASA ADS] [CrossRef] [Google Scholar]
  19. Dere, S., Fatima, M., Jagtap, R., Inamdar, U., & Shardoor, N. B., 2021, in 2021 7th International Conference on Advanced Computing and Communication Systems (ICACCS), Vol. 1, IEEE, 702 [Google Scholar]
  20. Doeser, L., Jamieson, D., Stopyra, S., et al. 2024, MNRAS, 535, 1258 [NASA ADS] [CrossRef] [Google Scholar]
  21. D’Isanto, A., & Polsterer, K. L., 2018, A&A, 609, A111 [Google Scholar]
  22. Efstathiou, G., Davis, M., White, S., & Frenk, C., 1985, ApJS, 57, 241 [NASA ADS] [CrossRef] [Google Scholar]
  23. Feng, L., 2023, IEEE J. Multiscale Multiphys. Comput. Tech., 8, 97 [Google Scholar]
  24. Freed, M., & Lee, J., 2013, in 2013 International Conference on Computational and Information Sciences, IEEE, 322 [Google Scholar]
  25. Gardner, J. P., Mather, J. C., Clampin, M., et al. 2006, Space Sci. Rev., 123, 485 [Google Scholar]
  26. Giusarma, E., Reyes, M., Villaescusa-Navarro, F., et al. 2023, ApJ, 950, 70 [NASA ADS] [CrossRef] [Google Scholar]
  27. Gondhalekar, Y., Bose, S., Li, B., & Cuesta-Lazaro, C., 2025, MNRAS, 536, 1408 [Google Scholar]
  28. González, R. E., Munoz, R. P., & Hernández, C. A., 2018, Astron. Comp., 25, 103 [Google Scholar]
  29. Hahn, O., Abel, T., & Kaehler, R., 2013, MNRAS, 434, 1171 [Google Scholar]
  30. Hansen, D. L., Mendoza, I., Liu, R., et al. 2022, Mach. Learn. Astrophys., 27 [Google Scholar]
  31. Hausen, R., & Robertson, B., 2022, arXiv e-prints [arXiv:2201.04714] [Google Scholar]
  32. He, S., Li, Y., Feng, Y., et al. 2019, Proc. Natl. Acad. Sci., 116, 13825 [NASA ADS] [CrossRef] [Google Scholar]
  33. Henghes, B., Pettitt, C., Thiyagalingam, J., Hey, T., & Lahav, O., 2021, MNRAS, 505, 4847 [CrossRef] [Google Scholar]
  34. Henghes, B., Thiyagalingam, J., Pettitt, C., Hey, T., & Lahav, O., 2022, MNRAS, 512, 1696 [NASA ADS] [CrossRef] [Google Scholar]
  35. Hiegel, J., Thélie, E., Aubert, D., et al. 2023, A&A, 679, A125 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  36. Hockney, R. W., & Eastwood, J. W., 2021, Computer Simulation Using Particles (Boca Raton: CRC Press) [Google Scholar]
  37. Hsu, A., Ho, M., Lin, J., et al. 2025, Open J. Astrophys., 8, 92 [Google Scholar]
  38. Humbird, K. D., Peterson, J. L., & McClarren, R. G., 2018, arXiv e-print [arXiv:1811.05852] [Google Scholar]
  39. Jamieson, D., Li, Y., de Oliveira, R. A., et al. 2023, ApJ, 952, 145 [NASA ADS] [CrossRef] [Google Scholar]
  40. Jamieson, D., Li, Y., Villaescusa-Navarro, F., Ho, S., & Spergel, D. N., 2025, J. Cosmology Astropart. Phys., 2025, 072 [Google Scholar]
  41. Jasche, J., & Lavaux, G., 2019, A&A, 625, A64 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  42. Jia, P., Zheng, Y., Wang, M., & Yang, Z., 2023, Astron. Comput., 42, 100687 [Google Scholar]
  43. Jindal, V., Liang, A., Singh, A., Ho, S., & Jamieson, D., 2023, arXiv e-prints [arXiv:2303.13056] [Google Scholar]
  44. Jo, Y., & Kim, J.-H. 2019, MNRAS, 489, 3565 [Google Scholar]
  45. Johnson, J., Alahi, A., & Fei-Fei, L., 2016, in Computer Vision-ECCV 2016: 14th European Conference, Amsterdam, The Netherlands (Berlin: Springer), 694 [Google Scholar]
  46. Kingma, D. P., & Welling, M., 2013, arXiv e-prints [arXiv:1312.6114] [Google Scholar]
  47. Kitaura, F.-S., & Heß, S., 2013, MNRAS, 435, L78 [Google Scholar]
  48. Kodi Ramanah, D., Charnock, T., Villaescusa-Navarro, F., & Wandelt, B. D., 2020, MNRAS, 495, 4227 [NASA ADS] [CrossRef] [Google Scholar]
  49. Laureijs, R., Amiaux, J., Arduini, S., et al. 2011, arXiv e-prints [arXiv:1110.3193] [Google Scholar]
  50. Legin, R., Ho, M., Lemos, P., et al. 2024, MNRAS, 527, L173 [Google Scholar]
  51. Li, P., Ilayda Onur, I., Dodelson, S., & Chaudhari, S., 2022, arXiv e-prints [arXiv:2205.07368] [Google Scholar]
  52. Liu, Y., 2023, Int. J. Mod. Phys. C, C34, 2350099 [Google Scholar]
  53. Luo, Z., Chen, J., Chen, Z., et al. 2025, ApJS, 279, 17 [Google Scholar]
  54. Mirza, M., & Osindero, S., 2014, arXiv e-prints [arXiv:1411.1784] [Google Scholar]
  55. Monaco, P., Theuns, T., & Taffoni, G., 2002, MNRAS, 331, 587 [Google Scholar]
  56. Nelson, D., Springel, V., Pillepich, A., et al. 2019, Comput. Astrophys. Cosmol., 6, 1 [NASA ADS] [CrossRef] [Google Scholar]
  57. Planck Collaboration VI., 2020, A&A, 641, A6 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  58. Prost, J., Thouvenin, P.-A., Sorce, J., & Chainais, P., 2025, hal-05042936 [Google Scholar]
  59. Rastegarnia, F., Mirtorabi, M., Moradi, R., Vafaei Sadr, A., & Wang, Y., 2022, MNRAS, 511, 4490 [Google Scholar]
  60. Reyes, E., & Estévez, P. A., 2020, in 2020 International Joint Conference on Neural Networks (IJCNN), IEEE, 1 [Google Scholar]
  61. Rezaei, S., McKean, J., Biehl, M., de Roo, W., & Lafontaine, A., 2022a, MNRAS, 517, 1156 [NASA ADS] [CrossRef] [Google Scholar]
  62. Rezaei, S., McKean, J. P., Biehl, M., & Javadpour, A., 2022b, MNRAS, 510, 5891 [NASA ADS] [CrossRef] [Google Scholar]
  63. Saadeh, D., Koyama, K., & Morice-Atkinson, X., 2024, MNRAS, 537, 448 [Google Scholar]
  64. Sanborn, S., Shewmake, C., Olshausen, B., & Hillar, C., 2022, arXiv e-prints [arXiv:2209.03416] [Google Scholar]
  65. Sanchez-Gonzalez, A., Godwin, J., Pfaff, T., et al. 2020, in International conference on machine learning, PMLR, 8459 [Google Scholar]
  66. Shamir, L., 2009, MNRAS, 399, 1367 [Google Scholar]
  67. Shandarin, S. F., & Zeldovich, Y. B., 1989, Rev Mod. Phys., 61, 185 [Google Scholar]
  68. Siddique, N., Paheding, S., Elkin, C. P., & Devabhaktuni, V., 2021, Ieee Access, 9, 82031 [Google Scholar]
  69. Springel, V., 2005, MNRAS, 364, 1105 [Google Scholar]
  70. Springel, V., Yoshida, N., & White, S. D., 2001, New Astron., 6, 79 [Google Scholar]
  71. Sweere, S. F., Valtchanov, I., Lieu, M., et al. 2022, MNRAS, 517, 4054 [Google Scholar]
  72. Tanimura, H., Aghanim, N., Bonjean, V., & Zaroubi, S., 2022, A&A, 662, A48 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  73. Tassev, S., Zaldarriaga, M., & Eisenstein, D. J., 2013, J. Cosmol. Astropart. Phys., 2013, 036 [CrossRef] [Google Scholar]
  74. Ullmo, M., 2022, PhD thesis, Université Paris-Saclay, France [Google Scholar]
  75. Ullmo, M., Decelle, A., & Aghanim, N., 2021, A&A, 651, A46 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  76. Vafaei Sadr, A., Vos, E. E., Bassett, B. A., et al. 2019, MNRAS, 484, 2793 [CrossRef] [Google Scholar]
  77. Villaescusa-Navarro, F., Anglés-Alcázar, D., Genel, S., et al. 2021, ApJ, 915, 71 [NASA ADS] [CrossRef] [Google Scholar]
  78. Villar, V. A., Cranmer, M., Berger, E., et al. 2021, ApJS, 255, 24 [NASA ADS] [CrossRef] [Google Scholar]
  79. Vogelsberger, M., Genel, S., Springel, V., et al. 2014, MNRAS, 444, 1518 [Google Scholar]
  80. Wiewel, S., Becher, M., & Thuerey, N., 2019, in Computer Graphics Forum (Hoboken: Wiley Online Library), 38, 71 [Google Scholar]
  81. Wilde, J., Serjeant, S., Bromley, J. M., et al. 2022, MNRAS, 512, 3464 [Google Scholar]
  82. Wu, J. F., & Kragh Jespersen, C., 2023, arXiv e-prints [arXiv:2306.12327] [Google Scholar]

1

See original article (Ullmo et al. 2021).

All Tables

Table 1

Architectures of the 2D and 3D TW.

All Figures

Thumbnail: Fig. 1 Refer to the following caption and surrounding text. Fig. 1

Architecture of the timewarper. A trained GAN’s generator is used as a readily built decoder. Only the encoder’s weights are changed during training. The same GAN’s truncated discriminator is used to compute a perceptual loss (see Eq. (1)).

In the text
Thumbnail: Fig. 2 Refer to the following caption and surrounding text. Fig. 2

Example of simulation-issued data: a 2D image from a 2D simulation (left) and a 3D cube (right) from a 3D simulation. These data are built by dividing an N-body simulation snapshot into pixels (or voxels) and counting the particles within each, creating a discrete density map. The map is then smoothed with a Gaussian filter and log-transformed, allowing cosmic structure to stand out starkly. This process creates data that are compatible with convolutional neural networks, which specialize in feature detection.

In the text
Thumbnail: Fig. 3 Refer to the following caption and surrounding text. Fig. 3

Six images from the 2D simulations at various redshifts (left), and their equivalent predictions of redshift z = 0 (right) inferred by the baseline TW. The true z = 0 simulation images are shown above the predicted images (upper right) for comparison.

In the text
Thumbnail: Fig. 4 Refer to the following caption and surrounding text. Fig. 4

Spectra, cross-correlation coefficient and Dice coefficient for baseline timewarper on 2D data. (a) Average power spectra from predictions from input redshifts z = 3 → 0 (blue scale), and average target spectrum at z = 0 (black). (b) Corresponding average cross-correlation coefficient between prediction and target over the same redshift range. (c) Average Dice coefficient between prediction and target (same blue color scale as in a and b).

In the text
Thumbnail: Fig. 5 Refer to the following caption and surrounding text. Fig. 5

Five images from the 3D simulations at various redshifts (left) and their equivalent predictions of redshift z = 0 (right) as inferred by the baseline TW. The true z = 0 simulation images are shown above the predicted images (top-right) for comparison.

In the text
Thumbnail: Fig. 6 Refer to the following caption and surrounding text. Fig. 6

Closeup of predictions of a single datum of the 3D simulations for inputs z = 0, 1, 2 and 3, using the baseline TW. We can see that as z grows, predictions are increasingly imprecise, favoring more homogeneous, underdense structures, compared to the ground truth’s dense concentrated structures.

In the text
Thumbnail: Fig. 7 Refer to the following caption and surrounding text. Fig. 7

Spectra, cross-correlation coefficient and Dice coefficient for baseline timewarper on 3D data. (a) Power spectra from predictions from input redshifts z = 3 → 0 (blue scale), and target spectrum at z = 0 (black). (b) Corresponding cross-correlation coefficient between prediction and target over the same redshift range. (c) Dice coefficient between prediction and target (same blue color scale as in a and b).

In the text
Thumbnail: Fig. 8 Refer to the following caption and surrounding text. Fig. 8

Batch (left) and validation set (right) losses for the velocities TW (red) and baseline TW (blue) at various input redshifts. Both panels share the same y-axis for easier comparison.

In the text
Thumbnail: Fig. 9 Refer to the following caption and surrounding text. Fig. 9

Example slice of a 3D simulation, showing the density field (left) and its associated velocity field (right), represented in (vx, vy, vz) to (R, G, B).

In the text
Thumbnail: Fig. 10 Refer to the following caption and surrounding text. Fig. 10

Five images from the 3D simulations at various redshifts (left), and their equivalent predictions of redshift z = 0 (right) as inferred by the (density + velocity) TW. The true z = 0 simulation images are shown above the predicted images (upper right) for comparison.

In the text
Thumbnail: Fig. 11 Refer to the following caption and surrounding text. Fig. 11

Spectra, cross-correlation and Dice coefficients for the baseline in blue, and (density + velocity) in red TW applied to 3D data. (a) Power spectra from predictions from input redshifts z = 3 → 0 (blue/red scales), and target spectrum at z = 0 (black). (b) Corresponding cross-correlation coefficient between prediction and target over the same redshift range. (c) Dice coefficient between prediction and target (same color scales as in a and b).

In the text

Current usage metrics show cumulative count of Article Views (full-text article views including HTML views, PDF and ePub downloads, according to the available data) and Abstracts Views on Vision4Press platform.

Data correspond to usage on the plateform after 2015. The current usage metrics is available 48-96 hours after online publication and is updated daily on week days.

Initial download of the metrics may take a while.