Open Access Article
Duc-Anh Daoa,
Minh-Quyet Ha
a,
Tien-Sinh Vu
a,
Shuntaro Takazawab,
Nozomu Ishiguro
b,
Yukio Takahashib,
Suzuki Masatoc,
Takashi Kakuboc,
Naoya Aminoc,
Hirosuke Matsuid,
Mizuki Tadae and
Hieu-Chi Dam
*ab
aJapan Advanced Institute of Science and Technology, 1-1 Asahidai, Nomi, Ishikawa 923-1292, Japan. E-mail: dam@jaist.ac.jp
bInternational Center for Synchrotron Radiation Innovation Smart, Tohoku University, 2-1-1 Katahira, Aoba, Sendai 980-8577, Japan
cMaterials Performance Laboratory, Research and Advanced Development Division, The Yokohama Rubber Co., Ltd, Oiwake, Hiratsuka, Kanagawa 254-8601, Japan
dSchool of Science, Kwansei Gakuin University, 1 Gakuen Kamigahara, Sanda, Hyogo 669-1330, Japan
eDepartment of Chemistry, Graduate School of Science, Nagoya University, Furo, Chikusa, Nagoya, Aichi 464-8602, Japan
First published on 20th October 2025
Understanding nanoscale material evolution—including phase transitions, structural deformations, and chemical reactions—under dynamic conditions remains a fundamental challenge in materials science. While advanced imaging techniques enable visualization of transformation processes, they typically capture only discrete temporal observations at specific time intervals. Consequently, intermediate stages and alternative pathways between captured images often remain unresolved, introducing ambiguity in analyzing material dynamics and transformation mechanisms. To address these limitations, we present a two-stage framework using deep generative models to probabilistically reconstruct intermediate transformations. Our framework is based on the hypothesis that generative models trained to reproduce experimental images inherently capture the dynamical processes that generated those observations. By integrating these trained generative models into Monte Carlo simulations, we generate plausible transformation pathways that interpolate unobserved intermediate stages. This approach enables the extraction of meaningful insights and the statistical analysis of material dynamics. This study also evaluates the framework's applicability across three phenomena: tantalum test chart translation, gold nanoparticle diffusion in polyvinyl alcohol solution, and copper sulfidation in heterogeneous rubber/brass composites. The generated transformations closely replicate experimental observations while revealing previously unrecognized dynamic behaviors for future experimental validation. These findings suggest that learned generative models encode physically meaningful continuity, enabling statistical interpolation of unobserved intermediate states and classification of transformation modes under sparse observational constraints.
Recent advanced imaging techniques, including scanning electron microscopy (SEM),8 transmission electron microscopy (TEM),9 and coherent X-ray diffraction imaging (CXDI),10–12 enable researchers to track structural and compositional changes during material transformations. These methods generate sequential snapshots of material states, often captured in real-time or under in situ conditions. Conventional analyses model dynamic behavior from these discrete observations, as illustrated in Fig. 1a. However, these techniques face a fundamental limitation: their discrete temporal sampling inherent in these techniques captures only isolated moments within continuous transformation processes. This constraint is analogous to reconstructing a continuous dynamical system from sparse temporal observations—while each captured state reveals crucial information about the system's evolution, the intermediate dynamics driving transitions between observed states remain inaccessible.
To overcome these observational limitations, methods capable of simulating the transitions between captured sparse snapshots are highly required. Ideally, while denser experimental sampling would yield more complete datasets,13,14 practical constraints in imaging equipment and measurement time often render this approach infeasible. Therefore, an alternative approach is to augment experimental data with computational methods that simulate plausible transitions between observed states.15 This strategy forms the conceptual foundation of our approach, addressing data sparsity through statistically grounded analyses that extend beyond directly measurable transitions.
With recent advances, deep generative models offer a promising avenue to address the limitations of observable transitions.16–19 Models such as generative adversarial networks (GANs),20 variational auto-encoders (VAEs),21 and diffusion models22 are designed to learn mappings from high-dimensional image data to compact latent spaces that preserve essential structural features. Crucially, we hypothesize that by learning to generate experimental images, these models implicitly capture the underlying dynamical processes that produce the observed material states. A key advantage of these latent representations is their local smoothness—small perturbations in latent coordinates typically yield gradual and coherent variations in the generated images, despite the absence of explicit physical constraints during training. This property is particularly valuable for modeling material transformations—such as defect motion or phase-boundary migration—where latent geometry can reflect physical continuity in structural evolution.23,24 Thus, exploring the latent space offers a potential mechanism to simulate hypothetical transitions that complement sparse experimental observations and statistically enrich the space of observed transformations.16,18
The local smoothness of latent spaces suggests a powerful approach to address the challenge of sparse observations. Given that small perturbations in latent coordinates typically yield gradual, coherent changes in generated images,25,26 we can systematically explore neighborhoods around observed states to generate material variations—an ensemble of plausible intermediate states that could occur during transformation. This concept arises from recognizing that nanoscale systems rarely follow deterministic paths. Instead, thermal fluctuations, kinetic barriers, and competing energetic pathways create multiple possible routes between observed states. By sampling these variations through controlled exploration of the latent space, we transform the limitation of discrete observations into an opportunity for richer analysis. Notably, while we cannot directly observe all intermediate states, we can utilize the learned generative model to generate statistically plausible alternatives. This approach parallels established methods in statistical mechanics, where ensemble averages provide more robust insights compared with single trajectories. However, implementing this strategy requires careful consideration of the extent to which the exploration of latent space can be conducted while maintaining physical validity.
Based on these insights, we propose a framework that integrates deep generative modeling with statistical sampling and distributional analysis to characterize material transformations from experimental data with limited transition coverage. Our approach uses a generative model to map compact latent representations to hypothetical material states, enabling efficient navigation on the latent space. To systematically sample local transitions, we implement Monte Carlo (MC) sampling27 in latent space to explore material variations around a given material state. This probabilistic strategy enables diverse sampling of latent codes under constraints that ensure the generated transitions remain physically plausible while capturing variation in possible outcomes (Fig. 1b). To further analyze dynamic behaviors, the generated states are grouped into variations that approximate the collective range of transformations possible from the same initial condition, rather than representing single predictions. This approach, based on local sampling and statistical aggregation, enables statistical interpolation of unobserved intermediate states and classification of transformation behaviors, even in the absence of explicit time-series data. Thus, the proposed framework extends generative models' representational capacity, enabling the interpretation of transformation behavior through physically grounded, data-driven reasoning.
For this construction, we implement GANs,28 selected for their effectiveness in generating high-quality images from limited training data. Our GAN approach constructs a latent space where the original data instances follow a multivariate normal distribution
enabling systematic sampling of the material state space. The adversarial training process involves a generator network that produces realistic material images, whereas a discriminator network learns to distinguish between generated and real experimental images, progressively improving generation quality. Compared with alternative approaches, GANs offer a practical trade-off between fidelity and efficiency. In contrast, the variational sampling principle in VAEs imposes an additional constraint by enforcing reconstructions through a latent distribution, which tends to blur fine-scale features and restricts analysis of delicate structural or compositional details, particularly at the nano- or microscale. Diffusion models29 require extensive computational resources and large datasets, making them less practical in this context.
The GAN consists of two competing neural networks: a generator G and a discriminator D, trained on a dataset
of experimental material images, as shown in Fig. 2a. Both networks use convolutional layers to capture the spatial features present in the material data. The generator G: Z → X maps a latent vector
from a d-dimensional latent space Z to a synthetic image
in image space X, simulating a possible material configuration. The discriminator D:
assigns higher scores to real images and lower scores to generated ones.
To train the model, we employ the Wasserstein loss function,30 which provides a stochastic objective for comparing real and generated distributions:
![]() | (1) |
and
are distributions of observed images x in dataset
over image space X and the prior distribution of latent vector z over the latent space Z, respectively. Accordingly,
denotes a material image sampled from the dataset
and
are latent vectors sampled from the multivariate normal distribution
over the latent space Z.
In material dynamics modeling, the generator is trained to produce images that simulate physically plausible dynamically evolving material states, such as those induced by diffusion, deformation, or sulfidation. These synthetic images represent hypothetical progressions along dynamic trajectories, capturing variations that may be absent from experimental observations. The discriminator acts as a critic, emphasizing realistic image features and enabling the model to internalize key structural and compositional characteristics of material evolution.
To improve training stability and image fidelity, we implement a gradient penalty31 along with a progressive growing strategy,32 which begins with low-resolution images and incrementally increases resolution. This multi-scale approach facilitates the learning of both coarse and fine features essential for capturing hierarchical material structures. The configuration of model architecture and training schedule across material systems follows the logic of progressive growing: training begins at low resolution to stabilize global structures, and layers with gradually reduced filter counts are added as resolution increases to capture finer details. Accordingly, higher-resolution datasets require deeper networks and finer resolution schedules, while batch sizes are scaled down to maintain stability, with stricter reductions for volumetric data due to cubic scaling (see Tables S1 and S2). All available experimental data were used for training without partitioning into separate sets, as the objective was not to predict unseen outcomes but to model the full distribution of material states. Model quality was assessed by validating that generated samples reflect the statistical and physical properties of the training data.
Our MC approach, illustrated in Fig. 2b, models short-range transformations within the GAN-generated latent space. We adopt a first-order Markov assumption where transformations depend primarily on the current state. This assumption enables us to focus on local changes rather than reconstructing full dynamic trajectories, which require densely sampled, high-resolution data rarely obtainable without specialized instrumentation. By constraining transitions to experimentally plausible magnitudes, we ensure physical relevance while enabling probabilistic analysis of material dynamics without the need to model long-range sequences.
We sample M latent vectors zi from the normal distribution
over the latent space Z and input them into the generator G to produce material state images
i = G(zi). For each sampled state as a reference, we generate N images of multiple transitioned states by perturbing its latent vector with a stochastic deviation εj (where j = 1, …, N), as follows:
i,j = G(zi,j) = G(zi + εj),
| (2) |
We determine γ by correlating the maximum observed micro-structural displacement over time intervals with the average latent space distances between adjacent time-step reconstructions. The uniform distribution ensures equal sampling probability within a controlled radius around each reference state, providing comprehensive coverage of local transformations.
Latent perturbations can be interpreted as surrogates for stochastic fluctuations in material states under thermal, chemical, or mechanical influences. Although the latent space is not explicitly linked to thermodynamic or kinetic parameters, its geometry reflects structural similarity, enabling the interpretation of small perturbations as physically feasible local transformations. Therefore, γ serves as a proxy for the expected magnitude of change under short-time dynamics or environmental variations. To select γ, we compare generated transformations to the structural differences observed between experimentally yielded states sampled at the highest available temporal resolution, ensuring that perturbations represent short-range, physically plausible transitions.
The sampled states
i,j, representing transitions associated with the reference state
i, are grouped into a set denoted as a material variation Vi. This set represents possible transformations associated with
i over short timescales, expressed as:
Vi = { i, i,1, i,2, …, i,N}.
| (3) |
Each material variation Vi is then characterized by the statistical distribution of reference state
i and its N transitioned states
i,j, reflecting patterns in the transformations occurring within Vi. For each case study, we design features that represent characteristic variations across the sampled state images, guided by domain knowledge of the associated physical processes. These features serve as representations of each material variation Vi.
Subsequently, we apply unsupervised clustering techniques to the space of material variation descriptors to identify representative transformation behaviors across the dataset. This approach enables systematic grouping of variations based on both discrete and continuous transformation patterns, offering insight into dominant dynamic modes. Notably, we do not impose directionality on these transitions. Given that the transformations are derived from differences between generated image samples in the GAN latent space, they are not constrained to a fixed direction of change. Instead, each pair of states can be interpreted as either a forward or backward transition, reflecting the symmetric nature of local variations in this representation. This general formulation is suitable for cases where transition order is uncertain or experimentally unresolved while still preserving local physical plausibility.
1. Image fidelity: evaluating whether the generated images of material states accurately reflect physically plausible configurations that align with real experimental observations.
2. Latent space smoothness: verifying that small perturbations in latent vectors produce gradual and coherent changes in generated images. This criterion assesses the model's capacity to reflect smooth physical transformations, such as defect movements or boundary migrations.
3. Insight consistency with prior studies: comparing the physical insights obtained from our generative analysis with those from established conventional approaches, emphasizing consistency rather than superiority.
To comprehensively evaluate these criteria, we selected three distinct yet progressively challenging case studies (Table 1), each designed to systematically increase the complexity of both the imaging data and the underlying physical mechanisms. The first case study served as a controlled, proof-of-concept scenario using two-dimensional (2D) phase images obtained through coherent X-ray diffraction imaging (CXDI) of a Ta test chart.33 These images captured simple, precisely controlled translational and rotational motions without internal structural changes, forming a baseline for validating the fundamental capabilities of the generative model.
Building on this foundational scenario, the second case study also employed CXDI to reconstruct phase images capturing the diffusion of gold nanoparticles (NPs) in a polyvinyl alcohol (PVA) solution.12 In this study, the dataset featured stochastic and complex NP motions, testing the framework's ability to model and interpret realistic diffusive processes.
The third case study further increased complexity by examining 3D volumetric images obtained using X-ray absorption fine structure computed tomography (XAFS-CT).1 These 3D images comprise three separate channels corresponding to Cu valence states (Cu0, Cu1+, Cu2+), offering insights into the chemical degradation (aging) processes within brass–rubber composites over time.
The model architecture is adapted for each case, with the number of convolutional layers and filter sizes tailored to the specific characteristics of each dataset. In the 2D cases (Ta test chart and NP diffusion), we employed 2D convolutional GANs. Conversely, for the 3D XAFS-CT case, we used a 3D convolutional GAN to better capture volumetric features. All models were trained on an NVIDIA A100 graphics processing unit (GPU) with 80 GB of memory. A complete description of the training configuration, including stage-wise growth schedules tailored for specific datasets, is provided in Section S1.
Throughout this study, we define a material variation as the set of synthetic material states generated by perturbing the latent vector of a single reference configuration. Thus, each variation represents an ensemble of physically admissible, short-range transitions accessible from that configuration and constitutes the fundamental unit of our analysis. For clarity, we adopt context-specific labels—NP variations for diffusion-driven displacements of NPs and clump variations for sulfidation-induced morphological changes in brass particles—while maintaining a unified underlying concept. In all cases, a material variation functions as a structured surrogate for local material dynamics, capturing the spectrum of plausible short-range evolution.
The dataset is acquired using CXDI to capture the horizontal translation of a Ta test chart moving at 340 nm s−1 under triangular aperture illumination.33 The raw dataset consists of 1755 phase-reconstructed frames, each representing a 7 ms time window of the Ta test chart's continuous motion. To enhance the statistical robustness of our generative model while maintaining physical realism, we augment the dataset through controlled in-plane rotations. Specifically, each original frame is supplemented with five additional images generated by applying random rotations within the range of [−30°, 30°], expanding the total dataset to 10
530 images. For visualization, images were center-cropped to highlight the most coherent region unaffected by peripheral artifacts, as shown in Fig. 3a. However, for model training, full image frames—including boundary artifacts—were retained to maximize the perceivable structural information. The GAN was trained on this dataset to capture the statistical distribution of physically plausible material states and their potential transformations (see Section S1.1 for details).
To evaluate whether the latent space encodes smooth, plausible transitions, we conduct an MC sampling step using the trained GAN. A latent vector z was sampled from the standard multivariate Gaussian prior and perturbed by additive noise drawn from a uniform distribution
(see more details in Section S3.1). While this range may seem large relative to the unit-variance prior, it corresponds to moderate deviations in high-dimensional latent spaces owing to their geometric properties (see Section S2). This process yields a local variation set Vz, consisting of images that represent material states that can transform between each other. Although these generated images are not temporally ordered, they collectively reflect plausible short-range transitions in the configuration space.
Fig. 3b and c show two representative image sequences, compiled post hoc from variations generated in Vz. These sequences are presented to illustrate that latent-space sampling captures the inherent mechanisms underlying the observed rigid-body dynamics: one representative sequence highlights horizontal translation and the other in-plane rotation. To construct these illustrative sequences, we selected subsets of generated images and sorted them according to quantitatively evaluated transformation magnitudes, such as translation offsets or rotation angles, computed via a template-matching procedure applied to locally perturbed latent samples (see Section S3.2 for details). Fig. 3b illustrates smooth translational motion along the x-axis, whereas Fig. 3c captures rotational steps of approximately 1–2 degrees. Full sequences are provided in Videos S1 and S2. Compared with the original experiment, the generated translation appears slightly slower, yet the spatial coherence across frames supports the interpretation that the latent space encodes physically meaningful, continuous variations. Intersection-over-union (IOU) distributions between consecutive generated frames and those in the experimental sequence are shown in Fig. 4a. The higher IOU values in the synthetic data suggest that the sampled transitions correspond to smaller structural changes than those observed experimentally, which is consistent with short-range interpolations.
To quantitatively verify the plausibility of the sampled transitions, we analyze their consistency with ground-truth rigid-body dynamics. Given that the transformation in this case study involves only rigid-body motion, which implies that the Ta test chart moves without internal deformation, any geometric inconsistencies in the generated images must originate from limitations in the model's latent representation. To detect such deviations, we track two structural features—slit-like probe lines labeled d1 and d2—across all images in the sampled variation set Vz, originating from a reference image Iz, as shown in Fig. 4b. We apply template-matching-based tracking34 to robustly localize and quantify the dimensions of these features throughout the variation set (more details of the process is shown in Section S3.2). As shown in Fig. 4b, the generated images consistently preserve the contrast between transmitted and absorbed regions, enabling clear delineation of the slits. Notably, the measured slit widths in the generated images are d1 = 309 ± 11 nm and d2 = 340 ± 12 nm, closely aligning with the reference widths in Iz and comparable to those in the observed dataset: d1 = 311 ± 13 nm and d2 = 324 ± 11 nm. The resulting deviations (26–39 nm) fall within approximately two pixels (1 pixel ≈ 20.72 nm), indicating that the generative model preserves key structural features and does not introduce significant deformations during localized transformations.
Overall, these results support the core hypothesis of our framework—that localized traversal in the GAN latent space, even without explicit temporal supervision, can expose transformation patterns consistent with physical dynamics.
The experimental dataset comprises 2000 sequentially captured, phase-reconstructed images from 2D CXDI, recording real-time NP movement in the PVA solution.12 Each image has a resolution of 142 × 142 pixels, with a spatial resolution of approximately 40.58 nm per pixel. These images represent snapshots of NP arrangements evolving under thermal agitation, providing an ideal dataset to evaluate the generative model's ability to represent and interpolate diffusive dynamics.
Beyond assessing the fidelity of generated images, we also evaluate the smoothness of the GAN's latent space. Specifically, we assess whether small perturbations in latent vectors lead to gradual and coherent changes in generated images. This criterion reflects the model's capacity to represent plausible short-range NP movements. To examine this hypothesis, we generate 100 perturbed variants around each of 1000 latent vectors sampled from a standard normal distribution. Perturbations are then applied by adding noise drawn from a uniform distribution
with the γ varied from 0.25 to 2.0 in increments of 0.25. This approach enables the analysis of how latent space distance influences structural variations in the generated images. To quantify how these perturbations affect image structure, we measure two metrics between reference and perturbed images: (1) image entropy difference, and (2) particle area difference. Image entropy35 measures the statistical complexity of grayscale distributions and reflects structural richness in the image. Notably, higher entropy indicates more complex or varied patterns.
Interpreting distances in high-dimensional latent spaces is challenging because absolute distances often lose meaning due to the curse of dimensionality.36,37 To address this, we measure the average Euclidean distance between each reference latent vector and its perturbed samples and normalize this average by the number of latent dimensions (see details provided in Section S2). This normalization yields an intuitive measure of locality within the latent space. Consequently, perturbation margins of γ = 0.25 and γ = 2 correspond to average per-dimension changes of approximately 0.025 and 0.20, respectively.
As shown in Fig. 5c, both metrics increase smoothly with larger γ, indicating that the latent space encodes a gradient of structural variation. At lower margins (e.g., γ = 0.25), generated variations show minimal differences from their reference images. Conversely, at higher margins, substantial differences in image entropy and particle area emerge, reflecting broader configurational changes in NP arrangements. Notably, a sampling margin of γ = 1 (equivalent to a per-dimension sampling margin of approximately 0.1) yields differences in entropy and particle area closely matching those observed between consecutive experimental frames. At this level, the entropy difference corresponds to the upper bound observed between consecutive experimental frames (0.26), and particle area differences remain below 200 pixels. This observation aligns with NP depletion rates of approximately 10 particles per second,12 indicating that latent vectors sampled within this margin produce configurations reflecting the natural stochastic motion of NPs over a timescale of about one second. Therefore, in subsequent simulations of dynamic behavior, we adopt γ = 1 to generate physically plausible transitions.
To quantitatively compare these NP variations, we construct a distance matrix using a metric scaled to account for the range and distribution of the descriptors (see Section S4.3 for details). We then apply hierarchical agglomerative clustering (HAC) using the average linkage method40,41 to identify structurally coherent groups based on geometric separation in descriptor space. This approach enables the classification of distinct diffusion modes by grouping trajectories with similar path characteristics. As shown in the dendrogram in Fig. 6c, three primary clusters emerge, suggesting distinct classes of NP motion within the PVA solution.
Fig. 6d shows the distribution of anisotropy and tortuosity values across these clusters. The first cluster exhibits directionally biased diffusion with lower tortuosity, indicating persistent movement potentially influenced by channel-like structures or local field gradients within the polymer.42 The second cluster exhibits random, isotropic diffusion characterized by high tortuosity and low directional persistence—features typical of classic Brownian motion and consistent with unstructured diffusion in homogeneous fluids.43 The third cluster consists of highly confined diffusion characterized by limited displacement and no dominant direction, suggesting transient immobilization or trapping—a phenomenon known to occur in polymer matrices and crowded soft-matter environments.44,45
To examine these sulfidation dynamics, we analyzed a 3D imaging dataset acquired using XAFS-CT1 at the SPring-8 synchrotron facility.46 This advanced imaging technique enables precise 3D mapping of elemental compositions and oxidation states, allowing spatially resolution of metallic Cu, monovalent Cu2S, and CuS distributions within rubber/brass composites. The dataset was generated through controlled aging experiments conducted in a previous study:1 approximately 2000 brass particles (Cu/Zn ratio = 75/25) were embedded within a rubber matrix, heated at 443 K for 10 min to form adhesive layers, and subsequently aged under controlled conditions. The aging process and corresponding XAFS-CT imaging were repeated over multiple intervals, producing imaging datasets at aging times of 0, 3, 14, and 28 days, denoted as Aging-0, Aging-3, Aging-14, and Aging-28, respectively.
From this time-resolved dataset, we systematically tracked the chemical evolution of individual brass particles throughout the entire aging period. Initially, 6590 particles were extracted from the experimental data (details of the data collection process are shown in Section S1.3). However, to ensure consistent volumetric representation within our generative modeling framework, we retained only particles whose dimensions fit entirely within a 32 × 32 × 32 voxel grid (voxel size = 0.65 μm per voxel). Particles exceeding this spatial constraint were excluded from subsequent modeling. This preprocessing step resulted in a final dataset comprising 4956 particles, each tracked across the aging intervals (0, 3, 14, and 28 days).
Beyond assessing the fidelity of generated images, we evaluated the smoothness of the GAN's latent space, specifically, whether small perturbations in latent vectors lead to gradual and coherent changes in generated images. We applied the same local perturbation protocol described in Section 3.3.1, generating 100 perturbed variants around each of 1000 latent vectors previously sampled from the standard normal distribution. Perturbations were applied by adding noise drawn from a uniform distribution
with the γ varied from 0.25 to 2.0 in increments of 0.25. To quantify the deviation of perturbed samples from their respective references, we measured two metrics: (1) the complement of the 3D IOU image (IOU3D), which captures shape-based dissimilarity (see Section S5.1), and (2) total compositional density difference, which quantifies voxel-wise changes in elemental distribution. As shown in Fig. 7d, both metrics gradually increase with increasing γ, indicating that the latent space encodes a continuous and interpretable spectrum of compositional transformations.
Notably, a sampling margin of γ = 0.5 (equivalent to a per-dimension sampling margin of approximately 0.05) yields an IOU3D ≈ 0.7 and total compositional mass deviations under 20%, closely matching observed differences between adjacent sulfidation stages.1 This suggests that latent vectors sampled within this margin produce brass clumps consistent with natural aging processes observed within the experimental time window. Therefore, in subsequent simulations of aging behavior, we adopt γ = 0.5 to generate physically plausible transitions.
Each sampled set defines a clump variation, representing the ensemble of admissible transformations around a given reference clump. From each set, we derived a single data instance in the form of a descriptor vector summarizing sulfidation-induced redistribution behaviors. Specifically, we computed the interquartile range (IQR) of voxel intensities for Cu, Cu2S, and CuS across all valid samples in the set. Subsequently, we calculated the variance of these IQR values as a descriptor of sulfidation-induced redistribution extent. Thus, each clump variation was represented by a 3D vector [Var(IQRCu), Var(IQRCu2S), Var(IQRCuS)], capturing compositional responses within its local transformation space (see Section S5.2). We applied Multidimensional scaling (MDS) to these descriptors, generating a 2D embedding space that visualizes the similarity among clump variations based on their transformation behaviors (Fig. 8a). Details of the embedding space construction are shown in Section S5.3. Notably, the embedded clump variations self-organized into three distinct groups—denoted as A, B, and C—each exhibiting characteristic transformation patterns.
To interpret the underlying dynamic behaviors, clump variations were mapped to experimentally temporal observations (Fig. 8b). Clump variations in Group A predominantly corresponded to the initial aging stage (Aging-0). Conversely, clump variations in Group B were primarily observed after 3 to 14 days, while those in Group C appeared mostly in later stages (14 to 28 days). Previous experimental studies indicate that early aging (0–14 days) primarily involves sulfidation and conversion of metallic Cu to monovalent Cu2S. By contrast, the later aging stage (14–28 days) is characterized by increased formation of bivalent CuS. This temporal alignment suggests that clump variations in Groups A and B are primarily associated with the sulfidation and consumption of Cu to form the monovalent Cu2S phase, while clump variations in Group C are more closely associated with the formation of the bivalent CuS phase.
Fig. 8c provides further support for the inferred sulfidation phases. Clump variations in Group A exhibited high metallic Cu content, indicating minimal sulfidation, whereas Groups B and C showed pronounced Cu2S and CuS enrichment, respectively. Further simulation of transitions within these groups revealed that clump variations in Groups A and B exhibited distinct degrees of monovalent sulfidation. Notably, the transformation amplitude in Group A was significantly lower compared with Group B. Conversely, Group C exhibited significant bivalent sulfidation, confirming distinct transformation dynamics at later stages.
The representative simulation of these clump variations, shown in Fig. 8d, revealed distinct patterns: Cu2S preferentially accumulated near clump surfaces during the early and intermediate stages, whereas CuS formation prominently occurred in interior regions during late stages. These spatial distributions closely match experimental observations, where early sulfidation (Cu2S formation) dominates interfacial regions during initial vulcanization, whereas deeper CuS formation is associated with progressive mechanical degradation observed in aging or cracked regions.1,47,48
Despite its effectiveness, the framework presents two main limitations. First, like any data-driven model, the proposed framework depends on the coverage and diversity of the training data; rare or under-sampled transformation patterns may not be reliably inferred. Second, the approach models transitions as isolated, memoryless perturbations, following a first-order Markovian assumption. Although this simplification facilitates analysis, it restricts the ability to capture trajectory-dependent behaviors such as hysteresis, irreversibility, or long-range causality. Addressing these limitations motivates future directions toward trajectory-level modeling. Incorporating temporal priors through latent ordinary differential equations (ODEs), time-conditioned generative models, or reinforcement learning-based latent navigation could enable the generation of physically consistent transformation sequences. Such advances would transition the framework from sampling local alternatives to simulating complete dynamic pathways governed by kinetic or thermodynamic constraints.
Nonetheless, even without explicit temporal structure, the current formulation provides insights into short-range transformation mechanisms. It also supports interpretable, probabilistic reasoning under sparse observational conditions, reveals emergent structural variation, and facilitates comparative analysis across heterogeneous datasets. Accordingly, the framework functions both as an independent tool for latent dynamics interpretation and as a foundational step toward generative simulation of material evolution. As machine learning continues to advance materials discovery, our study offers a principled and extensible approach to unveiling hidden dynamics from high-dimensional experimental observations.
Code availability: the Python implementation and custom code used to train the GAN model are publicly available on GitHub,51 and an archived version is stored in Zenodo52 to ensure long-term accessibility and reproducibility.
Supplementary information: comprehensive documentation of the test datasets, image preparation, model architecture, training regime, and analysis methods. It also presents additional results and images that support the main text and enable reproducibility. See DOI: https://doi.org/10.1039/d5dd00277j.
| This journal is © The Royal Society of Chemistry 2025 |