Michael
Alverson
*ac,
Sterling G.
Baird
a,
Ryan
Murdock
a,
(Enoch) Sin-Hang
Ho
b,
Jeremy
Johnson
b and
Taylor D.
Sparks
a
aDepartment of Materials Science & Engineering, University of Utah, Salt Lake City, 84112, Utah, USA. E-mail: michaeldavidalverson@gmail.com; sterling.baird@utah.edu; rynmurdock@gmail.com; sparks@eng.utah.edu
bDepartment of Chemistry & Biochemistry, Brigham Young University, Provo, 84602, Utah, USA. E-mail: enochho201315@gmail.com; jjohnson@chem.byu.edu
cThomas Lord Department of Computer Science, University of Southern California, Los Angeles, 90089, California, USA
First published on 6th December 2023
The idea of materials discovery has excited and perplexed research scientists for centuries. Several different methods have been employed to find new types of materials, ranging from the arbitrary replacement of atoms in a crystal structure to advanced machine learning methods for predicting entirely new crystal structures. In this work, we pursue three primary objectives. (I) Introduce CrysTens, a crystal encoding that can be used in a wide variety of deep learning generative models. (II) Investigate and analyze the relative performance of Generative Adversarial Networks (GANs) and Diffusion Models to find an innovative and effective way of generating theoretical crystal structures that are synthesizable and stable. (III) Show that the models that have a better “understanding” of the structure of CrysTens produce more symmetrical and realistic crystals and exhibit a better apprehension of the dataset as a whole. We accomplish these objectives using over fifty thousand Crystallographic Information Files (CIFs) from Pearson's Crystal Database.
Crystal generative models typically use variational auto-encoders (VAEs),27,31,35,37,41 generative adversarial networks (GANs),30,32,37,40,42 genetic algorithms,28 or substitution.29 One example of a crystal generative model based on VAEs is the Fourier-Transformed Crystal Properties (FTCP) representation.41 The FTCP representation is a crystal representation that not only incorporated both chemical and structural crystal properties but was invertible as well (property → chemistry + structure). When combined with a jointly trained VAE model, a probabilistic property-structured latent space was obtained that allowed for the generation of novel crystals with user-defined properties. Another important work in this area of study is the Crystal Diffusion Variational Autoencoder (CDVAE).39 By leveraging a diffusion process that pushes atomic coordinates to lower energy states and iterates atom types to satisfy bonding preferences, CDVAE significantly outperforms past attempts to perform material generation. Furthermore, CDVAE was capable of optimizing crystals for a given property and reconstructing a material from its latent space representation. Many examples exist for GANs as well. Constrained crystals deep convolutional generative adversarial network (CCDC-GAN)32 uses a post-processing filtering criteria to remove compounds with large formation energies based on a cutoff threshold. In another example, crystal structure is represented in a lean fashion by a set of atomic coordinates and unit cell parameters, with significantly reduced memory requirements compared to a voxel-based generative model, iMatGen.35 Physics Guided Crystal Generative Model (PGCGM)40 expands on CubicGAN42 while using base site atoms with physics-guided loss functions. For example, structures with atoms that are too crowded or too far apart are penalized via two terms in the loss function. We also note that normalizing flows, another promising direction, have been used in some crystal structure contexts38,43 including an application to molecular crystals.43
In this work, we compare the use of Generative Adversarial Networks and Diffusion Models in unstructured crystal generation. We illustrate the shortcomings of GANs in this space and show that Diffusion Models may offer a very promising alternative to previous methods. Furthermore, we introduce a standardized and image-like crystal embedding representation (CrysTens) that can be used in a wide array of image generation models with relative ease and minimal changes.
New machine learning crystal representations that capture chemistry (composition) and structure (periodicity) are important because they then allow us to utilize machine learning algorithms to identify and exploit patterns in data. Even the most experienced materials scientist domain expert would be unable to fully comprehend and leverage all the patterns in high-dimensional materials data for hundreds of thousands of crystal structures, and then use the discovered patterns to generate novel materials. Instead, scientists have relied on very low-order approximations and simplifications for generating new materials. Or, alternatively, they have relied on screening down lists of already discovered materials to identify those candidates that most closely match the desired properties using empirical relationships or computational materials science techniques such as molecular dynamics, Density Functional Theory (DFT), etc. However, given the immense size of the chemical space that is believed to exist (1060 materials) and our current microscopic subset of known materials (105–106 materials),3 it is unlikely that screening efforts alone are sufficient to find the new materials necessary to answer society's most pressing technological needs. Therefore it is abundantly clear that an intuitive and efficient method is needed to scan over the vast regions of untapped chemical space and select the groupings of materials it deems as stable and synthesizable.
Fortunately, machine learning and more specifically deep learning methods, have emerged as powerful complements to the human capacity for materials design.48,49 Within the field of material discovery, generative machine learning models are currently being investigated by a wide variety of research teams.39,44,50,51 Previously, the two most common types of generative machine learning models were the VAE and GANs. VAEs attempt to encode a sample of data into a lower dimensional latent space. The encoded samples are then decoded from latent space into potential new samples. However, VAEs differ from traditional autoencoders because they are simultaneously attempting to structure the latent space according to a predefined probability distribution. This makes VAEs an exciting candidate for efficient inverse design because materials that exhibit ideal properties may be located at the intersection of each property within latent space.52,53
The other common generative machine learning model for material discovery is GANs. GANs differ from VAEs in several key ways. First, GANs are composed of two separate neural network architectures: the generator and the discriminator. The generator's task is to create realistic samples of whatever data distribution it is trying to model. In the case of this work, the generator is attempting to create realistic crystal structures. The discriminator's job is to differentiate between samples that are real (taken from the original dataset) or generated. The discriminator and generator train against each other, continually improving until training is finished and the generator can be separated and used to create realistic samples. By using an adversarial approach, the generator can construct its own probability distribution of the data instead of requiring a pre-defined probability distribution like those used within VAEs. Although the training of GANs is game-theoretic in nature, they are not guaranteed to converge to a Nash equilibrium54 which can lead to performance issues. A more advanced variation of the traditional GAN is included in this work known as the Wasserstein GAN or WGAN. The differences between WGANs and GANs are discussed later.55,56
Recently, however, with the success of OpenAI's DallE-257 and Google's Imagen,58 Diffusion Models have quickly risen to achieve state-of-the-art performance for many types of generative modeling. Diffusion Models are inspired by non-equilibrium thermodynamics and operate by gradually destroying input training data by adding Gaussian noise (forward process) only to learn the transformations necessary to reconstruct each sample (backward process). By continually repeating this process, and incrementally adding more noise in each iteration, fully trained Diffusion Models are able to completely reconstruct a data sample from nothing more than noise. Diffusion Models take longer to train than GANs but do not suffer from many of the same deficiencies that GANs do such as mode collapse and extreme instability.
Fig. 1 Layer 1 of the crystal representation containing symmetrical rows and columns for structural information and a pairwise distance matrix with relative distances between all atoms in the basis. |
Fig. 2 Layers 2–4 of the crystal representation containing the same symmetrical rows and columns for structural information, but also having a directional graph for each dimension: x, y, and z. |
Finding a concise, efficient, and structurally informative representation for each CIF was an important step in the process of crystal structure generative modeling. Additionally, we wanted our representation to have image-like characteristics so that we could take advantage of high-performing image-generation models without a large amount of refactoring. After many iterations, including just simply listing the crystal structure parameters in a list, we eventually decided upon a representation that captured both the intricate parameters of each crystal as well as their interatomic components. The representation is a tensor of shape 64 × 64 × 4 and can be visualized in Fig. 1 and 2. The top twelve rows and leftmost twelve columns are symmetrical and list out all of the CIF-extracted information of a given crystal. The top (or leftmost) list is the atomic number of each of the atoms present in the crystal. They are listed from left to right (or top to bottom) until either fifty-two spots have been filled or the crystal has run out of atoms to place. If there is leftover space in the representation, zeros will be filled in for the remainder. Structures with more than 52 atoms in the basis are excluded. The same process is repeated for the three fractional coordinates (x, y, and z), the three lattice parameters (a, b, and c), the three lattice angles (α, β, and γ), and the space group number. Finally, a padding layer is inserted to separate the CIF-extracted information from the interatomic portion of the representation. For the first layer of the representation, the bottom-right 52 × 52 matrix is used to encode a pairwise distance matrix that relates each of the atoms together by their Euclidean distance. Within the latter three layers of the representation, a distance graph for each of the dimensions is represented to show the uni-dimensional relative distance between each atom.
The motivation behind the structure of this representation is to highlight the major components of crystal structures. There is a structural component that encodes symmetry, basis, and lattice information and an interatomic component that encodes the relative distances between atoms into the representation itself. With only the interatomic distances of a given crystal, it would be impossible to reconstruct the atomic numbers, space group, and various lattice parameters. Likewise, with only the structural components, it may be challenging for a convolution-based generative modeling algorithm to encode the relationships between various components which may lead to generated structures with unreasonable interatomic distances. Furthermore, there is a high degree of redundancy within our crystal structure representation by including both a pairwise distance matrix and unidimensional distance matrices as well as the repeating of lattice parameters, angles, and the space group number. The reason for this redundancy is that we would like to force the model to relate all of these different aspects in a way that can give rise to a greater ability to recognize and utilize crystal patterns within the PCD dataset during generation and a higher degree of noise mitigation during post-processing when all of these generated values are averaged. The primary strength of the CrysTens representation is that the structural aspects of a given crystal are fully encoded by including both the pairwise distance matrix and the distance graphs. This is desirable when using convolution-based image-generation models as they perform best when there is structurally informative data in the representation. Also, this gives the models that are trained on our representation several ways in which they can learn the patterns present in a crystal dataset. The repetition of lattice parameters, lattice angles, and the space group number present in our crystal representation may be viewed as a potential drawback as the model is forced to learn the same information several times, however, we chose to move forward with this representation as we felt it was the most natural representation.
We selected 64 as the length and width of each of our layers because of the commonly used deep learning heuristic of selecting powers of two. It could easily have been another value, however, the focus of this work is to show the structure of the CrysTens representation rather than the numerical intricacies or optimization of the representation. Furthermore, since 64 was selected and there are eleven different parameters plus a padding layer, only crystal structures with 52 atoms or less were selected for training and analysis. We redundantly included both the pairwise distance matrix and the three dimensional graphs to fully capture all of the different structural aspects of a given crystal structure, however, the representation can be changed to fit given constraints. We anticipate that future work could add layers to encode aspects related to chemistry such as the Oliynyk, magpie, mat2vec vector constituents in order to ensure realistic atom assignment.
Given a generated CrysTens, there will be a number of atoms that it will have predicted to be present in a crystal. We will refer to this value as A, and 0 < A ≤ 52 due to the nature of the current CrysTens representation. If the generative model was able to understand the symmetrical relationship within the CrysTens representation, then it will have A non-empty columns from left to right, and A non-empty rows from top to bottom for each of the four layers. Similarly, the single-valued structure parameters such as the lattice parameters (a, b, c), the lattice angles (α, β, γ), and the space group number will be repeated 2A times for each layer, a total of 8A occurrences. These values can be averaged to find the value that will ultimately be used when generating the output CIF. The variance among these repeated values is a good indicator of whether a given model “understands” the structure of the representation. If almost all 8A space group numbers are between 219 and 221, then there is a good chance that the model intended on a space group number of 220 and knew where to place these values. Generated CrysTens representations with small variances between values that are meant to be the same tended to produce more symmetrical and better-looking crystals, shown in Results. However, if there is a space group number spread between 110 and 220 there is a very good chance that the model is not able to pick up on the relationship between the different space group rows and columns, and the quality of the finalized crystal is likely to reflect that. Values that do not represent a single value throughout the entire CrysTens such as atomic number or fractional coordinates are instead only repeated a total of 8 times (two times for each layer) because each one corresponds to a single atom. The same variance check can be performed on each of these values as well.
However, the single best indicator of model performance comes from the use of the directional graphs (layers 2–4). Each spot in the directional graphs represents the relative difference between two atoms in all three dimensions. Therefore, with the knowledge of one atom and its relationship to another atom with the direction graph, the second atom's location can be deduced. Therefore, if we have 10 atoms predicted in a CrysTens representation, each with their respective fractional coordinates and a directional graph relating each of them, we can have 10 “guesses” as to where a given atom is supposed to be in the crystal (one guess from its own x, y, z coordinates and nine guesses from the relative distances of other atoms). This process is highlighted in two dimensions in Fig. 3. If there is a high degree of consistency between all of the coordinate predictions, the crystal that is produced tends to be more symmetrical and realistic. The average of the coordinate predictions for each atom is ultimately chosen as the point where a given atom is placed during the CIF reconstruction. Furthermore, with these reconstructed coordinates, a pairwise distance matrix can be created and compared to the generated pairwise distance matrix in layer 1 as an additional metric for model performance.
We are now left with an averaged value for the lattice parameters, lattice angles, and the space group number. We also have an average value for each of the coordinates of a given atom. In order to produce clean crystals that do not reflect the noise of the generative model, we added a few more post-processing steps.
As far as the atomic numbers are concerned, we found that even when there was low variance within the predicted atomic numbers we could still observe lists of atomic numbers such as (12.2, 12.6, 12.4, 12.5, 7.7, 7.4, 8.4, 8.5). When rounded to the nearest atomic number, this list would reflect (12, 13, 12, 13, 8, 7, 8, 9) → (Mg, Al, Mg, Al, O, N, O, F) instead of the list that would correspond to rule of parsimony: (12, 12, 12, 12, 8, 8, 8, 8) → (Mg, Mg, Mg, Mg, O, O, O, O). In order to rectify this inconsistency, we elected to use K-means clustering for the atomic numbers. For similar reasons, we chose to use K-means clustering for the coordinate values of x, y, and z across the different atoms. Now we could create a list of elements in the CIF as well as another list corresponding to their atomic positions. Using the averaged values from above, a Pymatgen lattice object can be constructed which is then used in conjunction with the element and coordinate lists to create a Pymatgen structure object. The structure object can be used to create a CIF.
• For lattice parameters, lattice angles, and space group number, average all 8A instances of each and use averages in the generated CIF.
• For each atomic number assignment, it will be repeated 8 times (2 times per layer). Average these 8 instances for all A atomic number assignments.
• Perform K-means clustering on averaged atomic number assignments, setting each to its centroid value and rounding to the nearest whole number for the final list of atomic number assignments.
• For atomic position assignments in each dimension, it will be repeated 8 times (2 times per layer). Average these 8 instances for all A unidimensional atomic position assignments.
• Perform the process described in Fig. 3.
• Perform K-means clustering on averaged atomic position assignments, setting each to its centroid value and rounding to the nearest whole number for the final list of atomic position assignments.
• Construct CIF with lattice parameters, lattice angles, space group number, atomic number list, and atomic position list.
The major shortcoming associated with the sigmoid activation function directly corresponds to the difficulty of the tasks assigned to the two neural networks. It is far easier to classify if a given sample is real or fake than it is to generate an entirely new sample. During the initial training epochs, before the generator has had any time to calculate the weights necessary for accurate sample generation, it will produce samples that are very obviously fake. When the discriminator has to decide between something that is clearly real and something that is obviously fake, it will start performing very well and will begin predicting with higher confidence (a value closer to either 1 or 0).
When plotted on a sigmoid activation function, low confidence predictions correspond to a large gradient (closer to 0.5) while high confidence predictions correspond to a progressively diminishing gradient (closer to 1 or 0). A prediction plotted within the large gradient zone of a sigmoid activation function corresponds with more-useful information being given back to the generator about how to improve its weights. Inversely, a prediction plotted within the diminishing gradient zone will provide the generator with less-useful information. The more confident the discriminator becomes with its predictions, the more the gradient and the usability of the information coming from the discriminator will be decreased (vanishing gradient problem). This leads to a repeating cycle where the discriminator continually improves and the performance of the generator ultimately stagnates.
Since the slope of the line is constant with respect to the relative performance between the critic and the generator (there is no low-gradient zone), the critic will never cease to grant the generator with useful data in terms of how to adjust its weights in order to produce more realistic samples.
Table 1 is used to show the ability of each model to capture the symmetrical characteristics of the CrysTens' structure. The average variance among each repeated parameter, angle, space group number, and fractional coordinates in the CrysTens representation for one thousand generated CIFs was calculated. Table 2 shows the average agreement between absolute and relative coordinates in a given CrysTens as to where to place each atom, as well as the average difference between the reconstructed pairwise distance matrix and the generated pairwise distance matrix for one thousand generated CIFs.
Model | σ Parameter 2 | σ Angle 2 | σ Spacegroup 2 | σ Fractionalcoordinates 2 |
---|---|---|---|---|
a One-thousand CrysTens representations were generated and the variance was averaged over all of them. | ||||
GAN | 582.23 | 1338.11 | 6322.1 | 5.57 × 10−2 |
WGAN | 31.18 | 695.03 | 3024.27 | 5.9 × 10−2 |
Diffusion | 3.5 × 10−1 | 8.46 × 10−1 | 4.56 | 4.36 × 10−4 |
Model | ΔCoordinate/atoma | Σ(ΔPairwise)/atomb |
---|---|---|
a The difference between the “absolute” x, y, and z values predicted and the “relative” position predicted by the direction graph. b The sum of the difference between the reconstructed pairwise distance matrix (by the final coordinates) and the generated pairwise distance matrix in layer 1 of CrysTens. c One-thousand CrysTens representations were generated and the values was averaged over all of them. | ||
GAN | 9.62 × 10−2 | 3.84 |
WGAN | 1.37 × 10−2 | 2.49 |
Diffusion | 5.51 × 10−4 | 4.73 × 10−1 |
Furthermore, we wanted to investigate how well each model modeled the PCD dataset as a whole. To do this, the distribution proportions of the parameters, angles, space group number, and atomic numbers were found for the 53856 CIFs in our dataset. The same distribution proportions were calculated for the set of one thousand generated CIFs generated by each model and the Kolmogorov–Smirnov (K–S) statistic was calculated for comparison between the distributions of real and generated parameters. Similar distributions will have a low K–S statistic (Fig. 4–6).
Fig. 4 The distribution proportion for the parameters a, b, and c for the real CIFs vs. the generated CIFs. |
Fig. 5 The distribution proportion for the angles α, β, and γ for the real CIFs vs. the generated CIFs. |
To investigate whether the correlation between CrysTens symmetry and CIF quality was reflected on an external metric, we used the Crystal Graph Convolutional Neural Network (CGCNN) to calculate the distribution of predicted formation energy, final energy, band gap, bulk moduli, shear moduli, Fermi energy, and Poisson ratio for the real CIFs as well as the generated CIFs for each model (Fig. 7).62 The K–S statistic is calculated for comparison between the distributions of real and generated predicted properties. This metric explores a model's quantitative ability to create CIFs that adhere to the real CIF property distributions without ever being exposed to CGCNN predictions during training. We recognize that CGCNN serves as an imperfect surrogate model because it was trained on a different dataset. Because of this, the CIF distributions will not match reality exactly, however, this metric provides proxy model performance indicators nonetheless. We also evaluated the CIFs based on their predicted energy above convex hull. In order to do this, we used M3GNet to predict the formation energy of a given crystal and compared it to the convex hull of its crystal system, which was created by PyMatGen (Fig. 8). M3GNet is a materials graph neural network that incorporates three-body interactions.63 The energy above convex hull calculation code is found in https://github.com/michaeldalverson/CrysTens. Finally, several generated CIFs from each model are shown to showcase the visual differences in the generated crystal quality (Fig. 9–11). The produced CrysTens' used for evaluation, used 3-means clustering for the atomic numbers and 6-means clustering for the potential distinct x, y, and z coordinates (see Post-processing summary). Manually selecting the K value before generating one thousand CIFs from each model does damage the overall performance of each model. This constraint was applied to each model however, and thus a comparison between each model with this limitation is valid.
Fig. 7 The distribution of CGCNN predicted values for the real CIFs and the CIFs generated by each of the models. |
Fig. 8 The distribution of M3GNet-predicted energy above convex hull for the CIFs generated by each of the models. |
When the distribution of lattice parameters, lattice angles, space group numbers, and atomic numbers of each model was checked against the values in the real distribution, we found that once again, our Diffusion Model performed best. In all areas, Diffusion Models performed exceedingly well, even capturing the peaks in the distribution of space group and atomic numbers (see Fig. 6), WGANs were not able to do this and Vanilla GAN distributions were even worse with indications of severe mode collapse.
Using pre-trained CGCNN models, several material properties were predicted for each CIF. The distribution of each predicted property for each model was compared against the distribution of that property across real CIFs. The violin plots qualitatively show the comparison between the generated and real CIF distributions. By looking at the shape of the violin plots with respect to that of the real distribution and the K–S statistics, it can be seen that both Diffusion Models and WGANs were capable of approximating the real distribution of each predicted property to an adequate degree, while GANs struggled, once again showing the mode collapse that occurred. In every case, Diffusion Models produced a CIF distribution that had a lower K–S statistic than both GANs and WGANs, implying a more similar distribution to the real CIFs in PCD. When investigating the M3GNet predicted energy above convex hull, we found that Diffusion Models and WGANs both performed far better than GANs. Diffusion Models, on average, produced CIFs that had a lower energy above convex hull than WGANs.
Finally, a set of CIFs was shown for each model. The CIFs generated by the Vanilla GAN represented exactly what was exhibited in the parameter and CGCNN prediction distributions. All of the CIFs found had very similar parameters and space group numbers. There was an extremely low degree of symmetry and general realness. Furthermore, the bounding boxes outlined by the lattice parameters were enormous in every CIF that we observed. The lack of “understanding” of CrysTens that the Vanilla GAN showed reflected in the CIFs it was capable of producing.
The CIFs produced by our WGAN were a positive shift in the correct direction. There was far more variability in the lattice parameters, lattice angles, space groups, and elements present when CIFs were generated. There are echoes of symmetrical components visible within the CIFs, however, symmetry would not be a word used to describe these CIFs. There is clearly room for improvement for the CIFs generated from our WGAN.
The CIFs generated by our Diffusion Models not only have a diverse set of lattice parameters, lattice angles, space group numbers, and elements present, but they are symmetrical and realistic looking. Although the CIFs produced are not perfect, there is a clear distinction between the CIFs produced by our Diffusion Model and our WGAN.
Fig. 12 The distribution of formation energy, as predicted by M3GNet for the CIFs generated by the Diffusion Model. |
Among the top 35 CIFs with the lowest M3GNet-predicted formation energy, we selected 6 of them to be rigorously analyzed using Vienna Ab Initio Simulation Package (VASP). To prepare the 6 CIFs, slight manual tuning was applied after post-processing to maximize the likelihood that the structure would be realistic and stable. The manual tuning consisted of rounding fractional coordinates to more probable locations such as (0.011, 0.502, 0.009) → (0.0, 0.5, 0.0), and correcting potentially erroneous atomic number assignments as a result of the rounding within post-processing (36:Kr → 37:Rb). In order to evaluate the stability of the 6 selected CIFs, we performed a series of 4 successive relaxation calculations for each crystal structure. Running multiple relaxation calculations using a conjugate gradient algorithm helps ensure the ions have converged to their instantaneous ground state in the primitive unit cell. We used the Perdew–Burke–Ernzerhof exchange–correlation functional modified for solids, a break condition of 10−5 eV for the ionic relaxation loop, and a smearing width of 0.1 eV. Atomic pseudopotentials developed by G. Kresse and Peter Blöchl were used.64,65 All the calculations were performed with a cutoff energy of 400 eV for the plane-wave basis set, the k-point grid as 3 × 3 × 3,66 and each relaxation calculation typically included about 20 self-consistent field (SCF) iterations to reach the energy cut-off value. We calculate the free energy for the pre-relaxed CIFs and for the relaxed structures, and we calculate the external pressure after each of the relaxation calculations to evaluate their stability. The calculated free energy for the pre-relaxed structures (E0), after the 4th relaxation (E4) and the external pressure on the unit cell in kb after the 1st (P1) and 4th (P4) relaxation are shown below (Fig. 13).
Fig. 13 VASP calculated free energy and external pressure for 6 CIFs produced by the Diffusion Model. |
As expected, all the free energies are lower and the external pressure values are closer to zero for the 4th relaxation, which indicates the VASP optimized the CIFs into even more stable structures. However, it is observed that the 6 CIFs exhibited low free energy and external pressure on the unit cell even from the pre-relaxation and 1st optimization. This is especially apparent for the NaF and SnYF3 structures, which E0 and E4 are very close and the external pressure values after the 1st optimization are very close to 0 kb. These are strong indications of the stability of the produced CIFs.
With all of the crystal information organized within a CIF text file, VESTA is able to create visualizations as seen in Fig. 14 that assist materials scientists in determining the structural components of a given crystal. Pymatgen, a materials informatics Python library, was used to extract relevant information from each CIF during CrysTens construction using a Python programming interface and the Pymatgen Structure object (see Fig. 15).69
Fig. 14 (left) A VESTA visualization of NaCl (right) a VESTA visualization of W6S8(PC18H15)6(C6H6).67 |
Each Unet has 256 base channels. The first Unet created a “low-resolution” version of the CrysTens that was size 32 × 32 × 4. The second Unet, took the output of the first Unet and created the actual CrysTens of size 64 × 64 × 4. Each Unet was trained separately for 250000 optimization steps with batch size 4. The original code is found at https://github.com/lucidrains/imagen-pytorch.
This journal is © The Royal Society of Chemistry 2024 |