New Machine Learning Tool Converts 2D Material Images into 3D Structures

Imperial College London

A new algorithm developed at Imperial College London can convert 2D images of composite materials into 3D structures.

The machine-learning algorithm could help materials scientists and manufacturers to study and improve the design and production of composite materials like battery electrodes and aircraft parts in 3D.

Using data from 2D cross-sections of composite materials, which are made by combining different materials with distinct physical and chemical properties, the algorithm can expand the dimensions of cross-sections to convert them into 3D computerized models. This allows scientists to study the different materials, or “phases,” of a composite and how they fit together.

The tool learns what 2D cross-sections of composites look like and scales them up so their phases can be studied in a 3D space. It could in the future be used to optimize the designs of these types of materials by allowing scientists and manufacturers to study the layered architecture of the composites.

The researchers found their technique to be cheaper and faster than creating 3D computer representations from physical 3D objects. It was also able to more clearly identify different phases within the materials, which is more difficult to do using current techniques.

Combining materials as composites allows you to take advantage of the best properties of each component, but studying them in detail can be challenging as the arrangement of the materials strongly affects the performance. The algorithm allows researchers to take their 2D image data and generate 3D structures with all the same properties, which allows them to perform more realistic simulations.

Studying, designing, and manufacturing composite materials in three dimensions is currently challenging. 2D images are cheap to obtain and give researchers high resolution, wide fields of view, and are very good at telling the different materials apart. On the other hand, 3D imaging techniques are often expensive and comparatively blurry. Their low resolution also makes it difficult to identify different phases within a composite.

For example, researchers are currently unable to identify materials within battery electrodes, which consist of ceramic material, carbon polymetric binders, and pores for the liquid phase, using 3D imaging techniques.

In this study, the researchers used a new machine learning technique called ‘deep convolutional generative adversarial networks’ (DC-GANs) which was invented in 2014.

This approach, where two neural networks are made to compete against each other, is at the heart of the tool for converting 2D to 3D. One neural network is shown the 2D images and learns to recognize them, while the other tries to make “fake” 3D versions. If the first network looks at all the 2D slices in the “fake” 3D version and thinks they’re “real,” then the versions can be used for simulating any material property of interest.

The same approach also allows researchers to run simulations using different materials and compositions much faster than was previously possible, which will accelerate the search for better composites.

The performance of many devices that contain composite materials, such as batteries, is closely tied to the 3D arrangement of their components at the microscale. However, 3D imaging these materials in enough detail can be painstaking. Researchers are hoping that their new machine learning tool will empower the materials design community by getting rid of the dependence on expensive 3D imaging machines in many scenarios.

Generating three-dimensional structures from a two-dimensional slice with generative adversarial network-based dimensionality expansion.