Using HPC and experiment, researchers continue to refine graphene production

Graphene could also be among the many most enjoyable scientific discoveries of the final century. Whereas it’s strikingly acquainted to us — graphene is taken into account an allotrope of carbon, which means that it basically the identical substance as graphite however in a distinct atomic construction — graphene additionally opened up a brand new world of potentialities for designing and constructing new applied sciences.

The fabric is two-dimensional, which means that every “sheet” of graphene is only one atom thick, however its bonds make it as robust as a number of the world’s hardest metallic alloys whereas remaining light-weight and versatile. This invaluable, distinctive mixture of properties have piqued the curiosity of scientists from a variety of fields, resulting in analysis in utilizing graphene for next-generation electronics, new coatings on industrial devices and instruments, and new biomedical applied sciences.

It’s maybe graphene’s immense potential that has consequently induced one among its greatest challenges — graphene is troublesome to supply in massive volumes, and demand for the fabric is frequently rising. Current analysis signifies that utilizing a liquid copper catalyst could also be a quick, environment friendly means for producing graphene, however researchers solely have a restricted understanding of molecular interactions taking place throughout these transient, chaotic moments that result in graphene formation, which means they can’t but use the strategy to reliably produce flawless graphene sheets.

So as to deal with these challenges and assist develop strategies for faster graphene manufacturing, a group of researchers on the Technical College of Munich (TUM) has been utilizing the JUWELS and SuperMUC-NG high-performance computing (HPC) techniques on the J├╝lich Supercomputing Centre (JSC) and Leibniz Supercomputing Centre (LRZ) to run high-resolution simulations of graphene formation on liquid copper.

A window into experiment

Graphene’s enchantment primarily stems from the fabric’s completely uniform crystal construction, which means that producing graphene with impurities is wasted effort. For laboratory settings or circumstances the place solely a small quantity of graphene is required, researchers can place a chunk of scotch tape onto a graphite crystal and “peel” away atomic layers of the graphite utilizing a method that resembles how one would use tape or one other adhesive to assist take away pet hair from clothes. Whereas this reliably produces flawless graphene layers, the method is gradual and impractical for creating graphene for large-scale functions.


Business requires strategies that might reliably produce high-quality graphene cheaper and quicker. One of many extra promising strategies being investigated entails utilizing a liquid metallic catalyst to facilitate the self-assembly of carbon atoms from molecular precursors right into a single graphene sheet rising on prime of the liquid metallic. Whereas the liquid affords the power to scale up graphene manufacturing effectively, it additionally introduces a number of issues, such because the excessive temperatures required to soften the everyday metals used, reminiscent of copper. When designing new supplies, researchers use experiments to see how atoms work together underneath a wide range of situations. Whereas technological advances have opened up new methods for gaining perception into atomic-scale habits even underneath excessive situations reminiscent of very excessive temperatures, experimental strategies don’t at all times enable researchers to look at the ultra-fast reactions that facilitate the proper modifications to a cloth’s atomic construction (or what points of the response might have launched impurities). That is the place laptop simulations might be of assist, nevertheless, simulating the habits of a dynamic system reminiscent of a liquid is just not with out its personal set of issues.

“The issue describing something like that is you must apply molecular dynamics (MD) simulations to get the precise sampling,” Andersen stated. “Then, after all, there’s the system measurement — you must have a big sufficient system to precisely simulate the habits of the liquid.” In contrast to experiments, molecular dynamics simulations provide researchers the power to take a look at occasions taking place on the atomic scale from a wide range of totally different angles or pause the simulation to give attention to totally different points.

Whereas MD simulations provide researchers insights into the motion of particular person atoms and chemical reactions that might not be noticed throughout experiments, they do have their very own challenges. Chief amongst them is the compromise between accuracy and value — when counting on correct ab initio strategies to drive the MD simulations, this can be very computationally costly to get simulations which are massive sufficient and final lengthy sufficient to precisely mannequin these reactions in a significant means.

Andersen and her colleagues used about 2,500 cores on JUWELS in durations stretching over multiple month for the latest simulations. Regardless of the large computational effort, the group might nonetheless solely simulate round 1,500 atoms over picoseconds of time. Whereas these might sound like modest numbers, these simulations had been among the many largest finished of ab initio MD simulations of graphene on liquid copper. The group makes use of these extremely correct simulations to assist develop cheaper strategies to drive the MD simulations in order that it turns into doable to simulate bigger techniques and longer timescales with out compromising the accuracy.

Strengthening hyperlinks within the chain

The group printed its record-breaking simulation work within the Journal of Chemical Physics, then used these simulations to match with experimental information obtained of their most up-to-date paper, which appeared in ACS Nano.

Andersen indicated that current-generation supercomputers, reminiscent of JUWELS and SuperMUC-NG, enabled the group to run its simulation. Subsequent technology machines, nevertheless, would open up much more potentialities, as researchers might extra quickly simulate bigger numbers or techniques over longer durations of time.

Andersen obtained her PhD in 2014, and indicated that graphene analysis has exploded throughout the identical interval. “It’s fascinating that the fabric is such a latest analysis focus — it’s nearly encapsulated in my very own scientific profession that folks have regarded intently at it,” she stated. Regardless of the necessity for extra analysis into utilizing liquid catalysts to supply graphene, Andersen indicated that the two-pronged method of utilizing each HPC and experiment can be important to additional graphene’s growth and, in flip, use in industrial and industrial functions. “On this analysis, there’s a nice interaction between concept and experiment, and I’ve been on either side of this analysis,” she stated.