To make computer chips, technologists around the world rely on atomic layer deposition (ALD), which can create films as fine as one atom thick. Businesses commonly use ALD to make semiconductor devices, but it also has applications in solar cells, lithium batteries and other energy-related fields. Manufacturers increasingly rely on ALD to make new types of films, but figuring out how to tweak the process for each new material takes time. A recently published study suggests that using artificial intelligence (AI) can be more efficient.
In the study, researchers at the U.S. Department of Energy’s (DOE) Argonne National Laboratory describe multiple AI-based approaches for optimising the ALD processes autonomously. Their work details the relative strengths and weaknesses of each approach, as well as insights that can be used to develop new processes more efficiently and economically.
Two different chemical vapours, known as precursors, adhere to a surface, adding a thin layer of film in the process. This all happens inside a chemical reactor and is sequential: one precursor is added and interacts with the surface, then any excess of it is removed. Afterwards, the second precursor is introduced then later removed, and the process repeats itself. In microelectronics, the ALD thin film might be used to electrically insulate nearby components in nanoscale transistors.
ALD excels at growing precise, nanoscale films on complex, 3D surfaces such as the deep and narrow trenches patterned into silicon wafers to manufacture today’s computer chips. This has motivated scientists worldwide to develop new thin-film ALD materials for future generations of semiconductor devices.
However, developing and optimising these new ALD processes is challenging and labour-intensive. Researchers have to consider many different factors that can alter the process, including:
- The complex chemistries between the molecular precursors
- Reactor design, temperature and pressure
- The timing for each dose of their precursors
In an effort to find ways of overcoming these challenges, U.S. scientists evaluated three optimisation strategies — random, expert system and Bayesian optimization — the latter two utilising different AI approaches.
Researchers evaluated their three strategies by comparing how they optimised the dosage and purge times of the two precursors used in ALD. Dosage time refers to the time period when a precursor is added to the reactor, while purge time refers to the time needed to remove excess precursor and gaseous chemical products.
Researchers aim to find the conditions that would achieve high and stable film growth in the shortest time. Scientists also judged the strategies on how quickly they converged on the ideal set of timings using simulations that represented the ALD process inside a reactor.
Linking their optimisation approaches to their simulated system let them measure film growth in real-time after each cycle, based on the processing conditions their optimisation algorithms generated. All of the algorithms provide a much faster way of converging to optimum combinations because of a real-time loop that connects with the reactor. This setup also made the process automatic for the two AI approaches by forming a closed-loop system.
In a closed-loop system, the simulation performs an experiment, gets the results, and feeds them to the AI tool. The AI tool then learns from it or interprets it in some way, and then suggests the next experiment. All of the processes happen without human input.
Despite some weaknesses, the AI approaches effectively determined the optimal dose and purge timings for different simulated ALD processes. This makes the study among the first to show that thin-film optimisation in real-time is possible using AI.
The adoption of AI opens up the possibility of using these types of approaches to rapidly optimise real ALD processes, a step that could potentially save manufacturers precious time and money when developing new applications in the future.