Connect with us


GPT-4 AI gets chemistry

The latest AI “large language model” could aid chemistry researchers, but limitations still hold it back from much more than the basics.

GPT-4, the latest version of the artificial intelligence system from OpenAI, the developers of ChatGPT, has proved useful in tackling chemistry challenges. 

However significant weaknesses remain. 

“It has a notable understanding of chemistry, suggesting it can predict and propose experimental results in ways akin to human thought processes,” says chemist Kan Hatakeyama-Sato, at the Tokyo Institute of Technology. Hatakeyama-Sato and his colleagues discuss their exploration of the potential of GPT-4 in chemical research in the journal Science and Technology of Advanced Materials: Methods.

GPT-4, which stands for Generative Pre-trained Transformer 4, belongs to a category of artificial intelligence systems known as large language models. These can gather and analyse vast quantities of information in search of solutions to challenges set by users. One advance for GPT-4 is that it can use information in the form of images in addition to text.

Although the specific datasets used for training GPT-4 have not been disclosed by its developers, it has clearly learned a significant amount of detailed chemistry knowledge. To analyse its capabilities, the researchers set the system a series of chemical tasks focused on organic chemistry – the chemistry of carbon-based compounds. This covered basic chemical theory, the handling of molecular data, predicting the properties of chemicals, the outcome of chemical processes and proposing new chemical procedures.

The results of the investigation were varied, revealing both strengths and significant limitations. GPT-4 displayed a good understanding of general textbook-level knowledge in organic chemistry. It was weak, however, when set tasks dealing with specialized content or unique methods for making specific organic compounds. It displayed only partial efficiency in interpreting chemical structures and converting them into a standard notation. One interesting feat was its ability to make accurate predictions for the properties of compounds that it had not specifically been trained on. Overall, it was able to outperform some existing computational algorithms, but fell short against others.

“The results indicate that GPT-4 can tackle a wide range of tasks in chemical research, spanning from textbook-level knowledge to addressing untrained problems and optimizing multiple variables,” says Hatakeyama-Sato. “Inevitably, its performance relies heavily on the quality and quantity of its training data, and there is much room for improvement in its inference capabilities.”

The researchers emphasise that their work was only a preliminary investigation, and that future research should broaden the scope of the trials and dig deeper into the performance of GPT-4 in more diverse research scenarios.

They also hope to develop their own large language models specializing in chemistry and explore their integration with existing techniques.

“In the meantime, researchers should certainly consider applying GPT-4 to chemical challenges, possibly using hybrid methods that include existing specialized techniques,” Hatakeyama-Sato concludes.

Subscribe to our free newsletter
To Top