2 code implementations • 8 Apr 2024 • Zihan Pengmei, Zimu Li
Graph Transformers have emerged as a powerful alternative to Message-Passing Graph Neural Networks (MP-GNNs) to address limitations such as over-squashing of information exchange.
1 code implementation • 2 Oct 2023 • Zihan Pengmei, Zimu Li, Chih-chan Tien, Risi Kondor, Aaron R. Dinner
We demonstrate SubFormer on benchmarks for predicting molecular properties from chemical structures and show that it is competitive with state-of-the-art graph transformers at a fraction of the computational cost, with training times on the order of minutes on a consumer-grade graphics card.
1 code implementation • 22 Aug 2023 • Zihan Pengmei, Junyu Liu, Yinan Shu
We show that the xxMD dataset involves diverse geometries which represent chemical reactions.
no code implementations • 14 Nov 2022 • Zimu Li, Zihan Pengmei, Han Zheng, Erik Thiede, Junyu Liu, Risi Kondor
Equivariant graph neural networks are a standard approach to such problems, with one of the most successful methods employing tensor products between various tensors that transform under the spatial group.