|
1 |
| ---- |
2 |
| -title: Matbench Discovery |
3 |
| -subtitle: Can machine learning identify stable crystals given unrelaxed structures? |
4 |
| -tags: |
5 |
| - - Python |
6 |
| - - machine learning |
7 |
| - - materials science |
8 |
| - - materials discovery |
9 |
| - - benchmark |
10 |
| - - ensembles |
11 |
| - - uncertainty estimation |
12 |
| -authors: |
13 |
| - - name: Janosh Riebesell |
14 |
| - orcid: 0000-0001-5233-3462 |
15 |
| - corresponding: true |
16 |
| - affiliation: 1, 2 |
17 |
| - - name: Rhys Goodall |
18 |
| - orcid: 0000-0002-6589-1700 |
19 |
| - affiliation: 1 |
20 |
| - - name: Anubhav Jain |
21 |
| - orcid: 0000-0001-5893-9967 |
22 |
| - affiliation: 2 |
23 |
| - - name: Kristin Persson |
24 |
| - orcid: 0000-0003-2495-5509 |
25 |
| - affiliation: 2 |
26 |
| - - name: Alpha Lee |
27 |
| - orcid: 0000-0002-9616-3108 |
28 |
| - affiliation: 1 |
29 |
| -affiliations: |
30 |
| - - Cavendish Laboratory, University of Cambridge, UK |
31 |
| - - Lawrence Berkeley National Laboratory, Berkeley, USA |
32 |
| -date: Jan 31, 2023 |
33 |
| -# To create a PDF from this markdown file, run: |
34 |
| -# ```sh |
35 |
| -# cd paper |
36 |
| -# pandoc path/to/this/file --output paper.pdf --citeproc |
37 |
| -# ``` |
38 |
| -# Requires `brew install pandoc` on macOS. Not identical but similar output as the |
39 |
| -# artifact generated by the JOSS GitHub action |
40 |
| -# https://github.com/marketplace/actions/open-journals-pdf-generator |
41 |
| ---- |
42 |
| - |
43 | 1 | <script>
|
44 | 2 | import MetricsTable from '$figs/metrics-table.svelte'
|
45 |
| - import { references } from './references.yaml' |
46 |
| - import { References } from '$lib' |
47 | 3 | import CumulativeClfMetrics from '$figs/cumulative-clf-metrics.svelte'
|
48 | 4 | import RollingMaeModels from '$figs/rolling-mae-vs-hull-dist-models.svelte'
|
49 | 5 | import { browser } from '$app/environment'
|
50 | 6 | </script>
|
51 | 7 |
|
52 |
| -# {title}<br><small>{subtitle}</small> |
53 |
| - |
54 |
| -<address> |
55 |
| - <span rel="author"> |
56 |
| - {@html authors |
57 |
| - .map((author) => `${author.name}<sup>${author.affiliation}</sup>`) |
58 |
| - .join(`, `)} |
59 |
| - </span> |
60 |
| - <span data-rel="affiliation"> |
61 |
| - {@html affiliations.map((affil, idx) => `${idx + 1}. ${affil}`).join(`<br/>`)} |
62 |
| - </span> |
63 |
| - <span style="font-weight: lighter;">{date}</span> |
64 |
| -</address> |
| 8 | +<summary> |
65 | 9 |
|
66 |
| -## Abstract |
| 10 | +We present a new machine learning (ML) benchmark for materials stability predictions named `Matbench Discovery`. A goal of this benchmark is to highlight the need to focus on metrics that directly measure their utility in prospective discovery campaigns as opposed to analyzing models based on predictive accuracy alone. Our benchmark consists of a task designed to closely simulate the deployment of ML energy models in a high-throughput search for stable inorganic crystals. We explore a wide variety of models covering multiple methodologies ranging from random forests to GNNs, and from one-shot predictors to iterative Bayesian optimizers and interatomic potential-based relaxers. We find M3GNet to achieve the highest F1 score of 0.58 and $R^2$ of 0.59 while MEGNet wins on discovery acceleration factor (DAF) with 2.94. Our results provide valuable insights for maintainers of high throughput materials databases to start using these models as triaging steps to more effectively allocate compute for DFT relaxations. |
67 | 11 |
|
68 |
| -We present a new machine learning benchmark for materials stability predictions called **Matbench Discovery**. The primary goal of this benchmark is to evaluate the effectiveness of machine learning energy models at accelerating the search for inorganic crystals and to determine the optimal methodology for this task. Specifically, we aim to answer the question of whether density functional theory emulators like M3GNet or one-shot predictors like Wrenformer perform better. To make the results easily accessible, we provide an online leaderboard with interactive plots that allow for custom model comparisons and looking at a variety of performance metrics. New models are easily added to our leaderboard, allowing this benchmark to grow into the future. We also make our results easily reproducible by releasing all training and test scripts as well as WandB logs of all our experiments. We hope our results provide valuable insights that motivate researchers in the field of materials discovery and builders of high throughput databases to start using these models as triaging steps to more effectively allocate compute for DFT relaxations. |
| 12 | +</summary> |
69 | 13 |
|
70 | 14 | ## Introduction
|
71 | 15 |
|
@@ -195,41 +139,3 @@ M3GNet's performance appears to be particularly affected by poor performance on
|
195 | 139 | ## Acknowledgements
|
196 | 140 |
|
197 | 141 | JR acknowledges support from the German Academic Scholarship Foundation (Studienstiftung) and gracious hosting as a visiting affiliate in the groups of [KP](https://perssongroup.lbl.gov/people) and [AJ](https://hackingmaterials.lbl.gov).
|
198 |
| - |
199 |
| -## References |
200 |
| - |
201 |
| -<References {references} /> |
202 |
| - |
203 |
| -<style> |
204 |
| - #abstract, |
205 |
| - #abstract + p { |
206 |
| - font-weight: 300; |
207 |
| - } |
208 |
| - address { |
209 |
| - text-align: center; |
210 |
| - } |
211 |
| - address span { |
212 |
| - margin: 1em; |
213 |
| - display: block; |
214 |
| - } |
215 |
| - /*auto-number HTML headings*/ |
216 |
| - h2 { |
217 |
| - counter-increment: h2; |
218 |
| - counter-reset: h3, h4; |
219 |
| - } |
220 |
| - h2::before { |
221 |
| - content: counter(h2); |
222 |
| - margin-right: 10pt; |
223 |
| - } |
224 |
| - h3 { |
225 |
| - counter-increment: h3; |
226 |
| - counter-reset: h4; |
227 |
| - } |
228 |
| - h3::before { |
229 |
| - content: counter(h2) '.' counter(h3); |
230 |
| - margin-right: 10pt; |
231 |
| - } |
232 |
| - sup.ref { |
233 |
| - text-transform: capitalize; |
234 |
| - } |
235 |
| -</style> |
0 commit comments