## 1 Introduction

“The basis of all human culture is language, and mathematics is a special kind of linguistic activity.” — Arnold and Manin (2000)

A growing number of machine learning works leverage datasets of mathematical expressions to perform various symbolic reasoning tasks. These tasks include symbolic regression (Koza, 1992; Petersen et al., 2021), symbolic integration (Lample and Charton, 2019), solving differential equations, (Lample and Charton, 2019), and freeform mathematical question/answering (Saxton et al., 2019)

. The expression datasets used to complete these tasks are typically generated procedurally using various rules or heuristics

(Lample and Charton, 2019; Saxton et al., 2019), crafted by hand (Uy et al., 2011), or manually extracted from existing texts (Udrescu and Tegmark, 2020). Yet, to the best of our knowledge, there exists no large-scale, customizable, and/or widely used dataset of mathematical expressions.We introduce a simple pipeline for extracting mathematical expression encodings directly from raw Wikipedia text. Wikipedia is an excellent source for extracting mathematical knowledge for several reasons. First, Wikipedia is ever-growing and spans virtually all scientific domains. Second, expressions embedded in Wikipedia are conveniently annotated via XML tags and follow standard markup encodings (i.e. LaTeX), facilitating extraction and parsing. Third, Wikipedia is structured into hierarchical categories, allowing users to extract customized expression corpora, e.g. based on a particular branch of science. Finally, raw Wikipedia data dumps are frequently updated (without requiring web-crawling), easy to access, and freely available.

We demonstrate the practical use of our expression dataset by using it to train a mathematical language model, then using the trained model as a prior in a downstream neural-guided search task. Specifically, we consider symbolic regression, the task of searching the space of tractable mathematical expressions to fit a dataset. Symbolic regression is an excellent testbed problem for symbolic search because it poses a large combinatorial search space, is computationally inexpensive, and has well-established suites of benchmark problems (White et al., 2013). We demonstrate that leveraging a pre-trained mathematical language model as a prior to guide the search improves the ability to recover symbolic expressions from data, and provides other advantages such as reduced semantically invalid expressions.

## 2 Related Work

Datasets of expressions. Various expression datasets have been proposed for symbolic tasks. Lample and Charton (2019)

procedurally generate expressions to perform symbolic mathematics, specifically to perform integration and solve ordinary differential equations. Expressions are randomly generated with hand-crafted rules, followed by a cleaning process to simplify and remove invalid expressions.

Saxton et al. (2019) released a dataset for mathematical reasoning, which is also generated through rules that sample answer and generates matching questions. Udrescu and Tegmark (2020) introduce a dataset of 120 expressions to be used as benchmarks for symbolic regression, manually pulled from the Feynman lectures on physics (Feynman et al., 1965). To the best of our knowledge, there exists no large-scale dataset (or dataset-generating pipeline) of real-world symbolic expressions.Mathematics as language. The perspective of viewing mathematics as a form of language dates as far back as the history of mathematics itself (Arnold and Manin, 2000). Within machine learning, Lample and Charton (2019) recently addressed symbolic mathematics as a machine translation problem, representing mathematical expressions as sequences and solving mathematical problems with a seq2seq model. We also consider mathematics as language to create a mathematical language model based on human-created, widely-used mathematical expressions in Wikipedia.

## 3 Methods

Extracting an expression corpus from Wikipedia.
Our pipeline for extracting expressions from Wikipedia is illustrated in Fig. 1.
Wikipedia provides its raw text data as XML (Extensible Markup Language) “dump” files.^{1}^{1}1https://dumps.wikimedia.org
Since it is written in a markup language, users can easily parse this dataset of rich information.
Specifically, we extract embedded mathematical expressions, which are annotated via <math> tags (Fig. 1b-c).
Raw expressions are represented by LaTeX text, which we convert into a tree-based representation—called an algebraic expression tree—via the computer algebra system SymPy (Meurer et al., 2017)^{2}^{2}2https://sympy.org/.
Finally, expressions are encoded as sequences of tokens corresponding to the pre-order traversal of the algebraic expression tree.
This data can then be readily used for downstream machine learning pipelines.

We also leverage Wikipedia’s hierarchical page categorizations.
As illustrated in Fig. 2, each category comprises subcategories and pages, forming a tree structure.
Pages are the user interface to access content in Wikipedia, while categories are metadata used to point users to related content.
To build a category tree, we need the edges connecting page and categories.
Edge information is only partially available in the XML dump files.
Therefore, we resort to the database provided by MediaWiki.^{3}^{3}3https://mediawiki.org
In particular, this database includes three relevant tables: Categorylinks, Category, and Page (Fig. 4 in the Appendix).
Using database management system such as MySQL, we use this to generate a category tree from a given root category.
Since categories may not be in strict hierarchy (i.e. they can be recursive), we specify a maximum search depth.
In this manner, users can customize the expression dataset to specific categories (e.g. scientific domain) of interest.

Training a mathematical language model.

By considering each symbolic token as a “word” from a dictionary of tokens, and each sequence of tokens (i.e. expression) as a “sentence,” we propose building a mathematical language model (MLM) to estimate the probability for a mathematical expression represented by a sequence of tokens.

Before training the MLM, we first filter and augment the dataset based on the operators and operands used in the downstream learning task. For instance, if the task does not use integrals, each expression containing an integral symbol could be simply removed. Instead, we choose two approaches to augment this data: replacing and splitting. When an integral appears while traversing the expressional tree, the whole subtree with a root of integral can be replaced with a terminal token. Another approach is splitting the tree to use the expression in the integrand for data augmentation.

Finally, we train a recurrent neural network (RNN) with our preprocessed and augmented expression dataset to create a MLM using the set of given symbols as the vocabulary. Specifically, we train the MLM to predict the next token in a sequence by minimizing the cross-entropy loss between the output of the RNN and the given label. This strategy is standard in natural language processing (NLP) for training language models that predict the next word or character given a partial sequence

(Mikolov et al., 2010).Integrating with deep symbolic regression. Deep symbolic regression (Petersen et al., 2021) uses an autoregressive recurrent neural network, or Controller, parameterized by , that emits a categorical distribution on which token to select next, conditioned on all previously selected tokens.

Here, we propose using the MLM to inform the Controller by directly adjusting the logits that are emitted by it, as in Fig. 3. Specifically, at time step , given input (e.g. the previously selected token), the Controller emits logits and updates its internal state :

Further, DSR allows incorporating in situ constraints to the search space (e.g. constraining nested trigonometric operators) via logits , whose values are either zero or negative infinity (see Petersen et al. (2021) for details). To incorporate the MLM into this generative process, we introduce a third vector of logits computed by the MLM:

Finally, the logit vectors are summed, a softmax operator is applied, and the resulting probability vector defines a categorical distribution used to sample the next token :

The hyperparameter

controls the strength of the MLM prior. We provide an interpretation of based on the temperature of the softmax function. Note the softmax function with temperature and logits is defined as: . Multiplying logits by yields:where inverse-softmax is defined up to an arbitrary constant. Thus, can be viewed as the inverse temperature of the MLM prior’s contribution to the final softmax. Higher temperatures (lower ) result in a lower influence of the MLM prior on the final softmax. At the extreme, corresponds to infinite temperature, in which case the MLM prior is ignored.

Pseudocode for DSR integrated with the MLM prior is provided in the Appendix. Note that the recurrent architecture used in the MLM can be completely different than the one used in DSR; this allows the MLM to be pre-trained offline.

## 4 Results and Discussion

Dataset statistics. From a raw Wikipedia dump file of 70 GB, we extracted 798,998 mathematical expressions across 41,763 different pages. As an example of using the category hierarchies, when we search pages under the category Physics with depth 3, we collect 67,404 expressions from 2,265 pages in 879 categories, while Physics itself has 25 pages with 1,374 expressions.

DSR without MLM | DSR with MLM | |||||

Benchmark | Recovery | Steps | Invalid | Recovery | Steps | Invalid |

100.% | 165.2 | 47.82 | 100.% | 99.47 | 56.75 | |

100.% | 264.6 | 35.86 | 100.% | 235.6 | 36.46 | |

100.% | 349.2 | 28.35 | 100.% | 329.1 | 26.93 | |

100.% | 672.2 | 17.65 | 100.% | 525.5 | 20.75 | |

76.% | 847.8 | 23.67 | 94.% | 672.8 | 22.81 | |

100.% | 189.3 | 45.26 | 100.% | 120.3 | 51.63 | |

35.% | 1513. | 11.02 | 27.% | 1620. | 10.91 | |

95.% | 601. | 31.81 | 99.% | 365.1 | 32.62 | |

100.% | 117.8 | 34.35 | 100.% | 95.52 | 27.34 | |

100.% | 368.3 | 17.55 | 100.% | 364.3 | 13.93 | |

100.% | 22.36 | 56.37 | 100.% | 12.58 | 41.13 | |

0.% | 2000. | 6.373 | 0.% | 2000. | 5.71 | |

Average: | 83.8% | 592.6 | 29.7% | 85.0% | 536.7 | 28.9% |

Application to symbolic regression. We demonstrate the value of the MLM by using it to inform the task of symbolic regression. For simplicity, we replicate the experimental setup and hyperparameters detailed in Petersen et al. (2021)

, leveraging the accompanying open-source package “Deep Symbolic Regression.” The only change is the introduction of the MLM prior as previously described, sweeping over inverse temperature hyperparameter

. The MLM is trained for 200 epochs using a recurrent neural network with a hidden layer size of 256, showing the cross-entropy loss going down to 0.367.

In Table 1, we report the recovery rate (fraction of runs in which the exact symbolic expression is found), average number of steps required to find the solution, and the fraction of invalid expressions (those that produce floating-point errors, e.g. overflows) produced during training, over 100 independent runs. The MLM prior show an improvement in recovery rate, requiring fewer steps and generating fewer invalid expressions. In particular, we note the dramatic reduction in steps required to find expressions , , and . Overall, these results show that the use of the MLM provides a better informed search in symbolic regression tasks.

## 5 Conclusion and Future Direction

We introduce a pipeline for generating a largescale “corpus” of mathematical expressions directly from Wikipedia, and demonstrate that a language model trained on this dataset can improve neural-guided search for the task of symbolic regression. Possible alternative uses of a MLM include auto-completion of writing expressions in LaTeX, or improving recognition of hand-written expressions.

### Acknowledgments

This work was performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under contract DE-AC52-07NA27344. Lawrence Livermore National Security, LLC. LLNL-CONF-820039.

## References

- Mathematics as profession and vocation. In Mathematics: Frontiers and Perspectives, pp. 153–159. Cited by: §1, §2.
- The feynman lectures on physics; vol. i. American Journal of Physics 33 (9), pp. 750–752. Cited by: §2.
- Genetic programming: on the programming of computers by means of natural selection. Vol. 1, MIT press. Cited by: §1.
- Deep learning for symbolic mathematics. arXiv preprint arXiv:1912.01412. Cited by: §1, §2, §2.
- SymPy: symbolic computing in python. PeerJ Computer Science 3, pp. e103. Cited by: §3.
- Recurrent neural network based language model. In Eleventh annual conference of the international speech communication association, Cited by: §3.
- Deep symbolic regression: recovering mathematical expressions from data via risk-seeking policy gradients. Proc. of the International Conference on Learning Representations. Cited by: §1, §3, §3, §4.
- Analysing mathematical reasoning abilities of neural models. arXiv preprint arXiv:1904.01557. Cited by: §1, §2.
- AI feynman: a physics-inspired method for symbolic regression. Science Advances 6 (16), pp. eaay2631. Cited by: §1, §2.
- Semantically-based crossover in genetic programming: application to real-valued symbolic regression. Genetic Programming and Evolvable Machines 12 (2), pp. 91–119. Cited by: §1.
- Better gp benchmarks: community survey results and proposals. Genetic Programming and Evolvable Machines 14 (1), pp. 3–29. Cited by: §1.

Comments

There are no comments yet.