Andrzej Odrzywo{ł}ek
Institute of Theoretical Physics, Jagiellonian University, 30-348 Krakow, Poland
E-mail: [email protected]
A single two-input gate suffices for all of Boolean logic in digital hardware. No comparable primitive has been known for continuous mathematics: computing elementary functions such as $\sin$, $\cos$, $\sqrt{\phantom{x}}$, and $\log$ has always required multiple distinct operations. Here we show that a single binary operator, $\operatorname{eml}(x,y)=\exp(x)-\ln(y),$ together with the constant $1$, generates the standard repertoire of a scientific calculator. This includes constants such as $e$, $\pi$, and $i$; arithmetic operations including $+$, $-$, $\times$, $/$, and exponentiation as well as the usual transcendental and algebraic functions. For example, $e^x= \operatorname{eml}(x,1)$, $\ln x= \operatorname{eml}(1, \operatorname{eml}(\operatorname{eml}(1,x),1))$, and likewise for all other operations. That such an operator exists was not anticipated; I found it by systematic exhaustive search and established constructively that it suffices for the concrete scientific-calculator basis. In EML (Exp-Minus-Log) form, every such expression becomes a binary tree of identical nodes, yielding a grammar as simple as $S!\to!1\mid \operatorname{eml}(S,S)$. This uniform structure also enables gradient-based symbolic regression: using EML trees as trainable circuits with standard optimizers (Adam), I demonstrate the feasibility of exact recovery of closed-form elementary functions from numerical data at shallow tree depths up to 4. The same architecture can fit arbitrary data, but when the generating law is elementary, it may recover the exact formula.
Executive Summary: The document explores a fundamental question in mathematics and computing: can the diverse set of elementary functions—such as sines, cosines, square roots, logarithms, and basic arithmetic—that power scientific calculations be generated from a single, simple operator, much like how a single NAND gate builds all digital logic? This matters now because these functions underpin everything from engineering simulations and data analysis to machine learning models and numerical solvers in physics and finance. Reducing them to one operator could streamline software, hardware, and AI tools, making computations more efficient and easier to discover or automate, especially as computational demands grow in fields like climate modeling and drug discovery.
The work set out to determine the minimal set of operations needed to reconstruct a standard scientific calculator's functions and to identify if a single binary operator could suffice for all of them.
To achieve this, the author started with a baseline of 36 common elements—constants like pi and e, unary functions like square root and sine, and binary operations like addition and exponentiation—and systematically tested reductions by removing items one by one. Using computational tools including Mathematica for symbolic checks and a custom Rust program for fast numerical verification, the process involved generating candidate expressions up to moderate complexity (equivalent to formula lengths of 7-9 steps) and confirming they matched targets across diverse input points, such as transcendental constants, to ensure general validity. This ablation and exhaustive search spanned months, drawing on hybrid numeric-symbolic methods to handle the vast search space without full symbolic proofs upfront.
The key discovery is a single binary operator called EML, defined as exp(x) minus the natural log of y, which, paired with the constant 1, reconstructs all 36 baseline elements—covering arithmetic, trigonometry, hyperbolics, inverses, and constants like pi, e, and i—through nested applications forming binary trees up to depth 8 for complex operations like multiplication. For instance, exponentiation simplifies to EML(x, 1), while logarithm requires a three-level nesting, and trigonometric functions emerge via complex intermediates. This reduces a 36-button calculator to effectively two: the digit 1 and EML. Close variants exist, such as one using constant e instead of 1, but EML proved the most versatile in tests. Finally, EML trees enable gradient-based optimization to recover exact formulas, like log(x), from noisy data in 100% of trials at shallow depths (up to 4 levels), snapping parameters to precise 0-1 values for symbolic exactness.
These results reveal that elementary functions form a far simpler, more unified class than traditionally viewed, akin to how digital circuits use identical gates—potentially cutting implementation costs in software libraries and hardware by standardizing evaluations, while improving safety and performance in numerical methods prone to diverse operator errors. Unlike prior reductions that stalled at 4-7 primitives (e.g., exp, log, and addition), EML achieves full coverage with one operator, challenging assumptions about inherent diversity and opening doors to interpretable AI that discovers laws from data, rather than opaque approximations. This could accelerate scientific workflows, from equation solving to formula invention, though it requires handling complex numbers internally even for real outputs.
Next steps should include integrating EML into open-source tools like Python's NumPy or specialized compilers for efficient evaluation, and piloting it in symbolic regression libraries to test formula discovery on real datasets from physics or biology. Developers could explore trade-offs: EML excels in uniformity but needs a fixed constant (1), so searching for a self-generating variant without one—perhaps a ternary operator—would enhance flexibility, though it might increase search complexity. For broader adoption, refine optimization for deeper trees to handle more intricate laws.
Confidence in the findings is high, backed by reproducible code, cross-verification in multiple languages (C, Python, PyTorch), and symbolic proofs in supplements, with numerical errors below machine precision. Limitations include reliance on complex arithmetic for real functions, potential floating-point issues at edges like zero, and heuristic search that might miss subtler operators; caution is advised for ultra-precise applications without custom handling.
Elementary functions such as exponentiation, logarithms and trigonometric functions are the standard vocabulary of STEM education. Each comes with its own rules and a dedicated button on a scientific calculator; our derivations rely on many of them simultaneously, even though we know they are heavily redundant and can be expressed through one another, e.g. $\sin{x} = \cos(x-\pi/2)$, $\sqrt{x}=x^{1/2}$, etc. They are the workhorse of quantitative science, appearing in basic and empirical laws and inside the engines of numerical methods like differential equation solvers, integration quadratures and Fourier analysis [1]. In digital electronics, a remarkable fact underlies universality: a single two-input gate, NAND (the Sheffer stroke), suffices to build any Boolean circuit [2]. Continuous mathematics has lacked such a primitive: calculators must expose many distinct buttons. Classical reductions, from logarithm tables [3, 4] and the slide rule through Euler's formula [5] to the exp-log representation [6] with algebraic adjunctions [7], reduced them to a few, but no further. Despite this, it remains unclear whether this apparent diversity is intrinsic, or whether a smaller generative basis exists. Here we show that the operator $\mathrm{eml}(x, y) = \exp(x) - \ln(y)$, together with the constant 1, does exactly that: it reconstructs arithmetic, all standard elementary transcendental functions, and constants including integers, fractions, radicals, $e$, $\pi$ and $i$. In simpler terms, a two-button calculator (1, eml) suffices for everything a full scientific calculator can do. Existence of the EML operator reveals that elementary functions are members of a much simpler class than previously recognized. Every EML expression is a binary tree of identical nodes, yielding an exceptionally simple grammar: $S \to 1 \mid \mathrm{eml}(S, S)$, a context-free language that is isomorphic to well-studied combinatorial objects like full binary trees and Catalan structures. Elementary formulas become circuits [8] composed of identical elements, much like digital hardware built from NAND gates. This uniform representation provides a complete and regular search space for continuous symbolic regression [9, 10]: parameterized EML trees can be optimized by standard gradient methods, and when the generating law is elementary, trained weights can snap to exact closed-form expressions. In effect, a single trainable architecture gains the potential to discover [11] any elementary formula from data. The EML operator may be the tip of an iceberg. Preliminary searches have already returned related operators with even stronger properties, including a ternary variant that requires no distinguished constant.
Everyone learns many mathematical operations in school: fractions, roots, logarithms, and trigonometric functions ($+, -, \times, /, \sqrt{\phantom{x}}, \sin, \cos, \log, \ldots$), each with its own rules and a dedicated button on a scientific calculator. Higher mathematics reveals that many of these are redundant: for example, trigonometric ones reduce to the complex exponential. How far can this reduction go? We show that it goes all the way: a single operation, $\operatorname{eml}(x, y)$, replaces every one of them. A calculator with just two buttons, EML and the digit 1, can compute everything a full scientific calculator does. This is not a mere mathematical trick. Because one repeatable element suffices, mathematical expressions become uniform circuits, much like electronics built from identical transistors, opening new ways to encoding, evaluating, and discovering formulas across scientific computing.
Section Summary: Single, reusable basic elements, like the NAND gate in logic or ReLU in machine learning, play a huge role in math, engineering, and biology by simplifying complex systems, even though experts debate their fundamental necessity. Everyday math functions such as sine, cosine, addition, and multiplication seem essential but are often redundant and can be rebuilt from simpler operations, as shown historically by logarithms turning multiplication into addition and Euler's formula linking exponentials to trigonometry. This paper explores reducing all standard real elementary functions to a single binary operator called EML, defined as exp(x) minus ln(y), paired with key constants, allowing practical computations with just repeated applications of this one operation.
Single, reusable primitives play a disproportionately large aesthetic and practical role in mathematics, engineering, and even biology. Widely known classical examples include the NAND gate (and its dual, Peirce Arrow, logical NOR) for Boolean 0/1 logic [2, 12], the operational amplifier [13] for positive and negative feedback processes, and, more recently, the rectified linear unit (ReLU) ''ramp'' activation function [14] in deep learning [15]. We also mention Wolfram's single axiom [16], K, S combinators from combinatory logic [17, 18], Interaction Combinators [19], and fuzzy versions of the Sheffer stroke [20]. Other well-known examples are one-instruction set computers (OISC), e.g. SUBLEQ [21], Conway's FRACTRAN [22] and the Rule 110 cellular automaton [16, 23].
Whether the existence of a single sufficient operator or element is conceptually crucial is disputed [24]. Nevertheless, its value for understanding, pedagogy, and public communication is substantial. Indeed, classical first-order logic does not need to single out NAND at all: it works perfectly well with a whole redundant family of connectives (XOR, AND, NOT, $\ldots)$ that we introduce for convenience in applications. But the realization that all of them are definable from a single primitive is a striking and conceptually deep structural fact. A similar kind of realization accompanies the recognition of DNA and RNA [25] as a nearly universal information carrier in evolutionary biology.
Sheffer-type elements are rare, and mining them typically requires time, compute, effort, and a bit of luck. They also appear in seemingly distant areas, for example in discrete geometry. The recently discovered solution to the “einstein[^1] problem” had a major impact on the tiling community [26]. The present author already has one such element in his portfolio, namely an igloo construction brick derived from the (2, 3, 5) spherical Möbius triangle [27].
[^1]: From German ein Stein, “one tile, ” not after Albert Einstein.
Elementary functions, for many students epitomized by the dreaded sine and cosine, play a central role in quantitative reasoning. They can be combined in countably many ways with integers, rational numbers, and mathematical constants such as $\pi$ and $e$, using the four basic arithmetic operations ($+, -, \times, /$) and exponentiation. Elementary functions are at the core of STEM education and at the foundation of modern technological civilization. This includes not only simple formulas or empirical models, but also most numerical algorithms, for example ordinary differential equation solver RKF45 [28], integration rules (Gaussian, Tanh-Sinh, $\ldots) [$ 29, 1], and the Fast Fourier Transform [30], whose twiddle factors are elementary functions. Finite expressions in elementary functions are supported[^2] in virtually all modern programming languages. While it is generally known that most standard functions and arithmetic operations are heavily redundant (e.g. $\tan{x} = \sin{x}/\cos{x}$, $\sqrt{x} = x^{1/2}$, $\operatorname{arsinh}{x} = \ln{(x+\sqrt{1+x^2})}$, $x/y = x \times y^{-1}$ etc.), especially in the complex domain (e.g. $\cosh{x} = \cos{(ix)}$, $\sinh{i x} = i \sin{x}$), they have never been regarded as candidates for a single primitive operator. Historically, two major milestones in understanding elementary arithmetic operations from this point of view are (i) the discovery of logarithms and the subsequent creation of tables [3, 4] followed by slide rule, and (ii) Euler’s formula $e^{i \pi}+1=0$ [5]. Logarithms were introduced to reduce multiplication to addition. The exp-log pair allows them to be expressed in terms of one another:
[^2]: Exponentiation, unlike the four basic arithmetic operations that are almost always built in as primitive operators, has often been treated as a higher-level operation exposed via library functions rather than as a dedicated infix operator.
$ x \times y = e^{\ln{x} + \ln{y}}, \quad x + y = \ln{\left(e^x \times e^y\right)}.\tag{1} $
Euler’s formula
$ e^{i \varphi}=\cos{\varphi} + i \sin{\varphi}\tag{2} $
shows that, once the imaginary unit $i!=!\sqrt{-1}$ is introduced, trigonometric functions can be viewed as another face of ' $\exp$ ' and ' $\ln$ ' [6]. This perspective leads naturally to what scientists call exp–log functions, namely finite expressions built from variables, named constants, arithmetic operations, together with $\exp$ and $\log$, in the spirit of differential and computer algebra [31, 32]. In the classical differential-algebraic setting, one often works with a broader notion of elementary function, defined relative to a chosen field of constants and allowing algebraic adjunctions [7], i.e., adjoining roots of polynomial equations (cf. Root in Wolfram Language [33]). That level of generality is not needed here. The present paper takes the ordinary scientific-calculator point of view: start from a concrete list of familiar constants, functions and operations, and ask how far they can be reduced without losing practical functionality. The precise starting list is given later in Table 1.
Many reductions inside that list are, of course, classical and well known; what seems not to have been investigated systematically is the endpoint of the process, namely reduction to a single binary operator paired with a single distinguished constant. In particular, the so-called ''broken calculator'' problem [34] is a popular formulation of computations with a reduced set of available keys or operations. I used this approach to construct a sequence of increasingly “primitive” yet fully functional calculators with 4, 3, 2, and finally a single operator. This operator is the EML (Exp–Minus–Log),
$ \operatorname{eml}{(x, y)} = \exp{(x)} - \ln{(y)}.\tag{3} $
Using the EML, a surprisingly simple binary operator, we can express any standard real elementary function in terms of repeated applications of Equation 3 to the input variables $x, y, z, \ldots$ and a single distinguished constant, 1. This constant is needed to neutralize the logarithmic term in Equation 3 via $\ln{(1)}!=!0$. Computations must be done in the complex domain, e.g: generating constants like $i$ and $\pi$ requires evaluating $\ln(-1)$, so $\operatorname{eml}(x, y)$ internally operates over $\mathbb{C}$ using the principal branch.
If we do not need external input variables, e.g. for use in an actual pocket calculator, two buttons are sufficient to retain full functionality: one binary operator, EML, and one terminal symbol, 1. No further reduction of operator count is possible, because at least one binary operator and at least one terminal symbol are required. The existence of such a binary operator, which is itself an elementary function, is somewhat unexpected.
In fact, the EML Sheffer operator Equation 3 is as simple as it appears, and in principle the article could end here; its consequences would eventually surface on their own. Nevertheless, it seems worthwhile to explain the methods used in the search for it (Section 2), to present the intermediate and final search results (Section 3), and to outline potential applications (Section 4) that are already visible on the horizon. The article concludes with a short summary, possible follow-up directions, and open questions.
Section Summary: The researcher used a methodical approach called ablation testing to identify the smallest possible set of basic operations needed to rebuild all 36 elements of a scientific calculator's toolkit, starting from constants like pi and e, unary functions like square roots and sines, and binary operations like addition and exponentiation. By repeatedly removing one element at a time and checking if the rest could still generate the originals, the goal was to see if an already efficient system, like Mathematica's, could be simplified further. Direct symbolic checks were too complex, so a hybrid numeric method was employed, substituting special mathematical constants into equations and using software to verify matches numerically, with full proofs provided separately.
I employed systematic 'ablation' testing to identify minimal operator sets for a fixed scientific-calculator starting list. The list, given in Table 1, contains 36 primitives: named constants and input variables, standard unary functions, and binary operations. I iteratively removed one element from this list (constant, function or binary operation) and verified whether the remaining set could still reconstruct original primitives. The Mathematica/Wolfram core language [33] instruction set (Table 2, 2nd row) served as a reference for an already-optimized and thoroughly tested for almost 40 years minimalist system. Research started out of curiosity whether this system could be reduced further.
:Table 1: Starting list of constants/variables, functions (unary) and binary operators used to initiate the reduction procedure. This table also fixes the concrete scientific-calculator notion of "elementary function" used throughout the paper: finite expressions built from these named symbols. The $\exp{(x)} = e^x$, $\ln{x}$ is natural logarithm, $\operatorname{inv}{(x)} = 1/x$ denotes reciprocal, $\operatorname{minus}{(x)} = -x$ is a sign flip. Square is denoted by $\operatorname{sqr}{(x)} = x^2$ and $\sigma{(x)} = 1/(1+e^{-x})$ is the logistic sigmoid. Trigonometric and hyperbolic functions with their inverses have usual meaning. In addition to four basic arithmetic binary operations we use arbitrary base logarithm $\log_x{y}$, exponentiation $\operatorname{pow}{(x, y)} = x^y$, arithmetic mean $\operatorname{avg}{(x, y)} = (x+y)/2$ and hypotenuse $\operatorname{hypot}{(x, y)} = \sqrt{x^2+y^2}$.
| Type | Elements | Count |
|---|---|---|
| Constants | $\pi, e, i, -1, 1, 2, x, y$ | 8 |
| Functions | $\exp, \ln, \operatorname{inv}, \operatorname{half}, \operatorname{minus}, \sqrt{\quad}, \operatorname{sqr}, \sigma$, $\sin, \cos, \tan, \arcsin, \arccos, \arctan$, $\sinh, \cosh, \tanh, \operatorname{arsinh}, \operatorname{arcosh}, \operatorname{artanh}$ | 20 |
| Operations | $+, -, \times, /$, $\log, \operatorname{pow}$, $\operatorname{avg}, \operatorname{hypot}$ | 8 |
| Total | 36 |
An operator set was deemed "complete" if it could reconstruct all primitives from Table 1, on the real axis where appropriate. This includes trigonometric functions and their inverses ($\sin$, $\cos$, $\tan$, $\arcsin$, $\ldots)$, hyperbolic functions and their inverses ($\sinh$, $\cosh$, $\tanh$, $\operatorname{arsinh}$, $\ldots)$, algebraic operations ($\sqrt{\cdot}$, reciprocal, $\ldots)$, and fundamental constants ($\pi$, $e$, $i$, -1, 0, 1, 2, $\ldots)$.
The challenge of executing such a test is illustrated by the following simple exercise from rational functions generation, attributed to [35]. Given only the three operations: $\operatorname{suc}(x) = x+1$ (successor), $\operatorname{pre}(x) = x-1$ (predecessor), and $\operatorname{inv}(x) = 1/x$ (reciprocal/inverse), let us compute negation, i.e., $-x$. The non-obvious solution is:
$ \operatorname{suc}(\operatorname{inv}(\operatorname{pre}(\operatorname{inv}(\operatorname{suc}(\operatorname{inv}(x)))))) = \frac{1}{\frac{1}{\frac{1}{x}+1}-1}+1 = -x. $
This exemplifies the formulas expected to be encountered in our search: nested expressions with depth 5-9 that defy intuitive construction. There is no regular method to find them automatically, except for (human-assisted) brute-force search, nowadays also enhanced using AI [36].
Direct symbolic verification proved computationally intractable. Typical Kolmogorov-style complexity $K$ [37] is usually $K \lesssim 7$. In practice it was searched up to $K=9$. Here K denotes the length of a Reverse Polish Notation (RPN) calculator program, equivalent to a formula. To speed up the search I developed a hybrid numeric bootstrapping verification approach, working as follows. We test whether a target operation, e.g. hypotenuse $\sqrt{x^2+y^2}$, can be expressed using a given operator set. Instead of symbolic manipulation on $x, y$, I substituted for $x$ and $y$ algebraically independent (not in exp-log class) transcendental constants, such as e.g., $x=\gamma \simeq 0.577216$ (Euler-Mascheroni constant), $y = A \simeq 1.28243$ (Glaisher-Kinkelin constant). Then, $\sqrt{\gamma^2+A^2}$ was evaluated numerically, and compared to all expressions generated from $1$ and 3. This method follows from the simple observation that any generally valid formula must also be valid at a single point. Under the Schanuel conjecture [38], coincidental equality between such expressions is vanishingly unlikely, making this a reliable sieve for formula equivalence candidates. In practice, I compute the double-precision numerical value of the operator sought, and then run constant recognition software (''inverse symbolic calculator'') on this result, which returns a candidate formula in the form of RPN calculator code [39] or as an expression tree, depending on the variant of the method used. The search procedure is heuristic and serves only to exclude evidently false formulas and discover candidate witnesses; independent verification is separate and is given in Supplementary Information (Part II), which provides symbolic checks, numerical cross-validation, and a constructive completeness proof sketch for the Table 1 class.
In detail, the search procedure worked as follows. Rather than searching for all elementary operations (Table 1) directly in pure EML form (which remains computationally infeasible even numerically), I employed iterative bootstrapping:
The verification procedure, VerifyBaseSet, was implemented using my own Mathematica SymbolicRegression package [39]. A version that is three orders of magnitude faster was recently translated by GPT Codex 5.3 into Rust [39], allowing for re-check of EML in seconds, rather than hours.
The ablation process, i.e. running above search with some of the elements from Table 1 removed, yielded progressively smaller calculator configurations (named Calc 3, 2, 1, 0 in Table 2), each requiring different primitives, constants in particular. Some configurations could generate required constants from arbitrary input on their own (via operations like $x-x=0, e^0=1, 0-1=-1, \ldots$ in Calc 2), while others required specific constants (e.g., Calc 1 requires $e$ or $\pi$, Wolfram requires $i$, cf. footnote in Table 2). Then the search stalled, and it became evident that the continuous Sheffer operator, if it exists, is not among the familiar named functions. For the final reduction, I began, guided by intuition gathered over many experiments, to enumerate elementary binary functions as candidate single operators, paired with similarly generated constants. Each candidate pair constant, operator was tested using the VerifyBaseSet procedure. This, after numerous failures, and a few discarded false-positives, revealed that the $\operatorname{eml}(x, y)$, given by Equation 3, paired with the constant 1, forms a complete basis for reconstructing the elementary functions. Later I realized that EML is not unique, discovering its close cousins: EDL (requiring constant $e$) and a third variant using $-\infty$ as a terminal symbol. Details are the subject of the next section.
Section Summary: The research demonstrates a progressive simplification of the basic operations needed to perform all functions on a scientific calculator, reducing from 36 initial buttons down to just three primitives: the constant 1 and a single binary operator called EML, defined as exp(x) minus ln(y). This minimal set, along with related variants like EDL, allows computation of essential math functions such as exponentials, logarithms, and multiplications through nested applications, as shown in a tree-like diagram tracing their derivations. Verification through code and detailed checks confirms that EML can regenerate all 36 elementary operations, paving the way for even simpler undiscovered primitives.

:Table 2: Progressive reduction from 36-button scientific calculator from Table 1 to the EML Sheffer operator. Calc 3, 2 and 0 accept any real constant or variable as input, generating all required values internally. The count-down column tallies all distinct primitives including one input variable; for systems marked ''none'' under Constants, any real value serves as the starting input. Wolfram Mathematica primitives are given for reference. Last lines are placeholders for speculated undiscovered primitives. The penultimate row is for a binary operator, hypothesized to generate, unlike EML, constants from arbitrary input. The last row is for a unary operator, which retains all good properties of well-tested activation functions in neural nets, while allowing for exact evaluation of elementary functions when combined with standard ''matrix'' arithmetic.
| Constants | Unary | Binary | Count-down | |
|---|---|---|---|---|
| Base-36 | cf. Table 1 | cf. Table 1 | cf. Table 1 | 36 |
| Wolfram$^{*}$ | $\pi, e, i$ | $\ln$ | $+, \times, \wedge$ | 7 |
| Calc 3 | none | $\exp, \ln, -x, 1/x$ | $+$ | 6 |
| Calc 2 | none | $\exp, \ln$ | $-$ | 4 |
| Calc 1 | $e$ or $\pi$ | none | $x^y, \log_{x}{y}$ | 4 |
| Calc 0 | none | $\exp$ | $\log_{x}{y}$ | 3 |
| EML | $1$ | none | $\operatorname{eml}{(x, y)}$ | 3 |
| ? | none | none | ? | 2 |
| ? | none | ? | $+, -, \times$ | 5 |
Table 2 presents the complete reduction sequence, in historical order. Each configuration represents a complete 'scientific calculator' capable of computing every expression built from the primitives listed in Table 1. Calc 3 (Table 2, 3rd row) was first to dethrone Wolfram Language primitives set (Table 2, 2nd row) by retaining negation and reciprocal alongside $\exp$, $\ln$, and addition. This system of 6 primitives is able to generate constants on its own via $x + (-x) = 0$, etc. Calc 2 achieved what Calc 3 does using only $\exp$, $\ln$ and subtraction, while still preserving ability to generate constants on its own. The non-commutative operator (subtraction) proved crucial, as it provides both expression-tree growth and inversion capabilities.
Calc 1 represents a fundamentally different, top-down approach, using binary exponentiation (a rank-3 hyper-operation) and its inverse (binary logarithm) as base operators instead of lower-rank operations such as addition and subtraction. This configuration works with $e$ or $\pi$ as the terminal constant. Despite extensive searches, no other constant was found for which Calc 1 works.
Calc 0 absorbs the constant $e$ into the $\exp$ function itself, reducing the system to 3 primitives. This configuration strongly suggested that a single binary operator might exist at all, motivating further search.
The final reduction to EML was achieved by recognizing a pattern: all minimal configurations involve pairs of inverse functions (including self-inverses) and non-commutative operations. Testing combinations of inverse functions at the input with asymmetric operations at the output yielded the first continuous sufficient operator. A month later I realized that it has at least two additional cousins: EDL and -EML:
$ \begin{align} \tag{a} \operatorname{eml}(x, y) &= \exp(x) - \ln(y) && \text{with constant } 1, \ \tag{b} \operatorname{edl}(x, y) &= \exp(x)/\ln(y) && \text{with constant } e, \ \tag{c}
The first successful discovery search run (cf. Figure 1) can be replicated with this 3-line Mathematica code:
Import["SymbolicRegression.m"]
EML[x_, y_] := Exp[x] - Log[y]
VerifyBaseSet[{1}, {}, {EML}]
where the package can also be imported directly from the repository [39]. Depending on computer speed, usually in less than an hour, the above procedure re-generates all 36 elementary operations from Table 1. For example, the natural logarithm becomes:
$ \ln{(z)} = \operatorname{eml}\left(1, \operatorname{eml}[\operatorname{eml}(1, z), 1] \right),\tag{5} $
and so on. The resulting EML expressions range from depth 1 (exponential, $e^x = \operatorname{eml}(x, 1)$) to depth 8 (multiplication), with most basic math functions requiring larger depths (Table 4). A much faster and more thorough test (multiple real transcendentals, both positive and negative, arbitrary-precision check) is provided by a Rust re-implementation of VerifyBaseSet [39], rust_verify.
The exhaustive search served only to discover candidate identities. Their verification is deferred to Supplementary Information, Part II, which contains symbolic simplification of the full discovery chain (Figure 1) in Wolfram Mathematica [33], independent numerical cross-checks across four implementations (C, NumPy, PyTorch, and mpmath), and a constructive completeness proof sketch.
Section Summary: The section explores practical ways to use EML expressions, starting with a Python-based compiler that transforms standard math formulas into a simplified EML form, which can then be computed symbolically in tools like Mathematica or numerically in programming languages, and even run on specialized hardware like a single-instruction calculator. This uniform structure represents elementary functions as binary trees, offering advantages for both digital and analog computing, such as building circuits for operations like logarithms or multiplications. While the approach handles most real numbers well, it requires adjustments for edge cases like zero or complex branches to ensure accuracy across different software environments.
The uniform tree structure of EML expressions suggests several directions for implementation and application.
The output of the VerifyBaseSet procedure provides the data (see Figure 1) required to reconstruct any primitive or composite elementary expression in terms of EML Sheffer, Equation 4a. I provide a prototype EML compiler, coded in Python, that converts formulas into pure EML form. An EML expression can be evaluated symbolically in Mathematica, or numerically in any IEEE754-compliant language. Pure EML form could also be executed on hardware (or an emulated machine) that has only a single instruction: the EML itself. In particular, the EML code can be executed on a single instruction stack machine, closely resembling a single-button RPN calculator. Pure-EML form could possibly be implemented efficiently in FPGA or analog circuits.
The simplest input-output example is provided by the $\ln x$ function. After compilation to EML form, we expect to obtain Equation 5, although the simplest possible form is not expected in general. The equivalent RPN code for $\ln$ is a sequence of $K=7$ instructions
$ {1, 1, x, \operatorname{eml}, 1, \operatorname{eml}, \operatorname{eml}}, $
or, denoting $\operatorname{eml} \to \text{E}$, as the RPN string 11xE1EE. For the expression tree, see Figure 2 (on top).
A few comments on the implementation are required. EML-compiled expressions work on the real axis, both positive and negative, except for a few isolated points, especially at zero and domain endpoints. Internal computations, for trigonometric functions in particular, must be done in the complex domain. Because the simplest form of the natural logarithm, Equation 5, obtained from EML, Equation 3, is equivalent to
$ \ln{z} = e- \log{\left(\frac{e^e}{z} \right)}, $
using the standard choice of principal branch for complex logarithm, we obtain a jump of $2 \pi i$ for the negative real axis, due to the $1/z$ term. This prevents use of e.g. some standard formulas for $i$, relying on $\ln(-1) = i \pi$, for which we get the wrong sign. A solution working for all real $x \neq 0$ is to redefine the branch for EML itself in such a way that $\ln{z}$ (and everything derived from it, cf. Figure 1) follows standard implementation of principal branch. Another option, used in EML compiler, is to manually correct $i$ sign.
EML-compiled formulas work flawlessly in symbolic Mathematica and IEEE754 floating-point, e.g. <math.h> in C. This is because some formulas internally might rely on the following properties of extended reals:
$ \ln{0} = -\infty, \quad e^{-\infty} = 0. $
These are properly implemented in Mathematica using symbolic processing, and in floating point using inf and signed zeros. But EML expressions in general do not work 'out of the box' in, e.g., pure Python/Julia or numerical Mathematica. In the first case, this is because special floats are trapped and raise errors. However, EML works in NumPy [40] and PyTorch [41]. In Mathematica, we have an automatically extensible range of floats leading to Overflow[]. Interestingly, the Lean 4 proof assistant [42] takes a different approach. Because Lean requires all functions to be total, it assigns the complex logarithm at zero a default "junk value" (Complex.log 0 = 0), causing the straightforward formalization of the EML chain to fail. As a bottom line I stress that all the above difficulties (edge cases) are not much different from those usually encountered in every kind of floating-point or symbolic computation. EML compiler is available for testing under EML_toolkit/EmL_compiler subdir of [39]; see also Supplementary Information, Sect. 4. Software and reproducibility.
Noteworthy, in EML notation, any elementary function expression tree is binary. The context-free grammar is trivial: $S \to 1 \mid \operatorname{eml}{(S, S)}$. For functions, input variables are added as additional terminal symbols (e.g. $x$ in the univariate case). This has many practical advantages over standard methods.

Some simple examples of tree/circuit representations are shown in Figure 2. The examples shown are natural logarithm, identity, negation $\operatorname{minus}(x)=-x$, reciprocal $\operatorname{inv}(x)=1/x$, and multiplication. Ability to compute the identity function using an EML tree of depth 4 allows some input variables to be moved down the tree (see next Subsection). Other elementary functions, e.g. trigonometric ones, have trees too large to be shown in print, cf. Table 4.
::: {caption="Table 3: EML Sheffer as a new kind of basic building block for symbolic and analog computations. Dot marks output, and, as Equation 3 is non-commutative, arrow determine chirality and operation order. First counterclockwise input after dot is $\exp{x}$, then $\ln{y}$, and dot EML ''perform'' subtraction. In Figure 2 we keep this convention to avoid confusion, but clockwise variant is valid as well. I'm following habits of OpAmp symbol usage here."}

:::
Circuits using EML operator as a new element (Table 3) might be useful for analog computing [8]. One of the old problems in this field is construction of predefined multivariate elementary functions [43]. Using EML compiler (Section 4.1), we can convert any expression (Figure 2) to such a circuit, with the topology of a binary tree. EML tree provides a uniform treatment of generic elementary functions.
::: {caption="Table 4: Complexity of various functions in EML tree representation. EML Compiler column gives RPN code length $K$ for expressions generated from EML compiler. The value of $K$ of EML formula can be computed using e.g. Mathematica LeafCount. For the identity function $x$, the compiler returns $x$ directly (leaf count 1); the shortest non-trivial EML expression have leaf count 9. Last column show results of direct exhaustive search for shortest expressions. Numbers in parentheses show length of formulas which do not use the extended reals ($\pm$ inf in floating-point). If search timed out, reached lower limit for $K$ is given."}

:::
Modern symbolic regression (SR) methods [9, 10, 44] aim to discover closed-form expressions from data, but they typically search over heterogeneous grammars involving many distinct operators. Knowledge of a single operator, Equation 3, which is sufficient to compute any elementary function, allows us to create a multiparameter ''master'' formula. Such a master formula, easily constructed in EML form due to binary expression tree, for some combinations of its parameters, is equivalent to elementary functions. By construction, such a general tree includes all possible formulas up to the given leaf count (tree depth). These trees are big by the standards of a typical mathematical analysis course, but small compared to what is in use in modern AI. For example, a full binary tree of depth $n$ has a total of $2^n-1$ branches and $2^n$ leaves. For the largest transformers with trillions ($10^{12}$) of parameters, the equivalent tree would have a depth of 40. The large values in Table 4 reflect the unoptimized prototype EML compiler (Subsect. Section 4.1); direct exhaustive search yields substantially shorter expressions, as the rightmost column demonstrates.
The master formula can be constructed as follows. For simplicity we present the univariate case; the construction extends to an arbitrary number of input variables. Both inputs to $\operatorname{eml}{(x, y)}$ can be: either 1, input variable $x$, or the result from the previous $\operatorname{eml}$, which we denote $f$. Let us represent such a general input by a linear combination
$ \alpha_i + \beta_i x + \gamma_i f.\tag{6} $
Now, setting specific values, we can recover all three cases
At lowest tree level (leaves), there are only 1 and x, so there are no parameters $\gamma_i$.
As an example, I present full level-2 master formula
$ \begin{align*} F(x) = \operatorname{eml} \biggl[&\alpha_1 + \beta_1 x + \gamma_1 \operatorname{eml}{(\alpha_3 + \beta_3 x, \alpha_4 + \beta_4 x)}, \ &\alpha_2 + \beta_2 x + \gamma_2 \operatorname{eml}{(\alpha_5 + \beta_5 x, \alpha_6 + \beta_6 x)} \biggr]. \end{align*} $
If we set $\alpha_1=0, \beta_1=1, \gamma_1=0$ and $\alpha_2=1, \beta_2=\gamma_2=0$, we recover $\exp(x)$. Using $\alpha_1=\alpha_2=1$ and all $\beta_i=\gamma_i=0$ we get the constant $e$. Setting $\alpha_1=\beta_1=0, \gamma_1=1$, $\alpha_2=1, \beta_2=\gamma_2=0$, $\alpha_3=0, \beta_3=1, \gamma_3=0$ and $\alpha_4=1, \beta_4=\gamma_4=0$, we obtain double exponential $\exp{(e^x)}$. The total number of parameters for the level-2 master formula is 14. In general, the level- $n$ master formula has $5 \times 2^n-6$ parameters, see Supplementary Information, Part III.
While one could, in principle, create such a master formula using more usual elementary functions (+, -, /, **, $\sin, \cos$, $\ldots$, etc.), it would be ridiculously complicated and would lack any regular structure. Moreover, in practical SR one typically works with a reduced subset of operations, running the risk that the chosen set is incomplete, i.e., unable to express all elementary functions. The EML master formula is complete by design.
Number of parameters in Equation 6 can be reduced to two using some switch function. Or, alternatively, in a more modern approach, one can treat parameters $\alpha_i, \beta_i, \gamma_i$ as logits, and pass them through softmax function [45] to convert them into probabilities normalized to $\alpha_i+\beta_i+\gamma_i=1$. In two examples below, I'll use both methods.
The simplest proof of concept is provided by fitting some numerical data obtained from example elementary function, the $\ln{x}$, using the complete level 3 binary tree with the above structure. Using simplex reparameterization, the number of free parameters reduces from 34 to 20. I managed to find all weights using plain NMinimize, a 'black-box' optimizer from Mathematica [33]. The mean squared fitting error is at the level of numerical precision, and the resulting formula, after rounding weights to the nearest vertex of the simplex (i.e. snapping each to 0 or 1), is exactly $\ln{x}$. Not only in the provided data range, but also beyond that. Generalization/extrapolation is nearly perfect in the above example. See Wolfram Mathematica notebook Log_fit.nb from EML toolkit [39] and Supplementary Information (Sect III.B) for details.
In practice, such a naive one-liner approach does not scale. Therefore, I performed several experiments using more recent techniques of Machine Learning. Simple functions of two variables, taken from composed EML, such as $\ln(e-\ln(e^x - \ln{y}))$, were used as targets. Python code was created, which used DTYPE = torch.complex128 data type. Main issues encountered during training of EML net were related to range overflow due to multiply composed exponentials, as well as floating-point errors (NaN) caused by the specific implementation of complex arithmetic. They can be eliminated by clamping arguments and values for $\exp{(x)}$, and careful (i.e. without impairing torch automatic differentiation) inspection of both real and imaginary parts. Parameters from Equation 6 were treated as logits. Optimization was a multi-stage process: first, usual training using a stochastic gradient optimizer (Adam), then a hardening phase pushing weights in Equation 6 towards 0 or 1. Finally, weights were clamped to exact symbolic values. Systematic experiments (over 1000 runs with varied seeds and initialization strategies) show that blind recovery from random initialization succeeds in 100% of runs at depth 2, approximately 25% at depths 3–4, and below 1% at depth 5. At depth 6, no blind recovery was observed in 448 attempts. When successful, the snapped weights yield mean squared errors at the level of machine epsilon squared ($\sim 10^{-32}$), consistent with exact symbolic recovery. Noteworthy, when the weights of the correct EML tree are perturbed by Gaussian noise, the optimization converges back to the exact values in 100% of runs, even for trees of depth 5 and 6. This demonstrates that the EML tree representation is valid and that the correct basins of attraction exist. Finding them from random initialization becomes harder with depth. If one manages to improve convergence of EML trees beyond proof of concept, possibly using another binary operator similar to 4 but with better properties (non-exponential asymptotics, no domain issues), then we could achieve symbolic regression of data using gradient-based methods [46]. See Supplementary Information for systematic training experiments, code and details.
Section Summary: The EML operator serves as a fundamental building block for creating and calculating real elementary functions like exponentials and logarithms, allowing many computations to be expressed in a unified form, though similar alternative operators also exist. Open questions include whether even more versatile versions can be found that don't require specific constants like infinity or e, if a binary operator without such pairings is possible, and whether a single-variable version could double as a neural network activation function to generate all elementary functions. While EML relies on complex numbers internally to handle real math—much like quantum computing uses them for probabilities—this approach enables clearer, interpretable results in neural-like structures compared to traditional opaque networks.
The operator EML, Equation 3, provides a single sufficient primitive from which real elementary functions can be constructed and evaluated. Consequently, a wide class of computations built from such functions can also be cast in EML form. It is not unique; several close variants are likely to exhibit similar properties, including EDL, Equation 4b, and the swapped-argument form $- \operatorname{eml}(y, x)$, Equation 4c. More operators of this kind exist. Perhaps an entire continuous family of them awaits discovery, with properties more convenient than Equation 3. For example, the requirement for one of the constants: $-\infty, 1, e, \ldots$ to be always present among terminal symbols makes its use less elegant and more complicated (cf. Subsect. Section 4.3) in comparison with, e.g., standard neural nets or the NAND gate. The latter is able to generate[^3] 0s and 1s out of 'anything'. The EML operator does not have this useful property. Whether an EML-type binary Sheffer working without pairing with a distinguished constant exists is an open question. Proving such impossibility for any given candidate is non-trivial: one might expect $f(x, x)$ being constant to suffice, but consider $B(x, y) ! = ! x !-! \tfrac{y}{2}$, for which $B(x, x) ! = ! \tfrac{x}{2}$ yet $B(B(x, x), x) ! = ! 0$. Such traps illustrate why systematic search is essential in this work. A ternary operator, $T(x, y, z)! = ! {e^x}/{\ln{x}} \times {\ln{z}}/{e^y}$, for which $T(x, x, x)! =! 1$ is next candidate for further analysis [47].
[^3]: E.g., $1 \equiv \text{NAND}(x, \text{NAND}(x, x))$. In practice, one will rarely use this property. Usually, we use directly available 0s and 1s, e.g. GND and $V_{cc}$ in digital circuits.
Whether a univariate Sheffer exists, serving simultaneously as a neural activation function [48] and as a generator of all elementary functions, remains open (see SI, Sect. 5).
One might complain that the EML representation of elementary functions requires complex arithmetic for real math, at least internally. Just as quantum computing uses complex amplitudes to compute real probabilities, EML uses complex intermediates to compute real elementary functions. This seems inevitable. We must somehow compute the imaginary unit $i$, $\pi$, and all trigonometric/hyperbolic functions via Euler's formula, Equation 2. For that, we use $\ln{x}$ for $x<0$. A continuous Sheffer working purely in the real domain seems impossible. My search for alternatives, e.g., using pairs of trigonometric/hyperbolic functions and their inverses instead of $\exp/\ln$, found nothing. Quite surprisingly, the requirement to use complex numbers internally causes only minor problems in practice of using Equation 3 in Computer Algebra Systems or numerical simulations.
Since standard activation functions are themselves elementary, any conventional neural network is a special case of an EML tree architecture. Current networks can learn symbolic algebra [49] and digit-level arithmetic [50], but their internal mechanisms remain opaque [11], and efficient exact evaluation of elementary functions as continuous real-valued operations is still beyond their reach. EML representations go further: as demonstrated in Subsect. Section 4.3, trained weights can snap to exact binary values, recovering closed-form elementary subexpressions alongside approximations. When this succeeds, the discovered circuits are legible as elementary function expressions — a form of interpretability unavailable to conventional architectures.
The core idea, the discovery of the EML Sheffer operator, the verification methodology, and results are entirely the author's own work. Large language models (including recent Claude, Grok, Gemini and ChatGPT) were used mainly for language editing and coding assistance.
Code, scripts, and supplementary reproducibility materials used to generate the figures, tables, and numerical results are available in the SymbolicRegressionPackage repository [39]. An archival snapshot of the exact submission version is deposited in Zenodo DOI: 10.5281/zenodo.19183008 (https://doi.org/10.5281/zenodo.19183008). The archival package contains the source code, scripts, and README files needed to rerun the reported results.
Computational resources were partially provided by Google Cloud Research Credits.
Extensive three-part Supporting Information is supplied as a separate SI Appendix PDF.
Section Summary: This references section lists a wide array of sources spanning centuries, from John Napier's 1614 book on logarithms to modern 2025 works on software like Mathematica, covering topics like mathematical logic, numerical computing, neural networks, and even DNA structure. It includes foundational texts on Boolean algebras and combinatory logic, alongside more specialized papers on AI-driven symbolic regression, cellular automata, and numerical integration methods. The collection draws from fields such as physics, biology, and computer science, providing historical and contemporary perspectives on computational and mathematical ideas.
[1] Press et al. (2007). Numerical Recipes: The Art of Scientific Computing. Cambridge University Press.
[2] Sheffer, Henry Maurice (1913). A Set of Five Independent Postulates for Boolean Algebras, with Application to Logical Constants. Transactions of the American Mathematical Society. 14(4). pp. 481–488. doi:10.1090/S0002-9947-1913-1500960-1.
[3] Napier, John (1614). Mirifici Logarithmorum Canonis Descriptio. Andrew Hart.
[4] Briggs, Henry (1624). Arithmetica Logarithmica. London.
[5] Cotes, Roger (1722). Harmonia Mensurarum, sive Analysis & Synthesis per Rationum & Angulorum Mensuras Promotae: Accedunt Alia Opuscula Mathematica. Cambridge.
[6] Liouville, Joseph (1835). Mémoire sur l'intégration d'une classe de fonctions transcendantes. Journal für die reine und angewandte Mathematik (Crelle's Journal). 13. pp. 93–118. doi:10.1515/crll.1835.13.93.
[7] Ritt, Joseph Fels (1948). Integration in Finite Terms: Liouville's Theory of Elementary Methods. Columbia University Press.
[8] Bernd Ulmann (2022). Analog Computing. De Gruyter Oldenbourg. doi:doi:10.1515/9783110787740. https://doi.org/10.1515/9783110787740.
[9] Udrescu, Silviu-Marian and Tegmark, Max (2020). AI Feynman: A physics-inspired method for symbolic regression. Science Advances. 6(16). pp. eaay2631. doi:10.1126/sciadv.aay2631.
[10] Cranmer et al. (2020). Discovering Symbolic Models from Deep Learning with Inductive Biases. Advances in Neural Information Processing Systems. 33. pp. 17429–17442.
[11] Liu et al. (2025). KAN: Kolmogorov–Arnold Networks. In International Conference on Learning Representations (ICLR). https://openreview.net/forum?id=Ozo7qJ5vZi.
[12] Jan Łukasiewicz (1963). Elements of Mathematical Logic. Pergamon Press.
[13] Black, H. S. (1934). Stabilized feedback amplifiers. The Bell System Technical Journal. 13(1). pp. 1-18. doi:10.1002/j.1538-7305.1934.tb00652.x.
[14] Vinod Nair and Geoffrey E. Hinton (2010). Rectified Linear Units Improve Restricted Boltzmann Machines. In Proceedings of the 27th International Conference on Machine Learning (ICML 2010). pp. 807–814.
[15] Ian Goodfellow et al. (2016). Deep Learning. MIT Press. https://www.deeplearningbook.org/.
[16] Wolfram, Stephen (2002). A New Kind of Science. Wolfram Media. https://www.wolframscience.com/nks/.
[17] Schönfinkel, M. (1924). Über die Bausteine der mathematischen Logik. Mathematische Annalen. 92(3). pp. 305-316. doi:10.1007/BF01448013. https://doi.org/10.1007/BF01448013.
[18] Curry, H.B. and Feys, R. (1958). Combinatory Logic. North-Holland Publishing Company. https://books.google.pl/books?id=fEnuAAAAMAAJ.
[19] Lafont, Yves (1997). Interaction Combinators. Inf. Comput.. 137(1). pp. 69–101. doi:10.1006/inco.1997.2643. https://doi.org/10.1006/inco.1997.2643.
[20] Michał Baczyński et al. (2022). On the Sheffer stroke operation in fuzzy logic. Fuzzy Sets and Systems. 431. pp. 110-128. doi:https://doi.org/10.1016/j.fss.2021.05.003. https://www.sciencedirect.com/science/article/pii/S0165011421001834.
[21] Oleg Mazonka and Alex Kolodin (2011). A Simple Multi-Processor Computer Based on Subleq. https://arxiv.org/abs/1106.2593. arXiv:1106.2593.
[22] Conway, J. H. (1987). FRACTRAN: A Simple Universal Programming Language for Arithmetic. doi:10.1007/978-1-4612-4808-8_2. https://doi.org/10.1007/978-1-4612-4808-8_2.
[23] Cook, Matthew (2004). Universality in Elementary Cellular Automata. Complex Systems. 15(1). pp. 1–40. doi:10.25088/ComplexSystems.15.1.1.
[24] Linsky, B. (2011). The Evolution of Principia Mathematica: Bertrand Russell's Manuscripts and Notes for the Second Edition. Cambridge University Press. https://books.google.pl/books?id=EHCMuK4gS-MC.
[25] Crick, Francis H. C. and Watson, James D. (1953). Genetical Implications of the Structure of Deoxyribonucleic Acid. Nature. 171(4361). pp. 964–967. doi:10.1038/171964b0.
[26] Smith et al. (2024). An aperiodic monotile. Combinatorial Theory. 4(1). doi:10.5070/C64163843. https://escholarship.org/uc/item/3317z9z9.
[27] Andrzej Odrzywolek (2014). How to Build the Perfect Igloo. Eureka. 63. pp. 1001110–1001111.
[28] Fehlberg, E. (1970). Klassische Runge-Kutta-Formeln vierter und niedrigerer Ordnung mit Schrittweiten-Kontrolle und ihre Anwendung auf Wärmeleitungsprobleme. Computing. 6(1). pp. 61-71. doi:10.1007/BF02241732. https://doi.org/10.1007/BF02241732.
[29] Hidetosi Takahasi and Masatake Mori (1974). Double Exponential Formulas for Numerical Integration. Publications of the Research Institute for Mathematical Sciences. 9(3). pp. 721–741. doi:10.2977/prims/1195192451.
[30] Cooley, James W. and Tukey, John W. (1965). An Algorithm for the Machine Calculation of Complex Fourier Series. Mathematics of Computation. 19(90). pp. 297–301. doi:10.1090/S0025-5718-1965-0178586-1.
[31] John Shackell (1990). Growth estimates for exp—log functions. Journal of Symbolic Computation. 10(6). pp. 611-632. doi:https://doi.org/10.1016/S0747-7171(08)80161-7. https://www.sciencedirect.com/science/article/pii/S0747717108801617.
[32] Richardson et al. (1996). Asymptotic Expansions of exp-log Functions. https://inria.hal.science/inria-00073832.
[33] Wolfram Research, Inc. (2025). Mathematica, Version 14.3. Champaign, IL, 2025. https://www.wolfram.com/mathematica.
[34] Dan Finkel (2017). Broken Calculator Warm Up. https://mathforlove.com/lesson/broken-calculator-warmup/. Accessed 2025-11-18.
[35] Newman, Donald J. (1982). A Problem Seminar. Springer-Verlag.
[36] Naskrecki, Bartosz and Ono, Ken (2025). Mathematical discovery in the age of artificial intelligence. Nature Physics. 21(10). pp. 1504–1506. doi:10.1038/s41567-025-03042-0. https://doi.org/10.1038/s41567-025-03042-0.
[37] Kolmogorov, A. N. (1965). Three Approaches to the Quantitative Definition of Information. Problems of Information Transmission. 1(1). pp. 1–7.
[38] Giuseppina Terzo (2008). Some Consequences of Schanuel's Conjecture in Exponential Rings. Communications in Algebra. 36(3). pp. 1171–1189. doi:10.1080/00927870701410694. https://doi.org/10.1080/00927870701410694.
[39] Odrzywołek, Andrzej (2026). SymbolicRegressionPackage: Basic building blocks for symbolic regression.. https://github.com/VA00/SymbolicRegressionPackage. Includes EML toolkit.
[40] Charles R. Harris et al. (2020). Array programming with NumPy. Nature. 585(7825). pp. 357–362. doi:10.1038/s41586-020-2649-2.
[41] Adam Paszke et al. (2019). PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Advances in Neural Information Processing Systems 32. pp. 8024–8035.
[42] de Moura, Leonardo and Ullrich, Sebastian (2021). The Lean 4 Theorem Prover and Programming Language. In Automated Deduction – CADE 28: 28th International Conference on Automated Deduction, Virtual Event, July 12–15, 2021, Proceedings. pp. 625–635. doi:10.1007/978-3-030-79876-5_37. https://doi.org/10.1007/978-3-030-79876-5_37.
[43] Bernd Ulmann (2024). Beyond zeros and ones – analog computing in the twenty-first century. International Journal of Parallel, Emergent and Distributed Systems. 39(2). pp. 139–151. doi:10.1080/17445760.2023.2296672. https://doi.org/10.1080/17445760.2023.2296672.
[44] Cranmer, Miles (2023). Interpretable Machine Learning for Science with PySR and SymbolicRegression.jl. arXiv:2305.01582.
[45] Jang et al. (2017). Categorical Reparameterization with Gumbel-Softmax. In Proceedings of the 5th International Conference on Learning Representations (ICLR). https://openreview.net/forum?id=rkE3y85ee.
[46] Petersen et al. (2021). Deep Symbolic Regression: Recovering Mathematical Expressions from Data via Risk-Seeking Policy Gradients. In International Conference on Learning Representations (ICLR). https://openreview.net/forum?id=m5Qsh0kBQG.
[47] Odrzywolek, Andrzej (2026). A ternary Sheffer operator for elementary functions?. Acta Physica Polonica B.
[48] Dubey et al. (2022). Activation Functions in Deep Learning: A Comprehensive Survey and Benchmark. Neurocomputing. 503. pp. 92–108. doi:10.1016/j.neucom.2022.06.111.
[49] Lample, Guillaume and Charton, François (2020). Deep Learning for Symbolic Mathematics. In Proceedings of the 8th International Conference on Learning Representations (ICLR). https://openreview.net/forum?id=S1eZYeHFDS.
[50] Lee et al. (2024). Teaching Arithmetic to Small Transformers. In The Twelfth International Conference on Learning Representations (ICLR). https://openreview.net/forum?id=dsUB4bst9S.