.

This work was partially supported by the ANR Gecko and ANR-09-JCJC-0098-01 MaGiX projects, and by the Digiteo 2009-36HD grant and Région Ile-de-France.

 Guessing singular dependencies

 Joris van der Hoeven

 CNRS, Département de Mathématiques Bâtiment 425 Université Paris-Sud 91405 Orsay Cedex France

 Email: vdhoeven@lix.polytechnique.fr

 Web: http://lix.polytechnique.fr/~vdhoeven

 January 5, 2015

 Given complex numbers , it is classical that linear dependencies with can be guessed using the LLL-algorithm. Similarly, given formal power series , algorithms for computing Padé-Hermite forms provide a way to guess relations with . Assuming that have a radius of convergence and given a real number , we will describe a new algorithm for guessing linear dependencies of the form , where have a radius of convergence . We will also present two alternative algorithms for the special cases of algebraic and Fuchsian dependencies. Keywords: guessing, asymptotic dependency, orthogonalization, analytic continuation, Fuchsian singularity, Padé-Hermite forms, algorithm A.M.S. subject classification: 30E10, 30B40, 41A21, 65B05

## 1.Introduction

### 1.1.General background and results

Consider an infinite sequence of complex numbers. If are the coefficients of a formal power series , then it is well-known [Pól37, Wil94, FS96] that the asymptotic behaviour of the sequence is closely related to the behaviour of the generating function near its dominant singularity. Now, if is the solution to some complicated equation, then it can be hard to compute the asymptotic behaviour using formal methods. On the other hand, the coefficients of such a solution can often be computed numerically up to a high order. With this numerical evidence at hand, it is natural to raise the following questions:

1. Can we guess the asymptotic behaviour of ?

2. Can we guess the behaviour of near its dominant singularity?

These questions can be regarded as part of the construction of a more general toolbox for the “experimental mathematician” [BBKW06, BD09]. More specifically, we advocate the systematic integration of “guessing tools” into symbolic computation packages. Indeed, current systems can be quite good at all kinds of formal manipulations. However, in the daily practice of scientific discovery, it would be helpful if these systems could also detect hidden properties, which may not be directly apparent or expected. Furthermore, the guessing tool is allowed to be heuristic, so that it only suggests hidden properties; at a second stage, one may then search for full proofs using other techniques.

One well-known tool in this direction is the LLL-algorithm [LLL82]. Given numbers , it can be used in order to guess relations of the form

 (1)

Given formal power series , algorithms for the computation of Padé-Hermite forms [BL94, Der94] can be used in order to guess linear relations

 (2)

A well-known implementation is provided by the Gfun package [SZ94]. Given a finite number of coefficients of a formal power series , the Gfun package is able to guess a closed form formula for or a linear differential equation with coefficients in satisfied by . Indeed, it suffices to take in (2) in order search for small linear differential equations satisfied by .

Unfortunately, many interesting formal power series do not admit closed form formulas and do not satisfy linear differential equations with polynomial coefficients. In that case, we can still use asymptotic extrapolation [Hoe09] in order to guess the asymptotic behaviour of the coefficients. However, this only provides us some rough idea about the behaviour of at its dominant singularity. In practice, it often happens that locally satisfies an algebraic or differential equation with analytic coefficients, even though these coefficients fail to be polynomials. For instance, combinatorics [Pól37, FS96] is full with examples of generating functions which are not algebraic, but whose dominant singularities are algebraic.

In this paper, we will describe two approaches to detect analytic dependencies on a compact disk: the first one assumes that we have an algorithm for the analytic continuation of and relies on the monodromy of at its singularities. The second approach is purely numerical and makes no special assumptions on . On our way, we will encounter various interesting related problems, such as the determination of the radius of convergence or the singularities of . We will also propose heuristic solutions to these problems, thereby extending the basic toolbox for experimenting with analytic functions.

### 1.2.Overview of the paper

Since all algorithms in this paper are directly or indirectly based on heuristics, it is important to investigate how much confidence we can attach to the computed results. In section 2, we will present a survey of different kinds of heuristic algorithms which are used in symbolic computation. Some of these heuristics are relatively benign when compared to others, so it will be useful to have some general insights on the reliability of different types of heuristic algorithms.

The remainder of the paper is divided into two main parts. In the first part (sections 3, 4 and 5), we will develop some basic tools for later use. In the second part, we turn to the main topic of this paper: the disclosure of local analytic relations between analytic functions. Two main types of analytic input functions will be considered:

AN1
Functions for which we can merely compute a large number of the Taylor coefficients at the origin.

AN2
Functions for which we have a reasonably efficient algorithm for their analytic continuation, so that we can also compute the Taylor coefficients of at other points besides the origin.

The second situation typically arises when is the solution of a differential or more general functional equation; see section 4.1 for a further discussion on analytic continuation.

One of the most basic problems concerning an analytic function which is only known by its Taylor coefficients, is to determine its radius of convergence. In section 3, we will present two methods. The first one is related Cauchy-Hadamard's formula and provides rough approximations under mild assumptions on . The second method provides much better approximations, but only works if admits a single dominant singularity of a simple type.

Building on the algorithm for approximating the radius of convergence of , we next turn our attention to the problem of locating its singularities. In section 4.2, we first restrict ourselves to the dominant singularities of (i.e. the singularities of minimal norm). Assuming AN2, we next present algorithms for the exploration of the Riemann surface of beyond its dominant singularities.

In the special case when is meromorphic on a compact disk , section 5 contains a special purpose algorithm for the determination of a polynomial such that is analytic on . This algorithm works under the hypothesis AN1 and induces better algorithms for the problems in sections 3 and 4 in this special case.

In the second part of this paper, we turn to the detection of dependencies between analytic functions on a compact disk . More precisely, assume that are fixed convergent power series. We are interested in the determination of linear dependencies of the form

 (3)

where are analytic on . Modulo a scaling , we may assume without loss of generality that .

In section 6, we assume AN2 and first consider the two special cases of algebraic and Fuchsian dependencies. In the case of algebraic dependencies, we take for a fixed function , and also require that . In the case of Fuchsian dependencies, we take for a fixed function , and again require that . The second algorithm only succeeds in the case when all singularities of on the disk are of Fuchsian type (see sections 3.2 and 6.2 for detailed definitions). The main idea behind the method is to compute the set of of functions generated by and its analytic continuation around its singularities in . If there exists an algebraic dependency, then this set is finite and we may use it to find the dependency. If there exists a Fuchsian dependency, then the set is contained in finite dimensional vector space, and we may construct the dependency from a basis of this vector space.

In section 7, we only assume AN1 and consider the general problem of determining dependencies of the form (3). We will describe a purely numerical algorithm based on Gram-Schmidt orthogonalization for finding such relations. The idea is to simply truncate the power series expansions and then apply the numerical method. We will prove that an asymptotic convergence theorem which shows that this approach is at least correct “at the limit”.

In the last section 8, we will present some numerical experiments with the algorithm from section 7. In section 8.1, we first consider some examples of relations which were recognized by the algorithm. In section 8.2, we will also examine the behaviour of the algorithm in the case when are analytically independent. In fact, the algorithm from section 7 still requires some human cooperation for deciding whether we really found a relation. Based on the numerical experiments, we conclude with some perspectives for the design of a fully automatic algorithm.

Acknowledgments. We would like to thank the three referees for their detailed and valuable comments on a first version of this paper.

## 2.On the use of heuristic algorithms

As soon as we enter the area of heuristic and non mathematically proven algorithms, it is natural to ask how much confidence can be attached to the output. In fact, there is a large spectrum of situations which can occur, depending on the precise nature of the heuristics. There may also be a trade-off between computational complexity and correctness: is it better to rely on a slow deterministic algorithm for polynomial factorisation over a finite field, or on a fast probabilistic algorithm?

Without striving for exhaustion, section 2.1 a catalogue of different kinds of heuristic algorithms which occur in the area of symbolic computation. Each of the heuristic algorithms considered in this paper will fit into one of these categories. We will also describe some standard strategies in order to increase the level of confidence and examine typical examples for which heuristic algorithms may fail.

### 2.1.Various kinds of heuristic algorithms

##### Correct modulo conjectures
A high confidence can typically be attached to algorithms which are correct modulo conjectures. One important example of this situation in computer algebra is zero testing of transcendental functions or constants.

For example, consider the class of exp-log constants, built up from the rationals using the operations , exp and log. The following conjecture [Lan71] is a major open problem in number theory:

Conjecture 1. (Schanuel) Let be complex numbers which are linearly independent over the rational numbers . Then the transcendence degree of

is at least .

From the computer algebra point of view, the conjecture implies [CP78] that all numerical relations between exp-log constants can be deduced from the usual rules and . Based on this fact, Richardson has given a zero test for exp-log constants [Ric97, Ric07] with the following properties: (1) if the algorithm terminates on a given input, then the result is correct; (2) if the conjecture holds, then the algorithm always terminates; (3) if the algorithm does not terminate on a given input, then a counterexample to the conjecture can be constructed from this input.

Similar situations occur in algorithmic number theory, depending on the correctness of Riemann's hypothesis [Bel11]. In this paper, the algorithms in sections 4.3 and 6 rely on zero tests for analytic functions. If we know that our analytic functions lie in special classes (such as the class of exp-log functions, or the class of differentially algebraic functions), then these zero tests might be reduced to suitable conjectures [Hoe01].

##### Reduction to other heuristic algorithms
Another frequent situation is that an algorithm relies on one of more other heuristic subalgorithms, with the property that, for a given input, we obtain the right output whenever all calls of the subalgorithm(s) return the correct results. From a theoretical point of view, this corresponds to correctness modulo one or more oracles.

For instance, in section 3, we will give heuristic algorithms for the computation of the radius of convergence of an analytic function. This will be a basic building block for many of the other algorithms in this paper. Nevertheless, the correctness of several of these other algorithms reduces to the correctness of this basic building block.

If, for input functions in a more restricted class, we may determine their convergence radii with larger accuracy, then this will immediately make all the other algorithms more reliable. For instance, if we consider a class (e.g. rational or algebraic functions) for which convergence radii can be computed exactly, then the algorithms in section 4.2 are no longer heuristic.

##### Correctness with probability one
In computer algebra, there are plenty of algorithms which are correct unless some degeneracy occurs. For instance, solving a system of polynomial equations using numerical homotopy methods [Ver96, SW05] is correct unless the chosen paths accidentally hit a singularity. Moreover, if paths are chosen at random, then such accidents occur with probability zero. Similarly, in order to test whether a real analytic function is zero, it suffices to pick a random number and test whether vanishes.

##### Correctness with high probability
In practice, there is no such thing as a random real number, since we can only represent numbers with finite precision. The above examples of heuristic algorithms which are correct “with probability ” are really correct with probability , where is the working precision. There are many other examples of heuristic algorithms which are correct with high probability. For instance, in order to show that two polynomials in have no common divisors, it suffices to pick a “sufficiently random” prime and show that the reductions of and in have no common divisors.

We also recall that there are two classical types of probabilistic algorithms. The examples mentioned above are only correct with high probability and fall into the class of so-called Monte Carlo algorithms. The second kind of Las Vegas algorithms are always correct, but only fast with high probability. For instance, all known fast algorithms for factoring polynomials over a finite field are probabilistic Las Vegas type algorithms [GG02].

##### Asymmetric correctness
We also would like to underline another aspect of heuristic tests, such as zero tests: it frequently occurs that only one direction of the test is difficult and based on heuristics, whereas the other direction is “easy”. For instance, assume that we want to test whether an exp-log constant is zero. If the constant is non zero, then it is easy to prove this by evaluating the constant using interval arithmetic [Moo66] using a sufficiently high precision and verify that zero is not in the interval enclosure of our constant. The difficult case, for which Richardson's algorithm [Ric07] relies on Schanuel's conjecture, is when the exp-log constant is zero and we want to prove this.

##### Numerical algorithms
Essentially all numerical algorithms are heuristic in the sense that they only compute approximate answers. Numerical analysts therefore have a lot of experience in dealing with heuristics and intentionally vague arguments. The very aim of numerical algorithms is often not to compute a mathematically sound result, but rather to design a method which “works” for solving a practical problem.

This does not withstand that experts in this area have a great sense of how accurate and trustworthy various methods are. Ideally speaking, any numerical algorithm comes with a detailed error analysis. Most textbooks on numerical analysis start with a section on the various sources of error and how to take into account the machine accuracy [PTVF07, section 1.1]. For more statistical algorithms, there is also a tradition of trying to quantify the level of trust that can be attached to the computed results. Finally, for many algorithms, it is possible to automate the error analysis by using interval arithmetic [Moo66].

##### Ultimate correctness
Classical numerical algorithms operate at a fixed precision . From a theoretical point of view [Wei00, Hoe11], it can be investigated whether a given numerical algorithm is at least “ultimately correct” in the sense that the computed output tends to the mathematically correct result if tends to infinity. If the numerical algorithm comes with a rigourous error analysis, then we have a means for controlling the distance between our approximations at a given precision and the ultimate limit. In particular, for any , we may then select a precision with .

However, there are many situation in which we have no means for controlling the distance , even though our algorithm is ultimately correct. For instance, the computation of the radius of convergence of a differentially algebraic function is undecidable [DL89], even though there exists an algorithm [Hoe05, Hoe07] for the computation of a sequence with . Similarly, for a large class of functions, the formula (4) below gives an approximation for the radius of convergence of a series (when computing with a precision of bits), but the formula is only ultimately correct (5).

### 2.2.Further considerations

##### Increasing the trustworthiness
Since the outcome of a heuristic algorithm cannot entirely be trusted, it is natural to search for ways to at least increase the trustworthiness of the result. Depending on the kind of heuristics that were used, this can either be done by rerunning the same algorithm for different parameters, or by combining the algorithm with other algorithms.

For instance, in numerical analysis, a frequent trick is to rerun the same algorithm for different working precisions and compare the results. Similarly, probabilistic algorithms of Monte Carlo type can be rerun several times in order to increase our confidence that we obtained the correct result.

We also mentioned the fact that equality testing of exp-log constants is quite asymmetric in the sense that proving equalities is usually much harder than proving inequalities. This suggests the use of a mixed strategy for this problem, which combines a heuristic equality test with a deterministic, interval arithmetic based, inequality prover. In case of equality, we may optionally run Richardson's more expensive algorithm and try to prove the equality.

In applications, we notice that heuristic algorithms are often used in a similar way: to speed up the main algorithm, which can be perfectly deterministic. A nice example is the computation of a Gröbner basis over the rationals using modular arithmetic. In Faugère's software, modular arithmetic is used in order to get a precise idea about those S-polynomials that reduce to zero, thereby avoiding many unnecessary reductions when performing the actual computations over the rationals [Fau94].

In the particular case of a heuristic algorithm whose purpose is to guess dependencies, we finally notice that we may always build in the trivial safeguard to check all proposed relations a posteriori; see also remark 10 in section 6.1.

##### Attacks
It is a good practice to attack heuristic algorithms and try to construct systematic examples on which the heuristics fail. Successful attacks usually indicate that the algorithm was either wrong, or that the heuristics only make sense under certain additional assumptions. Failed attacks generally increase the confidence in the algorithm and sometimes lead into theoretical insight on why precisely the heuristics work well.

One particular angle of attack concerns the computational complexity or the numerical stability. This means that we search for examples for which the heuristic algorithm is very slow or numerically unstable. This line of thought underlies much of the numerical experiments which are presented in section 8.

Instead of attacking the heuristic algorithms, it is also possible to look for interesting classes of examples on which the heuristic provably works (or works better). In this paper, some interesting special classes of analytic functions are meromorphic, algebraic and so-called Fuchsian functions on a compact disk (see section 6.2). Typical classes which have to be considered for counterexamples are lacunary series, functions with a natural boundary, or holonomic functions that are not Fuchsian.

## 3.Computing the radius of convergence

Let be an analytic function which is given by its power series at the origin. A natural problem is to compute the radius of convergence of . For sufficiently large classes of analytic functions, such as solutions to algebraic differential equations over the rationals, this problem is generally undecidable [DL89, Hoe07], although efficient and high quality algorithms for the computation of lower bounds for do exist in this case [Hoe07]. In this section, we will consider heuristic algorithms, and only assume that a large number of coefficients of are known numerically.

### 3.1.Rough approximation using the Newton polygon

The first idea which comes into our mind is to apply Cauchy-Hadamard's formula for the radius of convergence :

Given coefficients of , we may for instance use the approximation

 (4)

For most convergent power series , this formula is ultimately exact in the sense that

 (5)

This is in particular the case when is ultimately convex or ultimately concave (this follows from the fact that the sequence converges in an ultimately monotone way to , in this case). The set of for which (5) holds is also stable under the transformation for any . Of course, we may replace by in (5) for any . Notice however that the lacunary power series does not satisfy (5).

The formula (4) has the disadvantage that it has not been scaled appropriately: when replacing by , where is such that , we obtain different approximations for and . This drawback can be removed by replacing by for some appropriate coefficient index with . In fact, one can even do better, and compute using the formula for appropriate indices with and . Let us now show how to read off such indices from the numerical Newton diagram of .

Let , where we understand that . Then the Newton diagram of is the convex hull of the half lines

For a fixed , say , let , and consider the Newton diagram of . There exists a minimal subset with , such that the Newton diagram is also the convex hull of the half lines for . Graphically speaking, the are the vertices of the Newton diagram (see figure 1).

For a fixed , say , we may now determine the unique edge of the Newton diagram such that , and replace the formula (4) by

 (6)

For a large class of convergent power series , the formula (6) is again ultimately exact. The indices can be computed from the coefficients using a linear traversal in time . This yields an efficient algorithm for the approximation of which turns out to be more accurate than (4) in practice. The formula (6) has been implemented in the Mathemagix system [HLM+02]. For , the computed radius is usually correct up to one decimal digit in the relative error.

 Figure 1. Illustration of the numerical Newton diagram for and truncated at order . Taking , we get , and . The red line is the unique line which extends the edge , of slope .

### 3.2.Asymptotic extrapolation

The formula (6) usually yields a reasonable estimate for , even in very degenerate cases when there are several singularities at distance or close to . However, if admits a single isolated singularity at distance of the origin, with no other singularities at distance close to , then it is often possible to read off much more information about this singularity from the coefficients .

For instance, it frequently (always?) occurs that the quotients simply tend to for . Moreover, as we will show below, if the singularity at has a known type, then the approximation can be further improved. If nothing is known about the singularity at , then we may still try and see what happens if we apply such specialized algorithms for various frequent types of singularities: with a bit of luck, we will both obtain useful information about the type of singularity and the numerical value of .

We say that is algebraic at , if satisfies a polynomial equation

where are analytic functions at with . In that case, we have

with and ramification index , whence

 (7)

Using the E-algorithm [Wen01, BZ91], we may now compute simultaneous approximations for the first coefficients , , , , etc. of the expansion (7). It turns out that this strategy greatly improves the accuracy of the approximation of (see also [Hoe09]).

Similarly, we say that is Fuchsian at , if satisfies a linear differential equation

where and are analytic functions at with . In that case, the Taylor coefficients satisfy the asymptotic expansion

where , and the are polynomials in of degrees [Fab85, Poi86, Bir13, Was67]. Again, the E-algorithm or more general algorithms for asymptotic extrapolation [Hoe09] can be used to compute with a high accuracy. Notice that these algorithms also provide estimates for the accuracies of the computed approximations.

In cases where nothing particular is known about the behaviour of at its dominant singularity, the strategy of asymptotic extrapolation [Hoe09] still may work and both provide useful information about the nature of the singularity and the numerical value of .

## 4.Locating singularities of analytic functions

### 4.1.Analytic continuation

In order to explore the analytic function determined by the coefficients more closely, and in particular determine its singularities in a given region, it is useful to have a means for performing the analytic continuation of .

The way is given sometimes provides us with such a means. For instance, if is the solution to an initial value problem

then the continuation of at any sufficiently small is again the solution of an initial value problem

In [Hoe05, Hoe07] it is shown in detail how to turn this into an algorithm for the analytic continuation of . Notice also that the power series solutions to an initial value problem can be computed efficiently using the algorithms from [BK78, Hoe02].

In general, it is always possible to compute the continuation numerically. However, this kind of “poor man's” analytic continuation induces a big loss of accuracy. Let us illustrate this on the fundamental example

In order to compute with a precision of bits using the formula

we need to truncate this expansion at an order for which

Putting and using Stirling's formula, this yields

For instance, in order to expand at order with precision , we get

and we need an expansion of at the origin at order . In general, for every analytic continuation step, we need to multiply the expansion order by a constant factor which depends on the desired precision and the ratio between the step size and the radius of convergence.

An experimentally better strategy for poor man's analytic continuation is to postcompose with a suitable analytic function instead of . If admits a dominant singularity at , then should be chosen in such a way that has as its only singularity in the unit disk, and such that is as small as possible. One may typically take to be a composition of rational functions of the form . Yet another strategy would be to consider a Padé-Hermite approximation of and then to expand (instead of ) at .

### 4.2.Locating dominant singularities

Assuming that we have a means for analytic continuation (whether this means is efficient and accurate or only a “poor man's” solution), a natural question is to gather information about the singularities of . First of all, we may want to locate the dominant singularity (or singularities) of with a good precision.

Let us first assume that admits a unique dominant singularity at and no other singularities of norm . Section 3 provides us with heuristic algorithms for the computation of . More generally, for any with , the point is also the dominant singularity of , so we may compute using the same heuristic algorithms. Computing and for two such points, the singularity is now uniquely determined by its distances , and to , and (see figure 2).

Even if we do not know beforehand, then we apply the above method for with , for some . In order to check the correctness of the computed value of a posteriori, we also compute the radius of convergence of . Then is a dominant singularity of if and only if . If this check fails, then we simply divide by two and repeat the same procedure until the right value of is found.

In practice, the convergence radii can only be approximated, so all equality tests have to be replaced by approximate equality tests, for which the tolerance in relative precision is slightly larger than the estimated relative precision with which the convergence radii are computed. In particular, the above method computes with approximately the same accuracy as the radius of convergence .

Furthermore, if our procedure for analytic continuation is accurate enough, then we may “zoom in” on a singularity, and determine its location with higher precision. More precisely, assume that we have a rough approximation of the dominant singularity . For some small , we then compute an approximation of the dominant singularity of with . This yields the improved approximation of .

Assume now that admits more than one dominant singularity (or other singularities of norms close to ), but still a finite number of them. On the one hand, by computing the radii of convergence of for a sufficiently large number of points with , we may determine all dominant singularities. On the other hand, for any two adjacent dominant singularities and with , we may check the absence of other singularities in between by looking at the radii of convergence of and for some . This yields a criterion for deciding whether we took a “sufficiently large number of points ”.

From the numerical point of view, rough initial rough approximations can be further improved by zooming in. In figure 3 we have shown an example where we zoomed in on one among eight singularities. After zooming in, we see that all other singularities are far away, so they do not pollute the computation of the radius of convergence.

The generalized algorithm is particularly useful in the special case when there are two dominant singularities which are complex conjugates. This annoying situation prevents us from using the technique of asymptotic extrapolation in a direct way. Nevertheless, we may use it after zooming in on one of the two singularities.

 Figure 2. Determination of from its distances to , and .
 Figure 3. Zooming in on one among eight dominant singularities.

### 4.3.Exploring the Riemann surface

In the case when our method for analytic continuation allows us to turn around singularities, it is natural to ask whether we can explore the Riemann surface of beyond the dominant singularities. In general, it is always possible to approximate the Riemann surface of by an organically growing sequence of more and more precise Riemann surfaces of a special form; we refer to [Hoe07] for details. In this section, we will restrict ourselves to a more particular situation: given the closed disk of center and radius , we will assume that there exist a finite number of points in its interior such that is defined above . Our task is to find the points .

One complication is that some of the singularities may not directly be visible from the origin by following a straight line, but only after turning around one or more other singularities. In principle, more and more singularities might therefore be disclosed by making successive turns around the singularities which were already found. In order to ensure our algorithm to terminate, we will make the additional assumption

H

There exists a finite dimensional vector space of analytic functions above , which contains and all its analytic continuations.

This is in particular the case if

A

The Riemann surface of admits a finite number of sheets above .

In sections 6.1 and 6.2, we will encounter two natural situations in which the assumptions A resp. H are satisfied.

Consider a finite number of points together with a graph which admits as its vertices. We will say that is admissible if is connected and for every edge , the straightline segment between and lies in . Assume that is known at one of the points . For each , we may then define to be the vector space spanned by all analytic continuations of by following edges of . We may compute the collection of these vector spaces using the following algorithm:

Algorithm continue

Input: an analytic function at and an admissible graph with

Output: for each , a basis of

Step 1. [Initialize]

Set and for all

Step 2. [Saturate]

For any , and do

If , then

Set and repeat step 2

Step 3. [Terminate]

Return

Remark 2. The test requires an equality test for analytic functions. From the heuristic point of view, we may simply fix two numbers and test whether the first Taylor coefficients coincide up to a relative error .

Assume now that singularities in are known, say . We need a test in order to check whether , as well as a way to find at least one of the remaining singularities in if .

Let be such that for all and . We now consider a graph in the form of a hexagonal honeycomb, which fills up the disk and such that each edge has length (see also figure 4 below). Picking sufficiently at random, the graph is also admissible, so we may apply the algorithm continue. For each vertex , we also compute the minimum of the convergence radii of the elements in . We claim that if and only if

for all . Indeed, if , then the equality clearly holds. Assume for contradiction that the equality holds for all , even though . For each of the outermost vertices of the honeycomb with , we have . This implies that , whence lies in the interior of one of the cells of the honeycomb. Let with be such that is minimal. Now consider a vertex of the cell of the honeycomb which contains such that (see figure 4). Then . Now also implies that . Consequently, . This contradiction completes the proof of our claim.

Whenever (9) does not hold for some , then this indicates the presence of a singularity in near . It thus suffices to determine the dominant singularities of the elements in in order to find at least one of these missing singularities.

Remark 3. In practice, we only have heuristic algorithms for approximating the radii . Therefore, and as in the previous subsection, the test (9) should really be replaced by an approximate equality test.

Remark 4. We picked a graph in the form of a honeycomb in order to simplify the proof of our claim. In practice, it is more efficient to construct graphs with small edges near the singularities and larger edges elsewhere, such as Voronoï diagrams. Following [CC90, Hoe99], it can also be shown that the optimal shape of a polygon for turning around a singularity contains edges and not (at least in the case when is the solution to a differentially algebraic equation).

 Figure 4. Illustration of the proof of our claim that if and only if (9) holds for all . The colored region corresponds to the set of all points with .

## 5.Meromorphic functions

### 5.1.Computing the denominator on a compact disk

In the case when is meromorphic on the compact disk of radius , then there exists an alternative for the algorithms from section 4: we may directly search for a polynomial such that the radius of convergence of is strictly larger than . If such a polynomial exists, then we may select to be monic and of minimal degree; this particular polynomial will be called the denominator of on . If is not meromorphic on , then we define . Given , we also define the guarded denominator by if or and otherwise. Guarded denominators may be computed using simple linear algebra, as follows:

Algorithm denom

Input: the first coefficients of , a radius and a degree bound

Output: an approximation of a monic polynomial with ,

chosen of minimal degree , or (failed)

Step 1. [Initialize]

Step 2. [Determine ]

Solve the linear system

 (10)

Set

Step 3. [Terminate or loop]

Heuristically determine , based on the first coefficients of

If then return

If then return

Set and go to step 2

In order to study the reliability of this algorithm, let us first introduce some more notations. First of all, given integers and , we will denote

The approximate correctness of the algorithm relies on the following technical lemma which will be proved in the next section.

Lemma 5. If admits exactly roots in the closed unit disk, when counted with multiplicities, then the matrix norm of satisfies for .

Theorem 6. Let and assume that the test in the algorithm always returns true if and only if and . Then if and only if . Moreover, if , then

for and any .

Proof. Modulo a scaling , we may assume without loss of generality that . Furthermore, the proof is clear if , so we will assume that and . The test returns false as long as (or if ). Assume now that reaches in the algorithm. Then the computed satisfies

since

Now let . Given , we have . Using lemma 5, it follows that .

Remark 7. The algorithm uses a very simple -step search for finding a polynomial of minimal degree with . Using a binary search (doubling at each step at a first stage, and using a dichotomic search at a second stage), the number of steps can be reduced to .

### 5.2.Proof of lemma 5

Proof. Let us first consider the case when is a rational function with , , and such that only admits roots in the closed unit disk. We denote and for each . Using partial fraction decomposition, there exist numbers with such that

In particular, we may factor

where is a matrix with entries in and is a diagonal matrix with entries (and such that each occurs times). The functions with being linearly independent, the matrix is invertible in , whence

This completes the proof of the lemma in the special case.

In general, we may write , with as above and where is an analytic function on the closed unit disk. Then

since

for some . This completes the proof in the general case.

## 6.Detecting dependencies via analytic continuation

Keeping the same spirit as in section 5, we will now turn our attention to two larger classes of algebraic and Fuchsian analytic functions on the compact disk of radius . Given a function in one of these classes, we will show how to find the defining equation for , under the assumption that we have a high accuracy algorithm for its analytic continuation. As in remark 2, we will also assume the existence of a heuristic equality test for analytic functions.

### 6.1.Algebraic dependencies on a compact disk

Let be an analytic function which is given by its power series at the origin. Assume that can be continued analytically on a Riemann surface above the closed disk of radius minus a finite set of points . Let be the set of analytic functions on . We say that is algebraic on if there exists a polynomial relation

 (11)

with . In that case, we may normalize the relation such that has minimal degree and such that is a monic polynomial of minimal degree. In particular, all roots of are inside the disk .

Given a function which is algebraic on , this function clearly satisfies the assumption A in section 4.3, whence we may compute the singularities using the algorithm described there. Conversely, the example

shows that there are functions which satisfy the assumption A, but which are not algebraic on the disk .

Assume that is algebraic on . For each singularity , consider a path from the origin to a point near which avoids the other singularities, and let be the operator which performs an analytic continuation along , one turn around , followed by an analytic continuation along . With these operators, we may use the following algorithm for the detection of algebraic dependencies:

Algorithm alg_dep

Input: an analytic function above and bounds , and

Output: a normalized algebraic dependency (11) with and , or (failed)

Step 1. [Initialize]

Set

Step 2. [Saturate]

If then return

If for all , then go to step 3

Repeat step 2

Step 3. [Terminate]

Denote

Compute

For each , compute

If for some , then return

If then return

Return

Theorem 8. Modulo oracles for equality testing of analytic functions and the computation of guarded denominators and least common multiples, the algorithm alg_dep is correct.

Proof. Assume that satisfies a normalized relation (11), with and . Since only contains distinct roots of , we have in throughout the algorithm. In particular , and we ultimately obtain stabilization .

At this point, analytic continuation around any of the points leaves the polynomial invariant, so the coefficients of are analytic and single-valued on . On the other hand, given a singularity , each solution is also given by a convergent Puiseux series near , whence so are the coefficients of . Since the only Puiseux series without monodromy around are Laurent series, it follows that the coefficients of are meromorphic on .

By the minimality assumption on , it follows that and . Since each coefficient is meromorphic on , we may use the algorithm denom from the previous section in order to compute a polynomial of minimal degree such that . The monic least common multiple is nothing but the monic polynomial of minimal degree such that .

Remark 9. In a world without oracles, the algorithm remains correct in the following sense: if the tests and always return the right answer, then we will reach the line whenever the desired relation exists. For the final normalization step, we need a numerical algorithm for the computation of least common multiples which allows for small errors in the input coefficients. This is similar to the computation of approximate g.c.d.s for which there exists an extensive literature; see for example [CGTW95, KL96], or [Zen04] and references therein.

Remark 10. As an additional safeguard, we may heuristically compute the radii of convergence of and check whether they are indeed superior to .

### 6.2.Fuchsian dependencies on a compact disk

We say that is Fuchsian on , if satisfies a linear differential equation

 (12)

where , and is Fuchsian at each point in . Again, we may normalize (12) such that has minimal order and such that is a monic polynomial of minimal degree.

Given a function which is Fuchsian on , the function now satisfies the assumption H in section 4.3, so we may again compute the singularities using the algorithm described there. With as in the previous section, we may detect Fuchsian dependencies as follows:

Algorithm fuch_dep

Input: an analytic function above and bounds , and

Output: a normalized Fuchsian dependency (12) with and , or (failed)

Step 1. [Initialize]

Set

Step 2. [Saturate]

If then return

Let denote the vector space generated by a finite set of vectors

If for all , then go to step 3

for and with

Repeat step 2

Step 3. [Terminate]

Denote

Compute

in the skew polynomial ring , where denotes

For each , compute

If for some , then return

If then return

Return

Theorem 11. Modulo oracles for equality testing of analytic functions and the computation of guarded denominators and least common multiples, the algorithm fuch_dep is correct.

Proof. Assume that satisfies a normalized Fuchsian relation (12), with and . Throughout the algorithm, the set only contains linearly independent solutions to . Therefore, the smallest operator which vanishes on divides in on the right. In particular , and we ultimately obtain stabilization .

Consider one of the singularities . Since is Fuchsian at , the equation admits a fundamental system of solutions of the form

where and are convergent power series. The coefficients of lie in the field generated by all generalized power series of this form. Again, the only elements of with a trivial monodromy around are convergent Laurent series at . Since analytic continuation around leaves the operator unchanged, it follows that the coefficients of are meromorphic on . We conclude in a similar way as in the proof of theorem 8.

Remark 12. The hypothesis that admits at worse a Fuchsian singularity at is essential for the algorithm to work. For instance, in the case of the function

the monodromy around is trivial, whence applying the algorithm for would simply result in having at step 3. Even though has no monodromy around , this function is no longer a Laurent series. In fact, the desired vanishing operator has second order in this case, but it cannot be read off directly from .

## 7.Detecting dependencies via orthogonalization

In the case when we have no algorithm or no efficient algorithm for the analytic continuation of , the algorithms of section 6 can no longer be used. In this section, we will describe a purely numerical approach for the detection of analytic relations on a closed disk of radius . Modulo the change of variables , it suffices to consider the case when .

Given an order , we will denote by the set of power series with for all . When truncating the usual multiplication on at order , we give a ring structure, which is isomorphic to . We will denote by the Hilbert space of all power series with finite norm

Notice that any power series that converges on the closed unit disk has finite norm, but a power series with finite norm is only guaranteed to converge on the open unit disk. More generally, the norm of a vector is given by

The spaces can be considered as an increasing sequence of Hilbert spaces, for the restrictions of the norm on to the .

Let and assume that are power series with radii of convergence at least (notice that we do not assume the to be of finite norm). We want to solve the equation

 (13)

We will denote the vector space of all such relations by . Since the equation (13) involves an infinite number of coefficients, we need to consider its truncated version at a finite order . Replacing by their truncations in , we will search for non-trivial solutions of the equation

 (14)

In a way which will be made more precise below, we will require the norms of and to remain of the same orders of magnitude as the vector . We will denote by the vector space of all which satisfy (14). If the norms of the solutions to the truncated problems tend to a limit for , then we will prove that these solutions tend to a solution of the original problem (13).

Let us reformulate the truncated problem in terms of linear algebra. The series give rise to an matrix

The unknown series give rise to a row vector

Setting , we then have

Putting the first entries on the right hand side and grouping by packets of entries, it follows that encodes the relation . This reduces the problem to finding those vectors for which is of the same order of magnitude as the vector .

We start with the computation of a thin LQ decomposition of . This can for instance be done using the Gram-Schmidt process: starting with the first row, we orthogonally project each row on the vector space spanned by the previous rows. This results in a decomposition

where is a lower triangular matrix with ones on the diagonal and is an matrix, whose rows are mutually orthogonal (i.e. ). Now consider the matrix formed by the last rows of . Then each row gives rise to a relation (14), encoded by . Moreover, this relation is normal or -normal, in the sense that and for all . Since is the shortest vector in , the relation is also minimal in norm, among all -normal relations. Choosing the row for which is minimal, our algorithm simply returns the corresponding relation. Then our algorithm has the following fundamental property:

Proposition 13. The algorithm returns a normal relation for which is minimal.

Let us now return to the case when are no longer truncated, but all have a radius of convergence . A relation (13) is again said to be normal or -normal if and for all . Under the limit , we claim that our algorithm finds a minimal normal relation, if there exists a relation of the form (13):

Theorem 14.

1. Assume that . Then contains a minimal normal relation. For each , contains at most one minimal -normal relation.

2. Assume that is a minimal -normal relation. For each , let be the truncation of at order and consider the corresponding minimal -normal relation . Then the relations converge to in .

3. If , then the norms , with as in (b), are unbounded for .

Proof. A non trivial relation is easily normalized: we first divide by , where . We next divide by , where is largest with . Now the set of all -normal relations is a closed affine subspace of . The orthogonal projection of on this subspace yields a unique -normal relation of minimal norm. This proves (a).

Assume that there exists an -normal relation (13), and let be the minimal such relation. Given an order , consider the minimal -normal relation at this order. Truncation of this relation at a smaller order yields an -normal relation at order with , whence

 (15)

Moreover, since is the projection of on the affine space of -normal relations at order , we have and

 (16)

Similarly, truncation of the relation at finite order yields an -normal relation which satisfies

In particular, , so that

For , it follows that , whence

and

In other words, is a Cauchy sequence, which converges to a limit with . Now for each , we may write and denote by the truncation of the vector at order . By assumption, the inner products and vanish. Since for all and , we also have for , whence . Consequently, is an -normal relation in . By the minimality hypothesis on , we conclude that , which proves (b).

In general, the existence of a bound with

still ensures to be a Cauchy sequence, and its limit yields an -normal relation (13). This proves the last assertion (c).

Remark 15. For some very specific types of singularities, we would like to mention the existence of an alternative strategy for the determination of linear dependencies. This strategy relies on asymptotic extrapolation [Hoe09] and does not involve analytic continuations around singularities, as in section 6. Let us illustrate the method on the example of a function with an isolated smallest singularity at of the form

where and are analytic at . Then the asymptotic extrapolation of yields an asymptotic expansion of the form

Using singularity analysis [FS96], we may then recover the function from the coefficients . However, this technique only works in special cases, since the first terms of the asymptotic expansion may hide other terms, which need to be taken into account when searching for exact dependencies.

## 8.Numerical experiments

A first implementation of the guessing algorithm from section 7 has been made in the Mathemagix system [HLM+02]. It is instructive to test the algorithm on examples for which the existence or absence of linear dependencies is known. In particular, in order to turn theorem 14 into an algorithm which requires no manual tweaking, we need more precise information about the corresponding rates of convergence and divergence of the norms .

### 8.1.Examples of recognized relations

We have tested the orthogonalization algorithm on three examples of increasing difficulty, for which linear dependencies are known to exist. In order to avoid problems due to numerical instability, we have used a large working precision, of 1024 or 2048 bits.

##### Single poles
The easiest example on which we can run the algorithm is

Here follow the values for at different orders and :

The clearly converge to a limit with radius of convergence . It should be noticed that we do not have . In other words, the “best” numeric relation (from the point of view of -norms) does not coincide with the “best” algebraic relation (as found by using Padé-Hermite approximation). A closer examination of the result shows that

In particular, decreases if increases.

##### Logarithmic singularity
Our second example concerns a logarithmic singularity:

We obtain:

The convergence only becomes apparent at higher orders. Contrary to the previous example, it seems that the series in the computed relation all have a radius of convergence . The norms of the computed results are given by , , and .

##### Number of alcohols
A more interesting example from combinatorics is the enumeration of the number of alcohols of the form [Pól37]. Its generating series satisfies the functional equation

The asymptotic behaviour of the coefficients is determined by the asymptotic behaviour of at its singularity of smallest norm. Since for all , this so-called dominant singularity necessarily lies on the positive real axis, and coincides with the radius of convergence of . Using asymptotic extrapolation [Hoe09], we find that . In order to investigate for close to , we apply our algorithm to

The translation is done using power series evaluation until stabilization at the working precision. Although this was most straightforward to implement for this particular example, there exist better strategies for zooming in, as mentioned at the end of section 4.1. At different orders, we obtain:

The corresponding norms are given by , , and . Again, the convergence is rather slow, and the computed series all seem to have radius of convergence . Since , this means that we indeed guessed an analytic relation between and on the unit disk.

Remark 16. The last two examples show that our algorithm usually returns a relation whose series all have radius of convergence , if such a relation indeed exists. Larger radii of convergence can be obtained simply by running the algorithm on and scaling back the result.

### 8.2.Rate of divergence in case of independence

It is also instructive to run the algorithm on examples where the are known to be analytically independent. In that case, it is important to examine the behaviour of the norms for and find more precise criteria which will enable us to discard the existence of a nice dependency with a reasonable degree of certainty.

##### Logarithmic singularity

where . Running our algorithm directly on this function yields the following results for :

The results indeed do not converge and the corresponding sequence of norms , , and diverges at moderate exponential speed.

##### Various singularities
In the above example, it is interesting to study the rate of divergence of the sequence of norms for other values of . More generally, we can consider various types of singularities, such as

The series is a series whose coefficients are chosen according to a random uniform distribution on . The results are shown in table 1 below. For , and , it seems that the norm does not much depend on the precise type of singularity, but only on and the truncation order . For the last series , the dependencies on and approximately seem to follow the law

 (17)

This idealized law needs to be adjusted for functions , , with more common types of singularities. Although (17) remains asymptotically valid for large values of , the factor needs to be replaced by a smaller number for moderate values. For , this factor rather seems to be of the form , where the constant depends on the nature of the singularity. Also, the linear dependency of on is only reached for large values of .

 32 64 128 256 2.2960 3.7208 9.7394

Table 1. Computed values of for various types of singular behaviour and orders .

##### Various singularities and
We have also applied our algorithm jointly to several as above and studied the dependency of on . The results are shown in table 2 below. In the bottom rows, stand for distinct uncorrelated random series . In that case, the relation (17) generalizes to

 (18)

It also seems that the law can be adapted to functions with more common types of singularities, along similar lines as before. It would be interesting to have a theoretical explanation for the empirical laws (17) and (18). The division by in (18) is probably due to the fact that there are unknowns in the equation (13).

Of course, in the case that we are considering functions with several singularities on our disk of interest, only the singularities for which is smallest are “numerically visible” (except when there exist relations which not involve these dominant singularities ).

 32 64 128

Table 2. Computed values of for different input vectors of increasing dimensions .

### 8.3.Towards criteria for an automatic final decision

In order to design a blackbox algorithm for the detection of dependencies (13), we need to make a final decision on whether we declare the input functions to be dependent or not. In view of the empirical law (18), and whatever decision criterion we implement, the outcome can only be trusted if the following two conditions are satisfied:

1. The number should not be too large.

2. There exists a large constant for which all singularities of are concentrated inside the disk of radius or outside the disk of radius .

As usual, if we are interested in dependencies near an isolated singularity, then the second condition can often be achieved by zooming in on the singularity.

Our empirical observations in the previous section suggest that we should base the final decision on the empirical law (18). First of all, in the case of independence, we should observe a more or less geometric increase of the norm when running the algorithm for different orders . If only admits an isolated singularity in , then we may also compute its norm and compare the computed values of with the expected values, as given by the law (18), or a suitable adaptation of this law when becomes small. This method is most effective when is large, so it is usually recommended to zoom in on the singularity before doing this comparison. In fact, if the law (18) is satisfied for different zooming factors, then this may further increase our confidence that the input functions are independent.

For any numerical checks based on the law (18) or a refinement of it, we also recommend to precondition the input series . In particular, we recommend to multiply each by a suitable constant, ensuring that

whenever we apply our algorithm at order . Here is computed using one of the algorithms from section 3.

In fact, there is a trade-off between zooming in on the dominant singularity (thereby ensuring a large value of ) and using the original input coefficients (which are given with a higher accuracy and order). In order to get more insight in how far we should zoom in, we need to analyze the cost of the algorithm from section 7.

The current working precision should in general be taken larger than in order to keep the method numerically stable. Denoting by the cost for multiplying two -bit numbers, a naive implementation of the Gram-Schmidt orthogonalization procedure yields a total cost of . Denoting by the cost of multiplying two matrices with bit entries, and using a blockwise Gram-Schmidt procedure (similar to the procedure for matrix inversion in [Str69]), we obtain the better bound . However, the matrix from section 7 has a very special form. With more work, it might therefore be possible to save an additional factor , but we have not actively tried to do so yet.

Taking into account the effect of a possible zoom, we may next evaluate the computational cost in terms of the desired “output quality”. As a definition of the output quality, we may take the expected value of in the case when are independent. In order to obtain accurate results, we need to compute with a bit precision larger than . In terms of , the time complexity thus becomes

where is the exponent of matrix multiplication [Str69, Pan84, CW87, Vas12]. The complexity bound makes it clear that we should put a lot of effort into keeping large. If satisfies a differential equation, then zooming in can be done with low computational cost (see section 4.1). Otherwise, the required expansion order grows quickly with the inverse of the distance to the singularity. Indeed, for , the formula (8) becomes

where we recall that, in order to apply the orthogonalization process after zooming in, we need to expand at an order which is times larger than before.

## Bibliography

[BBKW06] D. Bailey, J.M. Borwein, V. Kapoor, and E.W. Weisstein. Ten problems in experimental mathematics. The American Mathematical Monthly, 113(6):481–509, 2006.

[BD09] J.M. Borwein and K.J. Devlin. The Computer As Crucible: An Introduction to Experimental Mathematics. A.K. Peters Series. A.K. Peters, 2009.

[Bel11] K. Belabas. Journées Nationales de Calcul Formel (2011), volume 2 of Les cours du CIRM, chapter Théorie algébrique des nombres et calcul formel. CEDRAM, 2011. Exp. No. 1, 40 pages, http://ccirm.cedram.org/ccirm-bin/item?id=CCIRM_2011__2_1_A1_0.

[Bir13] G.D. Birkhoff. Equivalent singular points of ordinary differential equations. Math. Ann., 74:134–139, 1913.

[BK78] R.P. Brent and H.T. Kung. Fast algorithms for manipulating formal power series. Journal of the ACM, 25:581–595, 1978.

[BL94] B. Beckermann and G. Labahn. A uniform approach for the fast computation of matrix-type Padé approximants. SIAM J. Matrix Analysis and Applications, 15:804–823, 1994.

[BZ91] C. Brezinski and R. Zaglia. Extrapolation Methods. Theory and Practice. North-Holland, Amsterdam, 1991.

[CC90] D.V. Chudnovsky and G.V. Chudnovsky. Computer algebra in the service of mathematical physics and number theory (computers in mathematics, stanford, ca, 1986). In Lect. Notes in Pure and Applied Math., volume 125, pages 109–232, New-York, 1990. Dekker.

[CGTW95] R. M. Corless, P. M. Gianni, B. M. Trager, and S. M. Watt. The singular value decomposition for polynomial systems. In Proc. ISSAC '95, pages 195–207. ACM Press, 1995.

[CP78] B. F. Caviness and M. J. Prelle. A note on algebraic independence of logarithmic and exponential constants. SIGSAM Bull., 12/2:18–20, 1978.

[CW87] D. Coppersmith and S. Winograd. Matrix multiplication via arithmetic progressions. In Proc. of the Annual Symposium on Theory of Computing, pages 1–6, New York City, may 25–27 1987.

[Der94] H. Derksen. An algorithm to compute generalized Padé-Hermite forms. Technical Report Rep. 9403, Catholic University Nijmegen, January 1994.

[DL89] J. Denef and L. Lipshitz. Decision problems for differential equations. The Journ. of Symb. Logic, 54(3):941–950, 1989.

[Fab85] E. Fabry. Sur les intégrales des équations différentielles linéaires à coefficients rationnels. PhD thesis, Paris, 1885.

[Fau94] J.-C. Faugère. Parallelization of Gröbner basis. In Hoon Hong, editor, Parallel and Symbolic Computation, volume 5 of Lect. Notes Series in Computing. Pasco'94, World Scientific, 1994.

[FS96] Ph. Flajolet and R. Sedgewick. An introduction to the analysis of algorithms. Addison Wesley, Reading, Massachusetts, 1996.

[GG02] J. von zur Gathen and J. Gerhard. Modern Computer Algebra. Cambridge University Press, 2-nd edition, 2002.

[HLM+02] J. van der Hoeven, G. Lecerf, B. Mourrain, et al. Mathemagix, 2002. http://www.mathemagix.org.

[Hoe99] J. van der Hoeven. Fast evaluation of holonomic functions. TCS, 210:199–215, 1999.

[Hoe01] J. van der Hoeven. Zero-testing, witness conjectures and differential diophantine approximation. Technical Report 2001-62, Prépublications d'Orsay, 2001.

[Hoe02] J. van der Hoeven. Relax, but don't be too lazy. JSC, 34:479–542, 2002.

[Hoe05] J. van der Hoeven. Effective complex analysis. JSC, 39:433–449, 2005.

[Hoe07] J. van der Hoeven. On effective analytic continuation. MCS, 1(1):111–175, 2007.

[Hoe09] J. van der Hoeven. On asymptotic extrapolation. JSC, 44(8):1000–1016, 2009.

[Hoe11] J. van der Hoeven. Journées Nationales de Calcul Formel (2011), volume 2 of Les cours du CIRM, chapter Calcul analytique. CEDRAM, 2011. Exp. No. 4, 85 pages, http://ccirm.cedram.org/ccirm-bin/fitem?id=CCIRM_2011__2_1_A4_0.

[KL96] N. K. Karmarkar and Y. N. Lakshman. Approximate polynomial greatest common divisors and nearest singular polynomials. In Y.N. Lakhsman, editor, Proc. ISSAC '96, pages 35–42, Zürich, Switzerland, July 1996. ACM Press.

[Lan71] S. Lang. Transcendental numbers and diophantine approximation. Bull. Amer. Math. Soc., 77/5:635–677, 1971.

[LLL82] A.K. Lenstra, H.W. Lenstra, and L. Lovász. Factoring polynomials with rational coefficients. Math. Ann., 261:515–534, 1982.

[Moo66] R.E. Moore. Interval Analysis. Prentice Hall, Englewood Cliffs, N.J., 1966.

[Pan84] V. Pan. How to multiply matrices faster, volume 179 of Lect. Notes in Math. Springer, 1984.

[Poi86] H. Poincaré. Sur les intégrales irrégulières des équations linéaires. Acta Math., 8:295–344, 1886.

[Pól37] G. Pólya. Kombinatorische Anzahlbestimmungen für Gruppen, Graphen und chemische Verbindungen. Acta Mathematica, 68:145–254, 1937.

[PTVF07] W.H. Press, S.A. Teukolsky, W.T. Vetterling, and B.P. Flannery. Numerical recipes, the art of scientific computing. Cambridge University Press, 3rd edition, 2007.

[Ric97] D. Richardson. How to recognise zero. JSC, 24:627–645, 1997.

[Ric07] D. Richardson. Zero tests for constants in simple scientific computation. MCS, 1(1):21–37, 2007.

[Str69] V. Strassen. Gaussian elimination is not optimal. Numer. Math., 13:352–356, 1969.

[SW05] A.J. Sommese and C.W. Wampler. The Numerical Solution of Systems of Polynomials Arising in Engineering and Science. World Scientific, 2005.

[SZ94] B. Salvy and P. Zimmermann. Gfun: a Maple package for the manipulation of generating and holonomic functions in one variable. ACM Trans. on Math. Software, 20(2):163–177, 1994.

[Vas12] V. Vassilevska Williams. Multiplying matrices faster than Coppersmith-Winograd. In ACM Symp. on Theory of Computing (STOC'12), 2012. To appear. Preliminary version available at http://cs.berkeley.edu/~virgi/matrixmult.pdf.

[Ver96] J. Verschelde. Homotopy continuation methods for solving polynomial systems. PhD thesis, Katholieke Universiteit Leuven, 1996.

[Was67] W. Wasow. Asymptotic expansions for ordinary differential equations. Dover, New York, 1967.

[Wei00] K. Weihrauch. Computable analysis. Springer-Verlag, Berlin/Heidelberg, 2000.

[Wen01] E. J. Weniger. Nonlinear sequence transformations: Computational tools for the acceleration of convergence and the summation of divergent series. Technical Report math.CA/0107080, Arxiv, 2001.

[Wil94] H.S. Wilf. Generating functionology. Academic Press, 2nd edition, 1994. http://www.math.upenn.edu/~wilf/DownldGF.html.

[Zen04] Z. Zeng. The approximate gcd of inexact polynomials Part I: a univariate algorithm. http://www.neiu.edu/~zzeng/uvgcd.pdf, 2004.