| 
      
                
. This work was
                partially supported by the ANR Gecko and ANR-09-JCJC-0098-01
                
            Given  
              | 
        
      Consider an infinite sequence 
 of complex
      numbers. If 
 are the coefficients of a formal
      power series 
, then it is
      well-known [Pól37, Wil94, FS96]
      that the asymptotic behaviour of the sequence 
 is
      closely related to the behaviour of the generating function 
 near its dominant singularity. Now, if 
      is the solution to some complicated equation, then it can be hard to
      compute the asymptotic behaviour using formal methods. On the other
      hand, the coefficients 
 of such a solution 
 can often be computed numerically up to a high order.
      With this numerical evidence at hand, it is natural to raise the
      following questions:
    
          Can we guess the asymptotic behaviour of 
?
        
          Can we guess the behaviour of 
 near its
          dominant singularity?
        
These questions can be regarded as part of the construction of a more general toolbox for the “experimental mathematician” [BBKW06, BD09]. More specifically, we advocate the systematic integration of “guessing tools” into symbolic computation packages. Indeed, current systems can be quite good at all kinds of formal manipulations. However, in the daily practice of scientific discovery, it would be helpful if these systems could also detect hidden properties, which may not be directly apparent or expected. Furthermore, the guessing tool is allowed to be heuristic, so that it only suggests hidden properties; at a second stage, one may then search for full proofs using other techniques.
      One well-known tool in this direction is the LLL-algorithm [LLL82].
      Given 
 numbers 
,
      it can be used in order to guess relations of the form
    
![]()  | 
        (1) | 
      Given 
 formal power series 
, algorithms for the computation of
      Padé-Hermite forms [BL94, Der94] can be
      used in order to guess linear relations
    
![]()  | 
        (2) | 
      A well-known implementation is provided by the 
, the
      
 or a linear differential equation
      with coefficients in 
 satisfied by 
. Indeed, it suffices to take 
      in (2) in order search for small linear differential
      equations satisfied by 
.
    
      Unfortunately, many interesting formal power series 
      do not admit closed form formulas and do not satisfy linear differential
      equations with polynomial coefficients. In that case, we can still use
      asymptotic extrapolation [Hoe09] in order to guess the
      asymptotic behaviour of the coefficients. However, this only provides us
      some rough idea about the behaviour of 
 at its
      dominant singularity. In practice, it often happens that 
 locally satisfies an algebraic or differential equation
      with analytic coefficients, even though these coefficients fail to be
      polynomials. For instance, combinatorics [Pól37, FS96] is full with examples of generating functions which are
      not algebraic, but whose dominant singularities are algebraic.
    
      In this paper, we will describe two approaches to detect analytic
      dependencies on a compact disk: the first one assumes that we have an
      algorithm for the analytic continuation of 
 and
      relies on the monodromy of 
 at its singularities.
      The second approach is purely numerical and makes no special assumptions
      on 
. On our way, we will
      encounter various interesting related problems, such as the
      determination of the radius of convergence or the singularities of 
. We will also propose heuristic
      solutions to these problems, thereby extending the basic toolbox for
      experimenting with analytic functions.
    
Since all algorithms in this paper are directly or indirectly based on heuristics, it is important to investigate how much confidence we can attach to the computed results. In section 2, we will present a survey of different kinds of heuristic algorithms which are used in symbolic computation. Some of these heuristics are relatively benign when compared to others, so it will be useful to have some general insights on the reliability of different types of heuristic algorithms.
      The remainder of the paper is divided into two main parts. In the first
      part (sections 3, 4 and 5), we
      will develop some basic tools for later use. In the second part, we turn
      to the main topic of this paper: the disclosure of local analytic
      relations between analytic functions. Two main types of analytic input
      functions 
 will be considered:
    
            Functions for which we can merely compute a large number of the
            Taylor coefficients 
 at the origin.
          
            Functions for which we have a reasonably efficient algorithm for
            their analytic continuation, so that we can also compute the
            Taylor coefficients of 
 at other points
            besides the origin.
          
      The second situation typically arises when 
 is
      the solution of a differential or more general functional equation; see
      section 4.1 for a further discussion on analytic
      continuation.
    
      One of the most basic problems concerning an analytic function 
 which is only known by its Taylor coefficients, is to
      determine its radius of convergence. In section 3, we will
      present two methods. The first one is related Cauchy-Hadamard's formula
      and provides rough approximations under mild assumptions on 
. The second method provides much better
      approximations, but only works if 
 admits a
      single dominant singularity of a simple type.
    
      Building on the algorithm for approximating the radius of convergence of
      
, we next turn our attention
      to the problem of locating its singularities. In section 4.2,
      we first restrict ourselves to the dominant singularities of 
 (i.e. the singularities of minimal norm).
      Assuming AN2, we next present algorithms for the
      exploration of the Riemann surface of 
 beyond its
      dominant singularities.
    
      In the special case when 
 is meromorphic on a
      compact disk 
, section 5 contains a special purpose algorithm for the determination
      of a polynomial 
 such that 
      is analytic on 
. This
      algorithm works under the hypothesis AN1 and induces
      better algorithms for the problems in sections 3 and 4 in this special case.
    
      In the second part of this paper, we turn to the detection of
      dependencies between analytic functions on a compact disk 
. More precisely, assume that 
      are fixed convergent power series. We are interested in the
      determination of linear dependencies of the form
    
![]()  | 
        (3) | 
      where 
 are analytic on 
. Modulo a scaling 
,
      we may assume without loss of generality that 
.
    
      In section 6, we assume AN2 and first
      consider the two special cases of algebraic and Fuchsian dependencies.
      In the case of algebraic dependencies, we take 
      for a fixed function 
, and
      also require that 
. In the
      case of Fuchsian dependencies, we take 
 for a
      fixed function 
, and again
      require that 
. The second
      algorithm only succeeds in the case when all singularities of 
 on the disk are of Fuchsian type (see sections 3.2
      and 6.2 for detailed definitions). The main idea behind the
      method is to compute the set of of functions generated by 
 and its analytic continuation around its singularities in
      
. If there exists an
      algebraic dependency, then this set is finite and we may use it to find
      the dependency. If there exists a Fuchsian dependency, then the set is
      contained in finite dimensional vector space, and we may construct the
      dependency from a basis of this vector space.
    
In section 7, we only assume AN1 and consider the general problem of determining dependencies of the form (3). We will describe a purely numerical algorithm based on Gram-Schmidt orthogonalization for finding such relations. The idea is to simply truncate the power series expansions and then apply the numerical method. We will prove that an asymptotic convergence theorem which shows that this approach is at least correct “at the limit”.
      In the last section 8, we will present some numerical
      experiments with the algorithm from section 7. In section
      8.1, we first consider some examples of relations which
      were recognized by the algorithm. In section 8.2, we will
      also examine the behaviour of the algorithm in the case when 
 are analytically independent. In fact, the algorithm from
      section 7 still requires some human cooperation for
      deciding whether we really found a relation. Based on the numerical
      experiments, we conclude with some perspectives for the design of a
      fully automatic algorithm.
    
Acknowledgments. We would like to thank the three referees for their detailed and valuable comments on a first version of this paper.
As soon as we enter the area of heuristic and non mathematically proven algorithms, it is natural to ask how much confidence can be attached to the output. In fact, there is a large spectrum of situations which can occur, depending on the precise nature of the heuristics. There may also be a trade-off between computational complexity and correctness: is it better to rely on a slow deterministic algorithm for polynomial factorisation over a finite field, or on a fast probabilistic algorithm?
Without striving for exhaustion, section 2.1 a catalogue of different kinds of heuristic algorithms which occur in the area of symbolic computation. Each of the heuristic algorithms considered in this paper will fit into one of these categories. We will also describe some standard strategies in order to increase the level of confidence and examine typical examples for which heuristic algorithms may fail.
      For example, consider the class of exp-log constants, built up
      from the rationals 
 using the operations 
, exp and log. The following
      conjecture [Lan71] is a major open problem in number
      theory:
    
        
 be complex numbers which are linearly
        independent over the rational numbers 
.
        Then the transcendence degree of
      
        
        is at least 
.
      
      From the computer algebra point of view, the conjecture implies [CP78] that all numerical relations between exp-log constants
      can be deduced from the usual rules 
 and 
. Based on this fact, Richardson
      has given a zero test for exp-log constants [Ric97, Ric07] with the following properties: (1) if the algorithm
      terminates on a given input, then the result is correct; (2) if the
      conjecture holds, then the algorithm always terminates; (3) if the
      algorithm does not terminate on a given input, then a counterexample to
      the conjecture can be constructed from this input.
    
Similar situations occur in algorithmic number theory, depending on the correctness of Riemann's hypothesis [Bel11]. In this paper, the algorithms in sections 4.3 and 6 rely on zero tests for analytic functions. If we know that our analytic functions lie in special classes (such as the class of exp-log functions, or the class of differentially algebraic functions), then these zero tests might be reduced to suitable conjectures [Hoe01].
For instance, in section 3, we will give heuristic algorithms for the computation of the radius of convergence of an analytic function. This will be a basic building block for many of the other algorithms in this paper. Nevertheless, the correctness of several of these other algorithms reduces to the correctness of this basic building block.
If, for input functions in a more restricted class, we may determine their convergence radii with larger accuracy, then this will immediately make all the other algorithms more reliable. For instance, if we consider a class (e.g. rational or algebraic functions) for which convergence radii can be computed exactly, then the algorithms in section 4.2 are no longer heuristic.
 is zero, it
        suffices to pick a random number 
 and test
        whether 
 vanishes.
      
” are really correct
        with probability 
, where
        
 is the working precision. There are many other
        examples of heuristic algorithms which are correct with high
        probability. For instance, in order to show that two polynomials 
 in 
 have no common divisors,
        it suffices to pick a “sufficiently random” prime 
 and show that the reductions of 
        and 
 in 
 have no common
        divisors.
      
      We also recall that there are two classical types of probabilistic
      algorithms. The examples mentioned above are only correct with high
      probability and fall into the class of so-called Monte Carlo algorithms.
      The second kind of Las Vegas algorithms are always correct, but only
      fast with high probability. For instance, all known fast algorithms for
      factoring polynomials over a finite field 
 are
      probabilistic Las Vegas type algorithms [GG02].
    
This does not withstand that experts in this area have a great sense of how accurate and trustworthy various methods are. Ideally speaking, any numerical algorithm comes with a detailed error analysis. Most textbooks on numerical analysis start with a section on the various sources of error and how to take into account the machine accuracy [PTVF07, section 1.1]. For more statistical algorithms, there is also a tradition of trying to quantify the level of trust that can be attached to the computed results. Finally, for many algorithms, it is possible to automate the error analysis by using interval arithmetic [Moo66].
. From a theoretical point of
        view [Wei00, Hoe11], it can be investigated
        whether a given numerical algorithm is at least “ultimately
        correct” in the sense that the computed output tends to the
        mathematically correct result if 
 tends to
        infinity. If the numerical algorithm comes with a rigourous error
        analysis, then we have a means for controlling the distance 
 between our approximations at a given precision 
 and the ultimate limit. In particular, for any
        
, we may then select a
        precision 
 with 
.
      
      However, there are many situation in which we have no means for
      controlling the distance 
,
      even though our algorithm is ultimately correct. For instance, the
      computation of the radius of convergence 
 of a
      differentially algebraic function is undecidable [DL89],
      even though there exists an algorithm [Hoe05, Hoe07]
      for the computation of a sequence 
 with 
. Similarly, for a large class of
      functions, the formula (4) below gives an approximation for
      the radius of convergence of a series (when computing with a precision
      of 
 bits), but the formula is only ultimately
      correct (5).
    
For instance, in numerical analysis, a frequent trick is to rerun the same algorithm for different working precisions and compare the results. Similarly, probabilistic algorithms of Monte Carlo type can be rerun several times in order to increase our confidence that we obtained the correct result.
We also mentioned the fact that equality testing of exp-log constants is quite asymmetric in the sense that proving equalities is usually much harder than proving inequalities. This suggests the use of a mixed strategy for this problem, which combines a heuristic equality test with a deterministic, interval arithmetic based, inequality prover. In case of equality, we may optionally run Richardson's more expensive algorithm and try to prove the equality.
In applications, we notice that heuristic algorithms are often used in a similar way: to speed up the main algorithm, which can be perfectly deterministic. A nice example is the computation of a Gröbner basis over the rationals using modular arithmetic. In Faugère's software, modular arithmetic is used in order to get a precise idea about those S-polynomials that reduce to zero, thereby avoiding many unnecessary reductions when performing the actual computations over the rationals [Fau94].
In the particular case of a heuristic algorithm whose purpose is to guess dependencies, we finally notice that we may always build in the trivial safeguard to check all proposed relations a posteriori; see also remark 10 in section 6.1.
One particular angle of attack concerns the computational complexity or the numerical stability. This means that we search for examples for which the heuristic algorithm is very slow or numerically unstable. This line of thought underlies much of the numerical experiments which are presented in section 8.
Instead of attacking the heuristic algorithms, it is also possible to look for interesting classes of examples on which the heuristic provably works (or works better). In this paper, some interesting special classes of analytic functions are meromorphic, algebraic and so-called Fuchsian functions on a compact disk (see section 6.2). Typical classes which have to be considered for counterexamples are lacunary series, functions with a natural boundary, or holonomic functions that are not Fuchsian.
      Let 
 be an analytic function which is given by
      its power series 
 at the origin. A natural
      problem is to compute the radius of convergence 
      of 
. For sufficiently large
      classes of analytic functions, such as solutions to algebraic
      differential equations over the rationals, this problem is generally
      undecidable [DL89, Hoe07], although efficient
      and high quality algorithms for the computation of lower bounds for 
 do exist in this case [Hoe07]. In this
      section, we will consider heuristic algorithms, and only assume that a
      large number of coefficients 
 of 
      are known numerically.
    
      The first idea which comes into our mind is to apply Cauchy-Hadamard's
      formula for the radius of convergence 
:
    
    
      Given 
 coefficients 
 of
      
, we may for instance use the
      approximation
    
![]()  | 
        (4) | 
      For most convergent power series 
,
      this formula is ultimately exact in the sense that
    
![]()  | 
        (5) | 
      This is in particular the case when 
 is
      ultimately convex or ultimately concave (this follows from the fact that
      the sequence 
 converges in an ultimately monotone
      way to 
, in this case). The
      set of 
 for which (5) holds is also
      stable under the transformation 
 for any 
. Of course, we may replace 
 by 
 in (5) for
      any 
. Notice however that the
      lacunary power series 
 does not satisfy (5).
    
      The formula (4) has the disadvantage that it has not been
      scaled appropriately: when replacing 
 by 
, where 
 is
      such that 
, we obtain
      different approximations for 
 and 
. This drawback can be removed by replacing
      
 by 
 for some appropriate
      coefficient index 
 with 
. In fact, one can even do better, and compute 
 using the formula 
 for
      appropriate indices 
 with 
      and 
. Let us now show how to
      read off such indices from the numerical Newton diagram of 
.
    
      Let 
, where we understand
      that 
. Then the Newton
      diagram of 
 is the convex hull of the half
      lines
    
    
      For a fixed 
, say 
, let 
,
      and consider the Newton diagram of 
.
      There exists a minimal subset 
 with 
, such that the Newton diagram is also the
      convex hull of the half lines 
 for 
. Graphically speaking, the 
      are the vertices of the Newton diagram (see figure 1).
    
      For a fixed 
, say 
, we may now determine the unique
      edge 
 of the Newton diagram such that 
, and replace the formula (4)
      by
    
![]()  | 
        (6) | 
      For a large class of convergent power series 
, the formula (6) is again ultimately
      exact. The indices 
 can be computed from the
      coefficients 
 using a linear traversal in time
      
. This yields an efficient
      algorithm for the approximation of 
 which turns
      out to be more accurate than (4) in practice. The formula
      (6) has been implemented in the 
,
      the computed radius is usually correct up to one decimal digit in the
      relative error.
    
![]()  | 
        
      The formula (6) usually yields a reasonable estimate for
      
, even in very degenerate
      cases when there are several singularities at distance 
      or close to 
. However, if
      
 admits a single isolated singularity 
 at distance 
 of the origin, with no
      other singularities at distance close to 
,
      then it is often possible to read off much more information about this
      singularity from the coefficients 
.
    
      For instance, it frequently (always?) occurs that the quotients 
 simply tend to 
 for 
. Moreover, as we will show below,
      if the singularity at 
 has a known type, then the
      approximation 
 can be further improved. If
      nothing is known about the singularity at 
,
      then we may still try and see what happens if we apply such specialized
      algorithms for various frequent types of singularities: with a bit of
      luck, we will both obtain useful information about the type of
      singularity and the numerical value of 
.
    
      We say that 
 is algebraic at 
, if 
      satisfies a polynomial equation
    
    
      where 
 are analytic functions at 
      with 
. In that case, we have
    
    
      with 
 and ramification index 
, whence
    
![]()  | 
        (7) | 
      Using the E-algorithm [Wen01, BZ91], we may
      now compute simultaneous approximations for the first
      coefficients 
, 
, 
,
      
, etc. of the
      expansion (7). It turns out that this strategy greatly
      improves the accuracy of the approximation of 
      (see also [Hoe09]).
    
      Similarly, we say that 
 is Fuchsian at
      
, if 
      satisfies a linear differential equation
    
    
      where 
 and 
 are analytic
      functions at 
 with 
.
      In that case, the Taylor coefficients 
 satisfy
      the asymptotic expansion
    
    
      where 
, 
      and the 
 are polynomials in 
      of degrees 
 [Fab85, Poi86,
      Bir13, Was67]. Again, the E-algorithm or more
      general algorithms for asymptotic extrapolation [Hoe09] can
      be used to compute 
 with a high accuracy. Notice
      that these algorithms also provide estimates for the accuracies of the
      computed approximations.
    
      In cases where nothing particular is known about the behaviour of 
 at its dominant singularity, the strategy of
      asymptotic extrapolation [Hoe09] still may work and both
      provide useful information about the nature of the singularity and the
      numerical value of 
.
    
      In order to explore the analytic function 
      determined by the coefficients 
 more closely, and
      in particular determine its singularities in a given region, it is
      useful to have a means for performing the analytic continuation of 
.
    
      The way 
 is given sometimes provides us with such
      a means. For instance, if 
 is the solution to an
      initial value problem
    
    
      then the continuation 
 of 
      at any sufficiently small 
 is again the solution
      of an initial value problem
    
    
      In [Hoe05, Hoe07] it is shown in detail how to
      turn this into an algorithm for the analytic continuation of 
. Notice also that the power series
      solutions to an initial value problem can be computed efficiently using
      the algorithms from [BK78, Hoe02].
    
      In general, it is always possible to compute the continuation 
 numerically. However, this kind of “poor
      man's” analytic continuation induces a big loss of accuracy. Let
      us illustrate this on the fundamental example
    
    
      In order to compute 
 with a precision of 
 bits using the formula
    
    
      we need to truncate this expansion at an order 
      for which
    
    
      Putting 
 and using Stirling's formula, this
      yields
    
      For instance, in order to expand 
 at order 
 with precision 
,
      we get
    
    
      and we need an expansion of 
 at the origin at
      order 
. In general, for every
      analytic continuation step, we need to multiply the expansion order by a
      constant factor which depends on the desired precision and the ratio
      between the step size and the radius of convergence.
    
      An experimentally better strategy for poor man's analytic continuation
      is to postcompose 
 with a suitable analytic
      function 
 instead of 
. If 
 admits a dominant
      singularity at 
, then 
 should be chosen in such a way that 
      has 
 as its only singularity in the unit disk,
      and such that 
 is as small as possible. One may
      typically take 
 to be a composition of rational
      functions of the form 
. Yet
      another strategy would be to consider a Padé-Hermite
      approximation 
 of 
 and
      then to expand 
 (instead of 
) at 
.
    
      Assuming that we have a means for analytic continuation (whether this
      means is efficient and accurate or only a “poor man's”
      solution), a natural question is to gather information about the
      singularities of 
. First of
      all, we may want to locate the dominant singularity (or singularities)
      of 
 with a good precision.
    
      Let us first assume that 
 admits a unique
      dominant singularity at 
 and no other
      singularities of norm 
.
      Section 3 provides us with heuristic algorithms for the
      computation of 
. More
      generally, for any 
 with 
, the point 
 is also the
      dominant singularity of 
, so
      we may compute 
 using the same heuristic
      algorithms. Computing 
 and 
      for two such points, the singularity 
 is now
      uniquely determined by its distances 
,
      
 and 
 to 
, 
 and 
      (see figure 2).
    
      Even if we do not know 
 beforehand, then we apply
      the above method for 
 with 
, for some 
.
      In order to check the correctness of the computed value of 
 a posteriori, we also compute the radius of
      convergence 
 of 
.
      Then 
 is a dominant singularity of 
 if and only if 
.
      If this check fails, then we simply divide 
 by
      two and repeat the same procedure until the right value of 
 is found.
    
      In practice, the convergence radii can only be approximated, so all
      equality tests have to be replaced by approximate equality tests, for
      which the tolerance in relative precision is slightly larger than the
      estimated relative precision with which the convergence radii are
      computed. In particular, the above method computes 
      with approximately the same accuracy as the radius of convergence 
.
    
      Furthermore, if our procedure for analytic continuation is accurate
      enough, then we may “zoom in” on a singularity, and
      determine its location with higher precision. More precisely, assume
      that we have a rough approximation 
 of the
      dominant singularity 
. For
      some small 
, we then compute
      an approximation 
 of the dominant singularity of
      
 with 
.
      This yields the improved approximation 
 of 
.
    
      Assume now that 
 admits more than one dominant
      singularity (or other singularities of norms close to 
), but still a finite number of them. On the
      one hand, by computing the radii of convergence of 
      for a sufficiently large number of points 
 with
      
, we may determine all
      dominant singularities. On the other hand, for any two adjacent dominant
      singularities 
 and 
 with
      
, we may check the absence of
      other singularities in between by looking at the radii of convergence of
      
 and 
 for some 
. This yields a criterion for
      deciding whether we took a “sufficiently large number of points
      
”.
    
From the numerical point of view, rough initial rough approximations can be further improved by zooming in. In figure 3 we have shown an example where we zoomed in on one among eight singularities. After zooming in, we see that all other singularities are far away, so they do not pollute the computation of the radius of convergence.
The generalized algorithm is particularly useful in the special case when there are two dominant singularities which are complex conjugates. This annoying situation prevents us from using the technique of asymptotic extrapolation in a direct way. Nevertheless, we may use it after zooming in on one of the two singularities.
      In the case when our method for analytic continuation allows us to turn
      around singularities, it is natural to ask whether we can explore the
      Riemann surface of 
 beyond the dominant
      singularities. In general, it is always possible to approximate the
      Riemann surface of 
 by an organically growing
      sequence of more and more precise Riemann surfaces of a special form; we
      refer to [Hoe07] for details. In this section, we will
      restrict ourselves to a more particular situation: given the closed disk
      
 of center 
 and radius
      
, we will assume that there
      exist a finite number of points 
 in its interior
      such that 
 is defined above 
. Our task is to find the points 
.
    
      One complication is that some of the singularities 
      may not directly be visible from the origin by following a straight
      line, but only after turning around one or more other singularities. In
      principle, more and more singularities might therefore be disclosed by
      making successive turns around the singularities which were already
      found. In order to ensure our algorithm to terminate, we will make the
      additional assumption
    
            There exists a finite dimensional vector space of analytic
            functions above 
, which
            contains 
 and all its analytic
            continuations.
          
This is in particular the case if
            The Riemann surface of 
 admits a finite
            number of sheets above 
.
          
In sections 6.1 and 6.2, we will encounter two natural situations in which the assumptions A resp. H are satisfied.
      Consider a finite number of points 
 together with
      a graph 
 which admits 
 as
      its vertices. We will say that 
 is
      admissible if 
 is connected and for
      every edge 
, the straightline
      segment between 
 and 
 lies
      in 
. Assume that 
 is known at one of the points 
. For each 
,
      we may then define 
 to be the vector space
      spanned by all analytic continuations of 
 by
      following edges of 
. We may
      compute the collection 
 of these vector spaces
      using the following algorithm:
    
        Algorithm continue
      
        
        at 
 and an admissible graph 
        with 
      
        
,
        a basis 
 of 
      
Step 1. [Initialize]
        Set 
 and 
 for all 
      
Step 2. [Saturate]
        For any 
, 
        and 
 do
      
        If 
, then
      
        Set 
 and repeat step 2
      
Step 3. [Terminate]
        Return 
      
      Remark 
 requires an equality test for analytic
      functions. From the heuristic point of view, we may simply fix two
      numbers 
 and test whether the first 
 Taylor coefficients coincide up to a relative error 
.
    
      Assume now that 
 singularities in 
 are known, say 
.
      We need a test in order to check whether 
,
      as well as a way to find at least one of the remaining singularities in
      
 if 
.
    
      Let 
 be such that 
 for all
      
 and 
.
      We now consider a graph 
 in the form of a
      hexagonal honeycomb, which fills up the disk 
 and
      such that each edge 
 has length 
      (see also figure 4 below). Picking 
      sufficiently at random, the graph 
 is also
      admissible, so we may apply the algorithm continue. For
      each vertex 
, we also compute
      the minimum 
 of the convergence radii of the
      elements in 
. We claim that
      
 if and only if
    
      for all 
. Indeed, if 
, then the equality clearly holds.
      Assume for contradiction that the equality holds for all 
, even though 
.
      For each of the outermost vertices 
 of the
      honeycomb with 
, we have
      
. This implies that 
, whence 
      lies in the interior of one of the cells of the honeycomb. Let 
 with 
 be such that 
      is minimal. Now consider a vertex 
 of the cell of
      the honeycomb which contains 
 such that 
 (see figure 4). Then 
. Now 
 also implies that
      
. Consequently, 
. This contradiction completes the proof of our
      claim.
    
      Whenever (9) does not hold for some 
, then this indicates the presence of a
      singularity in 
 near 
. It thus suffices to determine the dominant
      singularities of the elements in 
 in order to
      find at least one of these missing singularities.
    
      Remark 
. Therefore, and as in the previous subsection,
      the test (9) should really be replaced by an approximate
      equality test.
    
      Remark 
 edges and not 
 (at least in the case when 
 is
      the solution to a differentially algebraic equation).
    
![]()  | 
        
              Figure 4. Illustration of the proof
              of our claim that   | 
        
      In the case when 
 is meromorphic on the compact
      disk 
 of radius 
,
      then there exists an alternative for the algorithms from section 4: we may directly search for a polynomial 
      such that the radius of convergence of 
 is
      strictly larger than 
. If
      such a polynomial 
 exists, then we may select
      
 to be monic and of minimal degree; this
      particular polynomial 
 will be called the
      denominator of 
 on 
. If 
 is not meromorphic
      on 
, then we define 
. Given 
, we also define the guarded denominator
      
 by 
 if 
      or 
 and 
 otherwise.
      Guarded denominators may be computed using simple linear algebra, as
      follows:
    
        Algorithm denom
      
        
        coefficients of 
, a radius
        
 and a degree bound 
      
        
 with 
,
      
        chosen of minimal degree 
,
        or 
 (failed)
      
Step 1. [Initialize]
        
      
        Step 2. [Determine 
]
      
Solve the linear system
![]()  | 
          (10) | 
        Set 
      
Step 3. [Terminate or loop]
        Heuristically determine 
,
        based on the first 
 coefficients of 
      
        If 
 then return 
      
        If 
 then return 
      
        Set 
 and go to step 2
      
      In order to study the reliability of this algorithm, let us first
      introduce some more notations. First of all, given integers 
 and 
, we will
      denote
    
    The approximate correctness of the algorithm relies on the following technical lemma which will be proved in the next section.
        
 admits exactly 
 roots
        in the closed unit disk, when counted with multiplicities, then the
        matrix norm of 
 satisfies 
        for 
.
      
        
 and assume that the test 
        in the algorithm always returns true if and only if 
        and 
. Then 
        if and only if 
. Moreover,
        if 
, then
      
        
        for 
 and any 
.
      
      Proof. Modulo a scaling 
, we may assume without loss of generality that
      
. Furthermore, the proof is
      clear if 
, so we will assume
      that 
 and 
.
      The test 
 returns false as long as 
 (or if 
).
      Assume now that 
 reaches 
      in the algorithm. Then the computed 
 satisfies
    
    since
    
      Now let 
. Given 
, we have 
.
      Using lemma 5, it follows that 
.
    
      Remark 
-step
      search for finding a polynomial of minimal degree 
      with 
. Using a binary search
      (doubling 
 at each step at a first stage, and
      using a dichotomic search at a second stage), the number of steps can be
      reduced to 
.
    
      Proof. Let us first consider the case when 
 is a rational function with 
, 
, and
      such that 
 only admits roots in the closed unit
      disk. We denote 
 and 
 for
      each 
. Using partial fraction
      decomposition, there exist numbers 
 with 
 such that
    
    In particular, we may factor
    
      where 
 is a matrix with entries in 
 and 
 is a diagonal matrix with
      entries 
 (and such that each 
      occurs 
 times). The functions 
      with 
 being linearly independent, the matrix 
 is invertible in 
,
      whence
    
    This completes the proof of the lemma in the special case.
      In general, we may write 
,
      with 
 as above and where 
      is an analytic function on the closed unit disk. Then
    
    since
    
      for some 
. This completes the
      proof in the general case.
    
      Keeping the same spirit as in section 5, we will now turn
      our attention to two larger classes of algebraic and Fuchsian analytic
      functions on the compact disk 
 of radius 
. Given a function 
      in one of these classes, we will show how to find the defining equation
      for 
, under the assumption
      that we have a high accuracy algorithm for its analytic continuation. As
      in remark 2, we will also assume the existence of a
      heuristic equality test for analytic functions.
    
      Let 
 be an analytic function which is given by
      its power series 
 at the origin. Assume that 
 can be continued analytically on a Riemann surface
      
 above the closed disk 
 of
      radius 
 minus a finite set of points 
. Let 
 be the set of
      analytic functions on 
. We
      say that 
 is algebraic on 
 if there exists a polynomial relation
    
![]()  | 
        (11) | 
      with 
. In that case, we may
      normalize the relation such that 
 has minimal
      degree and such that 
 is a monic polynomial of
      minimal degree. In particular, all roots of 
 are
      inside the disk 
.
    
      Given a function 
 which is algebraic on 
, this function clearly satisfies
      the assumption A in section 4.3, whence we
      may compute the singularities 
 using the
      algorithm described there. Conversely, the example
    
    
      shows that there are functions which satisfy the assumption
      A, but which are not algebraic on the disk 
.
    
      Assume that 
 is algebraic on 
. For each singularity 
, consider a path 
 from the
      origin to a point near 
 which avoids the other
      singularities, and let 
 be the operator which
      performs an analytic continuation along 
,
      one turn around 
, followed by
      an analytic continuation along 
.
      With these operators, we may use the following algorithm for the
      detection of algebraic dependencies:
    
        Algorithm alg_dep
      
        
        above 
 and bounds 
,
        
 and 
      
        
 and 
, or 
 (failed)
      
Step 1. [Initialize]
        Set 
      
Step 2. [Saturate]
        If 
 then return 
      
        If 
 for all 
,
        then go to step 3
      
        
      
Repeat step 2
Step 3. [Terminate]
        Denote 
      
        Compute 
      
        For each 
, compute 
      
        If 
 for some 
,
        then return 
      
        
      
        If 
 then return 
      
        Return 
      
        
      Proof. Assume that 
      satisfies a normalized relation (11), with 
      and 
. Since 
      only contains distinct roots of 
,
      we have 
 in 
 throughout
      the algorithm. In particular 
,
      and we ultimately obtain stabilization 
.
    
      At this point, analytic continuation around any of the points 
 leaves the polynomial 
 invariant,
      so the coefficients of 
 are analytic and
      single-valued on 
. On the
      other hand, given a singularity 
,
      each solution 
 is also given by a convergent
      Puiseux series near 
, whence
      so are the coefficients of 
.
      Since the only Puiseux series without monodromy around 
      are Laurent series, it follows that the coefficients of 
      are meromorphic on 
.
    
      By the minimality assumption on 
,
      it follows that 
 and 
. Since each coefficient 
 is
      meromorphic on 
, we may use
      the algorithm denom from the previous section in order
      to compute a polynomial 
 of minimal degree 
 such that 
.
      The monic least common multiple 
 is nothing but
      the monic polynomial 
 of minimal degree such that
      
.
    
      Remark 
 and 
 always
      return the right answer, then we will reach the line 
      whenever the desired relation exists. For the final normalization step,
      we need a numerical algorithm for the computation of least common
      multiples which allows for small errors in the input coefficients. This
      is similar to the computation of approximate g.c.d.s for
      which there exists an extensive literature; see for example [CGTW95,
      KL96], or [Zen04] and references therein.
    
      Remark 
 and check whether they are indeed
      superior to 
.
    
      We say that 
 is Fuchsian on 
, if 
 satisfies a linear
      differential equation
    
![]()  | 
        (12) | 
      where 
, and 
      is Fuchsian at each point in 
.
      Again, we may normalize (12) such that 
      has minimal order 
 and such that 
      is a monic polynomial of minimal degree.
    
      Given a function 
 which is Fuchsian on 
, the function now satisfies the
      assumption H in section 4.3, so we may
      again compute the singularities 
 using the
      algorithm described there. With 
 as in the
      previous section, we may detect Fuchsian dependencies as follows:
    
        Algorithm fuch_dep
      
        
        above 
 and bounds 
,
        
 and 
      
        
 and 
, or 
 (failed)
      
Step 1. [Initialize]
        Set 
      
Step 2. [Saturate]
        If 
 then return 
      
        Let 
 denote the vector space generated by a
        finite set of vectors 
      
        If 
 for all 
,
        then go to step 3
      
        
 for 
 and 
 with 
      
Repeat step 2
Step 3. [Terminate]
        Denote 
      
        Compute 
      
        in the skew polynomial ring 
,
        where 
 denotes 
      
        For each 
, compute 
      
        If 
 for some 
,
        then return 
      
        
      
        If 
 then return 
      
        Return 
      
        
      Proof. Assume that 
      satisfies a normalized Fuchsian relation (12), with 
 and 
.
      Throughout the algorithm, the set 
 only contains
      linearly independent solutions to 
.
      Therefore, the smallest operator 
 which vanishes
      on 
 divides 
 in 
 on the right. In particular 
,
      and we ultimately obtain stabilization 
.
    
      Consider one of the singularities 
.
      Since 
 is Fuchsian at 
, the equation 
 admits a
      fundamental system of solutions of the form
    
    
      where 
 and 
 are convergent
      power series. The coefficients of 
 lie in the
      field 
 generated by all generalized power series
      of this form. Again, the only elements of 
 with a
      trivial monodromy around 
 are convergent Laurent
      series at 
. Since analytic
      continuation around 
 leaves the operator 
 unchanged, it follows that the coefficients of 
 are meromorphic on 
.
      We conclude in a similar way as in the proof of theorem 8.
    
      Remark 
 admits at worse a Fuchsian singularity at
      
 is essential for the algorithm to work. For
      instance, in the case of the function
    
    
      the monodromy around 
 is trivial, whence applying
      the algorithm for 
 would simply result in having
      
 at step 3. Even though 
      has no monodromy around 
,
      this function is no longer a Laurent series. In fact, the desired
      vanishing operator has second order in this case, but it cannot be read
      off directly from 
.
    
      In the case when we have no algorithm or no efficient algorithm for the
      analytic continuation of 
,
      the algorithms of section 6 can no longer be used. In this
      section, we will describe a purely numerical approach for the detection
      of analytic relations on a closed disk of radius 
. Modulo the change of variables 
, it suffices to consider the case when 
.
    
      Given an order 
, we will
      denote by 
 the set of power series 
 with 
 for all 
. When truncating the usual multiplication on
      
 at order 
,
      we give 
 a ring structure, which is isomorphic to
      
. We will denote by 
 the Hilbert space of all power series 
      with finite 
 norm
    
    
      Notice that any power series that converges on the closed unit disk has
      finite norm, but a power series with finite norm is only guaranteed to
      converge on the open unit disk. More generally, the norm of a vector
      
 is given by
    
    
      The spaces 
 can be considered as an increasing
      sequence of Hilbert spaces, for the restrictions of the norm on 
 to the 
.
    
      Let 
 and assume that 
 are
      power series with radii of convergence at least 
      (notice that we do not assume the 
 to be
      of finite norm). We want to solve the equation
    
![]()  | 
        (13) | 
      We will denote the vector space of all such relations 
      by 
. Since the equation (13) involves an infinite number of coefficients, we need to
      consider its truncated version at a finite order 
. Replacing 
 by their
      truncations in 
, we will
      search for non-trivial solutions of the equation
    
![]()  | 
        (14) | 
      In a way which will be made more precise below, we will require the
      norms of 
 and 
 to remain
      of the same orders of magnitude as the vector 
. We will denote by 
 the
      vector space of all 
 which satisfy (14).
      If the norms of the solutions to the truncated problems tend to a limit
      for 
, then we will prove that
      these solutions tend to a solution of the original problem (13).
    
      Let us reformulate the truncated problem in terms of linear algebra. The
      series 
 give rise to an 
      matrix
    
    
      The unknown series 
 give rise to a row vector
    
    
      Setting 
, we then have
    
    
      Putting the first 
 entries on the right hand side
      and grouping by packets of 
 entries, it follows
      that 
 encodes the relation 
. This reduces the problem to finding those vectors
      
 for which 
 is of the same
      order of magnitude as the vector 
.
    
      We start with the computation of a thin LQ decomposition of 
. This can for instance be done using the
      Gram-Schmidt process: starting with the first row, we orthogonally
      project each row on the vector space spanned by the previous rows. This
      results in a decomposition
    
    
      where 
 is a lower triangular 
      matrix with ones on the diagonal and 
 is an 
 matrix, whose rows are mutually orthogonal
      (i.e. 
). Now
      consider the matrix 
 formed by the 
 last rows of 
.
      Then each row 
 gives rise to a relation (14),
      encoded by 
. Moreover, this
      relation is normal or 
-normal,
      in the sense that 
 and 
      for all 
. Since 
 is the shortest vector in 
,
      the relation is also minimal in norm, among all 
-normal relations. Choosing the row 
      for which 
 is minimal, our algorithm simply
      returns the corresponding relation. Then our algorithm has the following
      fundamental property:
    
        
        for which 
 is minimal.
      
      Let us now return to the case when 
 are no longer
      truncated, but all have a radius of convergence 
. A relation (13) is again said to be
      normal or 
-normal
      if 
 and 
 for all 
. Under the limit 
, we claim that our algorithm finds a minimal
      normal relation, if there exists a relation of the form (13):
    
              Assume that 
. Then
              
 contains a minimal normal relation. For
              each 
, 
 contains at most one minimal 
-normal relation.
            
              Assume that 
 is a minimal 
-normal relation. For each 
, let 
 be the
              truncation of 
 at order 
              and consider the corresponding minimal 
-normal relation 
. Then the relations 
              converge to 
 in 
.
            
              If 
, then the norms
              
, with 
 as in 
.
            
      Proof. A non trivial relation 
      is easily normalized: we first divide 
 by 
, where 
. We next divide by 
,
      where 
 is largest with 
. Now the set of all 
-normal
      relations is a closed affine subspace of 
.
      The orthogonal projection of 
 on this subspace
      yields a unique 
-normal
      relation of minimal norm. This proves (a).
    
      Assume that there exists an 
-normal
      relation 
 be the minimal such relation. Given an order 
, consider the minimal 
-normal relation 
 at this
      order. Truncation of this relation at a smaller order 
      yields an 
-normal relation
      
 at order 
 with 
, whence
    
![]()  | 
        (15) | 
      Moreover, since 
 is the projection of 
 on the affine space of 
-normal
      relations at order 
, we have
      
 and
    
![]()  | 
        (16) | 
      Similarly, truncation of the relation 
 at finite
      order 
 yields an 
-normal
      relation 
 which satisfies
    
    
      In particular, 
, so that
    
    
      For 
, it follows that 
, whence
    
    and
    
      In other words, 
 is a Cauchy sequence, which
      converges to a limit 
 with 
. Now for each 
,
      we may write 
 and denote by 
      the truncation of the vector 
 at order 
. By assumption, the inner products
      
 and 
 vanish. Since 
 for all 
 and 
, we also have 
 for
      
, whence 
. Consequently, 
 is an
      
-normal relation in 
. By the minimality hypothesis on
      
, we conclude that 
, which proves (b).
    
      In general, the existence of a bound 
 with
    
    
      still ensures 
 to be a Cauchy sequence, and its
      limit yields an 
-normal
      relation (13). This proves the last assertion
      (c).
    
      Remark 
 with an isolated smallest singularity
      at 
 of the form
    
    
      where 
 and 
 are analytic
      at 
. Then the asymptotic
      extrapolation of 
 yields an asymptotic expansion
      of the form
    
    
      Using singularity analysis [FS96], we may then recover the
      function 
 from the coefficients 
. However, this technique only works in special
      cases, since the first 
 terms of the asymptotic
      expansion may hide other terms, which need to be taken into account when
      searching for exact dependencies.
    
      A first implementation of the guessing algorithm from section 7
      has been made in the 
.
    
We have tested the orthogonalization algorithm on three examples of increasing difficulty, for which linear dependencies are known to exist. In order to avoid problems due to numerical instability, we have used a large working precision, of 1024 or 2048 bits.
    
      Here follow the values for 
 at different orders
      and 
:
    
    
      The 
 clearly converge to a limit 
      with radius of convergence 
.
      It should be noticed that we do not have 
.
      In other words, the “best” numeric relation (from the point
      of view of 
-norms) does not
      coincide with the “best” algebraic relation (as found by
      using Padé-Hermite approximation). A closer examination of the
      result shows that
    
    
      In particular, 
 decreases if 
      increases.
    
    We obtain:
    
      The convergence only becomes apparent at higher orders. Contrary to the
      previous example, it seems that the series in the computed relation all
      have a radius of convergence 
.
      The norms of the computed results are given by 
, 
, 
 and 
.
    
 [Pól37].
        Its generating series satisfies the functional equation
      
    
      The asymptotic behaviour of the coefficients 
 is
      determined by the asymptotic behaviour of 
 at its
      singularity of smallest norm. Since 
 for all
      
, this so-called dominant
      singularity necessarily lies on the positive real axis, and coincides
      with the radius of convergence 
 of 
. Using asymptotic extrapolation [Hoe09],
      we find that 
. In order to
      investigate 
 for 
 close to
      
, we apply our algorithm to
    
    
      The translation 
 is done using power series
      evaluation until stabilization at the working precision. Although this
      was most straightforward to implement for this particular example, there
      exist better strategies for zooming in, as mentioned at the end of
      section 4.1. At different orders, we obtain:
    
    
      The corresponding norms are given by 
,
      
, 
 and
      
. Again, the convergence is
      rather slow, and the computed series all seem to have radius of
      convergence 
. Since 
, this means that we indeed guessed
      an analytic relation between 
 and 
 on the unit disk.
    
      Remark 
,
      if such a relation indeed exists. Larger radii of convergence 
 can be obtained simply by running the algorithm on 
 and scaling back the result.
    
      It is also instructive to run the algorithm on examples where the 
 are known to be analytically independent. In that
      case, it is important to examine the behaviour of the norms 
 for 
 and find more precise criteria
      which will enable us to discard the existence of a nice dependency with
      a reasonable degree of certainty.
    
    
      where 
. Running our algorithm
      directly on this function yields the following results for 
:
    
    
      The results indeed do not converge and the corresponding sequence of
      norms 
, 
, 
 and 
      diverges at moderate exponential speed.
    
. More generally, we can consider various
        types of singularities, such as
      
    
      The series 
 is a series whose coefficients are
      chosen according to a random uniform distribution on 
. The results are shown in table 1
      below. For 
, 
      and 
, it seems that the norm
      does not much depend on the precise type of singularity, but only on
      
 and the truncation order 
. For the last series 
,
      the dependencies on 
 and 
      approximately seem to follow the law
    
![]()  | 
        (17) | 
      This idealized law needs to be adjusted for functions 
, 
,
      
 with more common types of singularities.
      Although (17) remains asymptotically valid for large values
      of 
, the factor 
 needs to be replaced by a smaller number for moderate
      values. For 
, this factor
      rather seems to be of the form 
,
      where the constant 
 depends on the nature of the
      singularity. Also, the linear dependency of 
 on
      
 is only reached for large values of 
.
    
  | 
        ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||

        as above and studied the dependency of 
 on
        
. The results are shown in
        table 2 below. In the bottom rows, 
        stand for distinct uncorrelated random series 
. In that case, the relation (17)
        generalizes to
      ![]()  | 
        (18) | 
      It also seems that the law can be adapted to functions with more common
      types of singularities, along similar lines as before. It would be
      interesting to have a theoretical explanation for the empirical laws (17) and (18). The division by 
      in (18) is probably due to the fact that there are 
 unknowns in the equation (13).
    
      Of course, in the case that we are considering functions 
 with several singularities 
 on our
      disk of interest, only the singularities 
 for
      which 
 is smallest are “numerically
      visible” (except when there exist relations which not involve
      these dominant singularities 
).
    
In order to design a blackbox algorithm for the detection of dependencies (13), we need to make a final decision on whether we declare the input functions to be dependent or not. In view of the empirical law (18), and whatever decision criterion we implement, the outcome can only be trusted if the following two conditions are satisfied:
          The number 
 should not be too large.
        
          There exists a large constant 
 for which all
          singularities of 
 are concentrated inside the
          disk of radius 
 or outside the disk of radius
          
.
        
As usual, if we are interested in dependencies near an isolated singularity, then the second condition can often be achieved by zooming in on the singularity.
      Our empirical observations in the previous section suggest that we
      should base the final decision on the empirical law (18).
      First of all, in the case of independence, we should observe a more or
      less geometric increase of the norm 
 when running
      the algorithm for different orders 
.
      If 
 only admits an isolated singularity in 
, then we may also compute its norm
      
 and compare the computed values of 
 with the expected values, as given by the law (18),
      or a suitable adaptation of this law when 
      becomes small. This method is most effective when 
      is large, so it is usually recommended to zoom in on the singularity
      before doing this comparison. In fact, if the law (18) is
      satisfied for different zooming factors, then this may further
      increase our confidence that the input functions are independent.
    
      For any numerical checks based on the law (18) or a
      refinement of it, we also recommend to precondition the input series
      
. In particular, we recommend
      to multiply each 
 by a suitable constant,
      ensuring that
    
    
      whenever we apply our algorithm at order 
.
      Here 
 is computed using one of the algorithms
      from section 3.
    
      In fact, there is a trade-off between zooming in on the dominant
      singularity (thereby ensuring a large value of 
) and using the original input coefficients (which
      are given with a higher accuracy and order). In order to get more
      insight in how far we should zoom in, we need to analyze the cost of the
      algorithm from section 7.
    
      The current working precision 
 should in general
      be taken larger than 
 in order to keep the method
      numerically stable. Denoting by 
 the cost for
      multiplying two 
-bit numbers,
      a naive implementation of the Gram-Schmidt orthogonalization procedure
      yields a total cost of 
.
      Denoting by 
 the cost of multiplying two 
 matrices with 
 bit entries,
      and using a blockwise Gram-Schmidt procedure (similar to the procedure
      for matrix inversion in [Str69]), we obtain the better
      bound 
. However, the matrix
      
 from section 7 has a very special
      form. With more work, it might therefore be possible to save an
      additional factor 
, but we
      have not actively tried to do so yet.
    
      Taking into account the effect of a possible zoom, we may next evaluate
      the computational cost in terms of the desired “output
      quality”. As a definition of the output quality, we may take the
      expected value 
 of 
 in the
      case when 
 are independent. In order to obtain
      accurate results, we need to compute with a bit precision larger than
      
. In terms of 
, the time complexity thus becomes
    
    
      where 
 is the exponent of matrix multiplication
      [Str69, Pan84, CW87, Vas12].
      The complexity bound makes it clear that we should put a lot of effort
      into keeping 
 large. If 
      satisfies a differential equation, then zooming in can be done with low
      computational cost (see section 4.1). Otherwise, the
      required expansion order grows quickly with the inverse of the distance
      to the singularity. Indeed, for 
,
      the formula (8) becomes
    
    
      where we recall that, in order to apply the orthogonalization process
      after zooming in, we need to expand 
 at an order
      which is 
 times larger than before.
    
D. Bailey, J.M. Borwein, V. Kapoor, and E.W. Weisstein. Ten problems in experimental mathematics. The American Mathematical Monthly, 113(6):481–509, 2006.
J.M. Borwein and K.J. Devlin. The Computer As Crucible: An Introduction to Experimental Mathematics. A.K. Peters Series. A.K. Peters, 2009.
K. Belabas. Journées Nationales de Calcul Formel (2011), volume 2 of Les cours du CIRM, chapter Théorie algébrique des nombres et calcul formel. CEDRAM, 2011. Exp. No. 1, 40 pages, http://ccirm.cedram.org/ccirm-bin/item?id=CCIRM_2011__2_1_A1_0.
G.D. Birkhoff. Equivalent singular points of ordinary differential equations. Math. Ann., 74:134–139, 1913.
R.P. Brent and H.T. Kung. Fast algorithms for manipulating formal power series. Journal of the ACM, 25:581–595, 1978.
B. Beckermann and G. Labahn. A uniform approach for the fast computation of matrix-type Padé approximants. SIAM J. Matrix Analysis and Applications, 15:804–823, 1994.
C. Brezinski and R. Zaglia. Extrapolation Methods. Theory and Practice. North-Holland, Amsterdam, 1991.
D.V. Chudnovsky and G.V. Chudnovsky. Computer algebra in the service of mathematical physics and number theory (computers in mathematics, stanford, ca, 1986). In Lect. Notes in Pure and Applied Math., volume 125, pages 109–232, New-York, 1990. Dekker.
R. M. Corless, P. M. Gianni, B. M. Trager, and S. M. Watt. The singular value decomposition for polynomial systems. In Proc. ISSAC '95, pages 195–207. ACM Press, 1995.
B. F. Caviness and M. J. Prelle. A note on algebraic independence of logarithmic and exponential constants. SIGSAM Bull., 12/2:18–20, 1978.
              D. Coppersmith and S. Winograd. Matrix multiplication
              via arithmetic progressions. In Proc. of the 
              Annual Symposium on Theory of Computing, pages 1–6,
              New York City, may 25–27 1987.
            
H. Derksen. An algorithm to compute generalized Padé-Hermite forms. Technical Report Rep. 9403, Catholic University Nijmegen, January 1994.
J. Denef and L. Lipshitz. Decision problems for differential equations. The Journ. of Symb. Logic, 54(3):941–950, 1989.
E. Fabry. Sur les intégrales des équations différentielles linéaires à coefficients rationnels. PhD thesis, Paris, 1885.
J.-C. Faugère. Parallelization of Gröbner basis. In Hoon Hong, editor, Parallel and Symbolic Computation, volume 5 of Lect. Notes Series in Computing. Pasco'94, World Scientific, 1994.
Ph. Flajolet and R. Sedgewick. An introduction to the analysis of algorithms. Addison Wesley, Reading, Massachusetts, 1996.
J. von zur Gathen and J. Gerhard. Modern Computer Algebra. Cambridge University Press, 2-nd edition, 2002.
J. van der Hoeven, G. Lecerf, B. Mourrain, et al. Mathemagix, 2002. http://www.mathemagix.org.
J. van der Hoeven. Fast evaluation of holonomic functions. TCS, 210:199–215, 1999.
J. van der Hoeven. Zero-testing, witness conjectures and differential diophantine approximation. Technical Report 2001-62, Prépublications d'Orsay, 2001.
J. van der Hoeven. Relax, but don't be too lazy. JSC, 34:479–542, 2002.
J. van der Hoeven. Effective complex analysis. JSC, 39:433–449, 2005.
J. van der Hoeven. On effective analytic continuation. MCS, 1(1):111–175, 2007.
J. van der Hoeven. On asymptotic extrapolation. JSC, 44(8):1000–1016, 2009.
J. van der Hoeven. Journées Nationales de Calcul Formel (2011), volume 2 of Les cours du CIRM, chapter Calcul analytique. CEDRAM, 2011. Exp. No. 4, 85 pages, http://ccirm.cedram.org/ccirm-bin/fitem?id=CCIRM_2011__2_1_A4_0.
N. K. Karmarkar and Y. N. Lakshman. Approximate polynomial greatest common divisors and nearest singular polynomials. In Y.N. Lakhsman, editor, Proc. ISSAC '96, pages 35–42, Zürich, Switzerland, July 1996. ACM Press.
S. Lang. Transcendental numbers and diophantine approximation. Bull. Amer. Math. Soc., 77/5:635–677, 1971.
A.K. Lenstra, H.W. Lenstra, and L. Lovász. Factoring polynomials with rational coefficients. Math. Ann., 261:515–534, 1982.
R.E. Moore. Interval Analysis. Prentice Hall, Englewood Cliffs, N.J., 1966.
V. Pan. How to multiply matrices faster, volume 179 of Lect. Notes in Math. Springer, 1984.
H. Poincaré. Sur les intégrales irrégulières des équations linéaires. Acta Math., 8:295–344, 1886.
G. Pólya. Kombinatorische Anzahlbestimmungen für Gruppen, Graphen und chemische Verbindungen. Acta Mathematica, 68:145–254, 1937.
W.H. Press, S.A. Teukolsky, W.T. Vetterling, and B.P. Flannery. Numerical recipes, the art of scientific computing. Cambridge University Press, 3rd edition, 2007.
D. Richardson. How to recognise zero. JSC, 24:627–645, 1997.
D. Richardson. Zero tests for constants in simple scientific computation. MCS, 1(1):21–37, 2007.
V. Strassen. Gaussian elimination is not optimal. Numer. Math., 13:352–356, 1969.
A.J. Sommese and C.W. Wampler. The Numerical Solution of Systems of Polynomials Arising in Engineering and Science. World Scientific, 2005.
B. Salvy and P. Zimmermann. Gfun: a Maple package for the manipulation of generating and holonomic functions in one variable. ACM Trans. on Math. Software, 20(2):163–177, 1994.
              V. Vassilevska Williams. Multiplying matrices faster
              than Coppersmith-Winograd. In 
 ACM
              Symp. on Theory of Computing (STOC'12), 2012. To appear.
              Preliminary version available at http://cs.berkeley.edu/~virgi/matrixmult.pdf.
            
J. Verschelde. Homotopy continuation methods for solving polynomial systems. PhD thesis, Katholieke Universiteit Leuven, 1996.
W. Wasow. Asymptotic expansions for ordinary differential equations. Dover, New York, 1967.
K. Weihrauch. Computable analysis. Springer-Verlag, Berlin/Heidelberg, 2000.
E. J. Weniger. Nonlinear sequence transformations: Computational tools for the acceleration of convergence and the summation of divergent series. Technical Report math.CA/0107080, Arxiv, 2001.
H.S. Wilf. Generating functionology. Academic Press, 2nd edition, 1994. http://www.math.upenn.edu/~wilf/DownldGF.html.
Z. Zeng. The approximate gcd of inexact polynomials Part I: a univariate algorithm. http://www.neiu.edu/~zzeng/uvgcd.pdf, 2004.