Actually, I lie, slightly. I did find one instance of `numerical
stability' used in reference to the FFT, and that is on wikipedia (so
now we all know it must be true). There I presume the context is
signal processing rather than polynomial multiplication. I did track
down one article which specifically states that the Rader-Brennen FFT
is numerically unstable and it gave a reason why (the size of the
twiddle factors used), though I don't recall a definition of what they
meant by numerical stability. Again, I assume this concept is well-
understood by the signal processing people. How that concept actually
applies to polynomial multiplication I don't know.

What I imagine is the correct way of looking at the problem is via
pontrjagin duality, where the group algebras of Z/nZ and its dual are
isomorphic via the fourier transform. I suppose that polynomials
modulo x^n - 1 are complex valued functions on Z/nZ and that the
numerical stability of FFT's is examined in that context. Even as I
say that though, I am uncertain how inexact representations of R are
handled under this interpretation....

I'm sure it is an interesting problem. Just not one I have much time
to study theoretically at this time.

Bill.

On Apr 30, 9:57 am, Bill Hart <goodwillh...@googlemail.com> wrote:
> On Apr 30, 6:58 am, rjf <fate...@gmail.com> wrote:
>
>
>
>
>
> > On Apr 29, 10:58 am, Robert Bradshaw <rober...@math.washington.edu>
> > wrote:
>
> > > On Apr 29, 2010, at 8:30 AM, rjf wrote:
>
> > > > (RJF)Again, I see no definition of what you mean by accuracy in the 
> > > > result
> > > > of polynomial multiplication.The easiest position to take is that of 
> > > > MPFR--
> > > considering the inputs as exact rationals, how far off is the output  
> > > (say, coefficient by coefficient) from the actual result. One could  
> > > also use other measures, such as an L^2 norm over some compact region  
> > > (like [-1,1]). What may make sense for many applications is some kind  
> > > of a "slopped absolute" precision, as has been discussed with p-adic  
> > > polynomials. These would have good behavior on |x| < a or |x| > a  
> > > depending on the slope.
>
> > There is a large literature on the meaning of the "approximate GCD" of
> > two polynomials with floating-point coefficients.  The term
> > "stability"
> > also has a pretty good definition, and it does not seem to correspond
> > to how you are using it here.
>
> Approximate GCD? That's a curious concept. What is it used for? I
> can't imagine defining a GCD in this context as divisibility is an
> exact phenomenon.
>
> I hear the term numerical stability used quite a lot. The two contexts
> I've encountered it are in linear algebra and in differential
> equations. Actually, I might have heard it used in polynomial
> evaluation too.
>
> I happily admit that I've no idea of a definition in the case of
> polynomial multiplication. However, regardless of what the correct
> terminology is, the problem I am studying is precisely the one stated,
> which I took to be the subject of the thread.
>
> I'm not interested in numerical stability in any of the other senses
> that I know of, at this point. I don't even claim that what I am
> studying is a useful notion, But feel free to suggest some references
> if there is something I ought to read. Oh wait, every single published
> article out there is nonsense. Damn, I forgot. You don't recommend any
> articles.
>
> Bill.
>
>
>
>
>
>
>
> > > It all really depends on your application, and currently we don't  
> > > define, let alone make guarantees, on the precision of the result (and  
> > > nor does Maxima as far as I understand).
>
> > Maxima was designed around exact arithmetic, and generally offers to
> > convert floats to their corresponding exact rationals before doing
> > anything
> > that requires arithmetic. It makes no claims about floats per se,
> > partly
> > because saying anything sensible is a challenge, compounded by
> > particular issues that are not mathematical but practical .. overflow,
> > different machine floating point formats, etc
>
> > > We do, however, try to avoid  
> > > algorithms that are known to be numerically unstable.
>
> > The lack of serious understanding of numerical computation in Sage is
> > unfortunate. It might be cured by someone taking an interest in the
> > topic
> > and taking a course or two.
>
> > > Fortunately for us, many asymptotically fast algorithms do have  
> > > cutoffs well within the size ranges we regularly encounter, and low  
> > > enough that it makes a significant difference in the runtime.
>
> > There are a few, but not many compared to the huge number that you
> > will encounter
> > if you read certain theoretical journals.
>
> > If you care about specifics of the failure of theoretical computer
> > scientists
> > to address computation, write to me privately.
>
> > RJF
>
> > --
> > To post to this group, send an email to sage-devel@googlegroups.com
> > To unsubscribe from this group, send an email to 
> > sage-devel+unsubscr...@googlegroups.com
> > For more options, visit this group 
> > athttp://groups.google.com/group/sage-devel
> > URL:http://www.sagemath.org
>
> --
> To post to this group, send an email to sage-devel@googlegroups.com
> To unsubscribe from this group, send an email to 
> sage-devel+unsubscr...@googlegroups.com
> For more options, visit this group athttp://groups.google.com/group/sage-devel
> URL:http://www.sagemath.org

-- 
To post to this group, send an email to sage-devel@googlegroups.com
To unsubscribe from this group, send an email to 
sage-devel+unsubscr...@googlegroups.com
For more options, visit this group at http://groups.google.com/group/sage-devel
URL: http://www.sagemath.org

Reply via email to