On Thu, Mar 15, 2012 at 6:53 AM, Daniel Krenn <kr...@aon.at> wrote: > How do I change the default precision used? E.g. I want to enter {{{a > = 1.2}}} and want that a is an element of RealField(100) without > explicitly telling to use that field each time. > > What does not work is: > sage: RR = RealField(100) > sage: a = 1.2 > sage: a.parent() > Real Field with 53 bits of precision
You could do the following to force any floating point number entered to have exactly 100 bits of precision by default: sage: RealNumber = lambda x: RealField(100)(x) sage: 2.5 2.5000000000000000000000000000 sage: 4.2992038490283409823094820938492834082093482834 4.2992038490283409823094820939 Or, you could even force all the floating point numbers that you enter to be intervals with 100 bits of precision: sage: RealNumber = lambda x: RealIntervalField(100)(x) sage: 2.5 2.5000000000000000000000000000000? sage: sin(2.5) 0.598472144103956494051854702186? William -- To post to this group, send email to sage-support@googlegroups.com To unsubscribe from this group, send email to sage-support+unsubscr...@googlegroups.com For more options, visit this group at http://groups.google.com/group/sage-support URL: http://www.sagemath.org