| """Python implementations of some algorithms for use by longobject.c. |
| The goal is to provide asymptotically faster algorithms that can be |
| used for operations on integers with many digits. In those cases, the |
| performance overhead of the Python implementation is not significant |
| since the asymptotic behavior is what dominates runtime. Functions |
| provided by this module should be considered private and not part of any |
| public API. |
| |
| Note: for ease of maintainability, please prefer clear code and avoid |
| "micro-optimizations". This module will only be imported and used for |
| integers with a huge number of digits. Saving a few microseconds with |
| tricky or non-obvious code is not worth it. For people looking for |
| maximum performance, they should use something like gmpy2.""" |
| |
| import re |
| import decimal |
| try: |
| import _decimal |
| except ImportError: |
| _decimal = None |
| |
| # A number of functions have this form, where `w` is a desired number of |
| # digits in base `base`: |
| # |
| # def inner(...w...): |
| # if w <= LIMIT: |
| # return something |
| # lo = w >> 1 |
| # hi = w - lo |
| # something involving base**lo, inner(...lo...), j, and inner(...hi...) |
| # figure out largest w needed |
| # result = inner(w) |
| # |
| # They all had some on-the-fly scheme to cache `base**lo` results for reuse. |
| # Power is costly. |
| # |
| # This routine aims to compute all amd only the needed powers in advance, as |
| # efficiently as reasonably possible. This isn't trivial, and all the |
| # on-the-fly methods did needless work in many cases. The driving code above |
| # changes to: |
| # |
| # figure out largest w needed |
| # mycache = compute_powers(w, base, LIMIT) |
| # result = inner(w) |
| # |
| # and `mycache[lo]` replaces `base**lo` in the inner function. |
| # |
| # If an algorithm wants the powers of ceiling(w/2) instead of the floor, |
| # pass keyword argument `need_hi=True`. |
| # |
| # While this does give minor speedups (a few percent at best), the |
| # primary intent is to simplify the functions using this, by eliminating |
| # the need for them to craft their own ad-hoc caching schemes. |
| # |
| # See code near end of file for a block of code that can be enabled to |
| # run millions of tests. |
| def compute_powers(w, base, more_than, *, need_hi=False, show=False): |
| seen = set() |
| need = set() |
| ws = {w} |
| while ws: |
| w = ws.pop() # any element is fine to use next |
| if w in seen or w <= more_than: |
| continue |
| seen.add(w) |
| lo = w >> 1 |
| hi = w - lo |
| # only _need_ one here; the other may, or may not, be needed |
| which = hi if need_hi else lo |
| need.add(which) |
| ws.add(which) |
| if lo != hi: |
| ws.add(w - which) |
| |
| # `need` is the set of exponents needed. To compute them all |
| # efficiently, possibly add other exponents to `extra`. The goal is |
| # to ensure that each exponent can be gotten from a smaller one via |
| # multiplying by the base, squaring it, or squaring and then |
| # multiplying by the base. |
| # |
| # If need_hi is False, this is already the case (w can always be |
| # gotten from w >> 1 via one of the squaring strategies). But we do |
| # the work anyway, just in case ;-) |
| # |
| # Note that speed is irrelevant. These loops are working on little |
| # ints (exponents) and go around O(log w) times. The total cost is |
| # insignificant compared to just one of the bigint multiplies. |
| cands = need.copy() |
| extra = set() |
| while cands: |
| w = max(cands) |
| cands.remove(w) |
| lo = w >> 1 |
| if lo > more_than and w-1 not in cands and lo not in cands: |
| extra.add(lo) |
| cands.add(lo) |
| assert need_hi or not extra |
| |
| d = {} |
| for n in sorted(need | extra): |
| lo = n >> 1 |
| hi = n - lo |
| if n-1 in d: |
| if show: |
| print("* base", end="") |
| result = d[n-1] * base # cheap! |
| elif lo in d: |
| # Multiplying a bigint by itself is about twice as fast |
| # in CPython provided it's the same object. |
| if show: |
| print("square", end="") |
| result = d[lo] * d[lo] # same object |
| if hi != lo: |
| if show: |
| print(" * base", end="") |
| assert 2 * lo + 1 == n |
| result *= base |
| else: # rare |
| if show: |
| print("pow", end='') |
| result = base ** n |
| if show: |
| print(" at", n, "needed" if n in need else "extra") |
| d[n] = result |
| |
| assert need <= d.keys() |
| if excess := d.keys() - need: |
| assert need_hi |
| for n in excess: |
| del d[n] |
| return d |
| |
| _unbounded_dec_context = decimal.getcontext().copy() |
| _unbounded_dec_context.prec = decimal.MAX_PREC |
| _unbounded_dec_context.Emax = decimal.MAX_EMAX |
| _unbounded_dec_context.Emin = decimal.MIN_EMIN |
| _unbounded_dec_context.traps[decimal.Inexact] = 1 # sanity check |
| |
| def int_to_decimal(n): |
| """Asymptotically fast conversion of an 'int' to Decimal.""" |
| |
| # Function due to Tim Peters. See GH issue #90716 for details. |
| # https://github.com/python/cpython/issues/90716 |
| # |
| # The implementation in longobject.c of base conversion algorithms |
| # between power-of-2 and non-power-of-2 bases are quadratic time. |
| # This function implements a divide-and-conquer algorithm that is |
| # faster for large numbers. Builds an equal decimal.Decimal in a |
| # "clever" recursive way. If we want a string representation, we |
| # apply str to _that_. |
| |
| from decimal import Decimal as D |
| BITLIM = 200 |
| |
| # Don't bother caching the "lo" mask in this; the time to compute it is |
| # tiny compared to the multiply. |
| def inner(n, w): |
| if w <= BITLIM: |
| return D(n) |
| w2 = w >> 1 |
| hi = n >> w2 |
| lo = n & ((1 << w2) - 1) |
| return inner(lo, w2) + inner(hi, w - w2) * w2pow[w2] |
| |
| with decimal.localcontext(_unbounded_dec_context): |
| nbits = n.bit_length() |
| w2pow = compute_powers(nbits, D(2), BITLIM) |
| if n < 0: |
| negate = True |
| n = -n |
| else: |
| negate = False |
| result = inner(n, nbits) |
| if negate: |
| result = -result |
| return result |
| |
| def int_to_decimal_string(n): |
| """Asymptotically fast conversion of an 'int' to a decimal string.""" |
| w = n.bit_length() |
| if w > 450_000 and _decimal is not None: |
| # It is only usable with the C decimal implementation. |
| # _pydecimal.py calls str() on very large integers, which in its |
| # turn calls int_to_decimal_string(), causing very deep recursion. |
| return str(int_to_decimal(n)) |
| |
| # Fallback algorithm for the case when the C decimal module isn't |
| # available. This algorithm is asymptotically worse than the algorithm |
| # using the decimal module, but better than the quadratic time |
| # implementation in longobject.c. |
| |
| DIGLIM = 1000 |
| def inner(n, w): |
| if w <= DIGLIM: |
| return str(n) |
| w2 = w >> 1 |
| hi, lo = divmod(n, pow10[w2]) |
| return inner(hi, w - w2) + inner(lo, w2).zfill(w2) |
| |
| # The estimation of the number of decimal digits. |
| # There is no harm in small error. If we guess too large, there may |
| # be leading 0's that need to be stripped. If we guess too small, we |
| # may need to call str() recursively for the remaining highest digits, |
| # which can still potentially be a large integer. This is manifested |
| # only if the number has way more than 10**15 digits, that exceeds |
| # the 52-bit physical address limit in both Intel64 and AMD64. |
| w = int(w * 0.3010299956639812 + 1) # log10(2) |
| pow10 = compute_powers(w, 5, DIGLIM) |
| for k, v in pow10.items(): |
| pow10[k] = v << k # 5**k << k == 5**k * 2**k == 10**k |
| if n < 0: |
| n = -n |
| sign = '-' |
| else: |
| sign = '' |
| s = inner(n, w) |
| if s[0] == '0' and n: |
| # If our guess of w is too large, there may be leading 0's that |
| # need to be stripped. |
| s = s.lstrip('0') |
| return sign + s |
| |
| def _str_to_int_inner(s): |
| """Asymptotically fast conversion of a 'str' to an 'int'.""" |
| |
| # Function due to Bjorn Martinsson. See GH issue #90716 for details. |
| # https://github.com/python/cpython/issues/90716 |
| # |
| # The implementation in longobject.c of base conversion algorithms |
| # between power-of-2 and non-power-of-2 bases are quadratic time. |
| # This function implements a divide-and-conquer algorithm making use |
| # of Python's built in big int multiplication. Since Python uses the |
| # Karatsuba algorithm for multiplication, the time complexity |
| # of this function is O(len(s)**1.58). |
| |
| DIGLIM = 2048 |
| |
| def inner(a, b): |
| if b - a <= DIGLIM: |
| return int(s[a:b]) |
| mid = (a + b + 1) >> 1 |
| return (inner(mid, b) |
| + ((inner(a, mid) * w5pow[b - mid]) |
| << (b - mid))) |
| |
| w5pow = compute_powers(len(s), 5, DIGLIM) |
| return inner(0, len(s)) |
| |
| |
| # Asymptotically faster version, using the C decimal module. See |
| # comments at the end of the file. This uses decimal arithmetic to |
| # convert from base 10 to base 256. The latter is just a string of |
| # bytes, which CPython can convert very efficiently to a Python int. |
| |
| # log of 10 to base 256 with best-possible 53-bit precision. Obtained |
| # via: |
| # from mpmath import mp |
| # mp.prec = 1000 |
| # print(float(mp.log(10, 256)).hex()) |
| _LOG_10_BASE_256 = float.fromhex('0x1.a934f0979a371p-2') # about 0.415 |
| |
| # _spread is for internal testing. It maps a key to the number of times |
| # that condition obtained in _dec_str_to_int_inner: |
| # key 0 - quotient guess was right |
| # key 1 - quotient had to be boosted by 1, one time |
| # key 999 - one adjustment wasn't enough, so fell back to divmod |
| from collections import defaultdict |
| _spread = defaultdict(int) |
| del defaultdict |
| |
| def _dec_str_to_int_inner(s, *, GUARD=8): |
| # Yes, BYTELIM is "large". Large enough that CPython will usually |
| # use the Karatsuba _str_to_int_inner to convert the string. This |
| # allowed reducing the cutoff for calling _this_ function from 3.5M |
| # to 2M digits. We could almost certainly do even better by |
| # fine-tuning this and/or using a larger output base than 256. |
| BYTELIM = 100_000 |
| D = decimal.Decimal |
| result = bytearray() |
| # See notes at end of file for discussion of GUARD. |
| assert GUARD > 0 # if 0, `decimal` can blow up - .prec 0 not allowed |
| |
| def inner(n, w): |
| #assert n < D256 ** w # required, but too expensive to check |
| if w <= BYTELIM: |
| # XXX Stefan Pochmann discovered that, for 1024-bit ints, |
| # `int(Decimal)` took 2.5x longer than `int(str(Decimal))`. |
| # Worse, `int(Decimal) is still quadratic-time for much |
| # larger ints. So unless/until all that is repaired, the |
| # seemingly redundant `str(Decimal)` is crucial to speed. |
| result.extend(int(str(n)).to_bytes(w)) # big-endian default |
| return |
| w1 = w >> 1 |
| w2 = w - w1 |
| if 0: |
| # This is maximally clear, but "too slow". `decimal` |
| # division is asymptotically fast, but we have no way to |
| # tell it to reuse the high-precision reciprocal it computes |
| # for pow256[w2], so it has to recompute it over & over & |
| # over again :-( |
| hi, lo = divmod(n, pow256[w2][0]) |
| else: |
| p256, recip = pow256[w2] |
| # The integer part will have a number of digits about equal |
| # to the difference between the log10s of `n` and `pow256` |
| # (which, since these are integers, is roughly approximated |
| # by `.adjusted()`). That's the working precision we need, |
| ctx.prec = max(n.adjusted() - p256.adjusted(), 0) + GUARD |
| hi = +n * +recip # unary `+` chops back to ctx.prec digits |
| ctx.prec = decimal.MAX_PREC |
| hi = hi.to_integral_value() # lose the fractional digits |
| lo = n - hi * p256 |
| # Because we've been uniformly rounding down, `hi` is a |
| # lower bound on the correct quotient. |
| assert lo >= 0 |
| # Adjust quotient up if needed. It usually isn't. In random |
| # testing on inputs through 5 billion digit strings, the |
| # test triggered once in about 200 thousand tries. |
| count = 0 |
| if lo >= p256: |
| count = 1 |
| lo -= p256 |
| hi += 1 |
| if lo >= p256: |
| # Complete correction via an exact computation. I |
| # believe it's not possible to get here provided |
| # GUARD >= 3. It's tested by reducing GUARD below |
| # that. |
| count = 999 |
| hi2, lo = divmod(lo, p256) |
| hi += hi2 |
| _spread[count] += 1 |
| # The assert should always succeed, but way too slow to keep |
| # enabled. |
| #assert hi, lo == divmod(n, pow256[w2][0]) |
| inner(hi, w1) |
| del hi # at top levels, can free a lot of RAM "early" |
| inner(lo, w2) |
| |
| # How many base 256 digits are needed?. Mathematically, exactly |
| # floor(log256(int(s))) + 1. There is no cheap way to compute this. |
| # But we can get an upper bound, and that's necessary for our error |
| # analysis to make sense. int(s) < 10**len(s), so the log needed is |
| # < log256(10**len(s)) = len(s) * log256(10). However, using |
| # finite-precision floating point for this, it's possible that the |
| # computed value is a little less than the true value. If the true |
| # value is at - or a little higher than - an integer, we can get an |
| # off-by-1 error too low. So we add 2 instead of 1 if chopping lost |
| # a fraction > 0.9. |
| |
| # The "WASI" test platform can complain about `len(s)` if it's too |
| # large to fit in its idea of "an index-sized integer". |
| lenS = s.__len__() |
| log_ub = lenS * _LOG_10_BASE_256 |
| log_ub_as_int = int(log_ub) |
| w = log_ub_as_int + 1 + (log_ub - log_ub_as_int > 0.9) |
| # And what if we've plain exhausted the limits of HW floats? We |
| # could compute the log to any desired precision using `decimal`, |
| # but it's not plausible that anyone will pass a string requiring |
| # trillions of bytes (unless they're just trying to "break things"). |
| if w.bit_length() >= 46: |
| # "Only" had < 53 - 46 = 7 bits to spare in IEEE-754 double. |
| raise ValueError(f"cannot convert string of len {lenS} to int") |
| with decimal.localcontext(_unbounded_dec_context) as ctx: |
| D256 = D(256) |
| pow256 = compute_powers(w, D256, BYTELIM, need_hi=True) |
| rpow256 = compute_powers(w, 1 / D256, BYTELIM, need_hi=True) |
| # We're going to do inexact, chopped arithmetic, multiplying by |
| # an approximation to the reciprocal of 256**i. We chop to get a |
| # lower bound on the true integer quotient. Our approximation is |
| # a lower bound, the multiplication is chopped too, and |
| # to_integral_value() is also chopped. |
| ctx.traps[decimal.Inexact] = 0 |
| ctx.rounding = decimal.ROUND_DOWN |
| for k, v in pow256.items(): |
| # No need to save much more precision in the reciprocal than |
| # the power of 256 has, plus some guard digits to absorb |
| # most relevant rounding errors. This is highly significant: |
| # 1/2**i has the same number of significant decimal digits |
| # as 5**i, generally over twice the number in 2**i, |
| ctx.prec = v.adjusted() + GUARD + 1 |
| # The unary "+" chops the reciprocal back to that precision. |
| pow256[k] = v, +rpow256[k] |
| del rpow256 # exact reciprocals no longer needed |
| ctx.prec = decimal.MAX_PREC |
| inner(D(s), w) |
| return int.from_bytes(result) |
| |
| def int_from_string(s): |
| """Asymptotically fast version of PyLong_FromString(), conversion |
| of a string of decimal digits into an 'int'.""" |
| # PyLong_FromString() has already removed leading +/-, checked for invalid |
| # use of underscore characters, checked that string consists of only digits |
| # and underscores, and stripped leading whitespace. The input can still |
| # contain underscores and have trailing whitespace. |
| s = s.rstrip().replace('_', '') |
| func = _str_to_int_inner |
| if len(s) >= 2_000_000 and _decimal is not None: |
| func = _dec_str_to_int_inner |
| return func(s) |
| |
| def str_to_int(s): |
| """Asymptotically fast version of decimal string to 'int' conversion.""" |
| # FIXME: this doesn't support the full syntax that int() supports. |
| m = re.match(r'\s*([+-]?)([0-9_]+)\s*', s) |
| if not m: |
| raise ValueError('invalid literal for int() with base 10') |
| v = int_from_string(m.group(2)) |
| if m.group(1) == '-': |
| v = -v |
| return v |
| |
| |
| # Fast integer division, based on code from Mark Dickinson, fast_div.py |
| # GH-47701. Additional refinements and optimizations by Bjorn Martinsson. The |
| # algorithm is due to Burnikel and Ziegler, in their paper "Fast Recursive |
| # Division". |
| |
| _DIV_LIMIT = 4000 |
| |
| |
| def _div2n1n(a, b, n): |
| """Divide a 2n-bit nonnegative integer a by an n-bit positive integer |
| b, using a recursive divide-and-conquer algorithm. |
| |
| Inputs: |
| n is a positive integer |
| b is a positive integer with exactly n bits |
| a is a nonnegative integer such that a < 2**n * b |
| |
| Output: |
| (q, r) such that a = b*q+r and 0 <= r < b. |
| |
| """ |
| if a.bit_length() - n <= _DIV_LIMIT: |
| return divmod(a, b) |
| pad = n & 1 |
| if pad: |
| a <<= 1 |
| b <<= 1 |
| n += 1 |
| half_n = n >> 1 |
| mask = (1 << half_n) - 1 |
| b1, b2 = b >> half_n, b & mask |
| q1, r = _div3n2n(a >> n, (a >> half_n) & mask, b, b1, b2, half_n) |
| q2, r = _div3n2n(r, a & mask, b, b1, b2, half_n) |
| if pad: |
| r >>= 1 |
| return q1 << half_n | q2, r |
| |
| |
| def _div3n2n(a12, a3, b, b1, b2, n): |
| """Helper function for _div2n1n; not intended to be called directly.""" |
| if a12 >> n == b1: |
| q, r = (1 << n) - 1, a12 - (b1 << n) + b1 |
| else: |
| q, r = _div2n1n(a12, b1, n) |
| r = (r << n | a3) - q * b2 |
| while r < 0: |
| q -= 1 |
| r += b |
| return q, r |
| |
| |
| def _int2digits(a, n): |
| """Decompose non-negative int a into base 2**n |
| |
| Input: |
| a is a non-negative integer |
| |
| Output: |
| List of the digits of a in base 2**n in little-endian order, |
| meaning the most significant digit is last. The most |
| significant digit is guaranteed to be non-zero. |
| If a is 0 then the output is an empty list. |
| |
| """ |
| a_digits = [0] * ((a.bit_length() + n - 1) // n) |
| |
| def inner(x, L, R): |
| if L + 1 == R: |
| a_digits[L] = x |
| return |
| mid = (L + R) >> 1 |
| shift = (mid - L) * n |
| upper = x >> shift |
| lower = x ^ (upper << shift) |
| inner(lower, L, mid) |
| inner(upper, mid, R) |
| |
| if a: |
| inner(a, 0, len(a_digits)) |
| return a_digits |
| |
| |
| def _digits2int(digits, n): |
| """Combine base-2**n digits into an int. This function is the |
| inverse of `_int2digits`. For more details, see _int2digits. |
| """ |
| |
| def inner(L, R): |
| if L + 1 == R: |
| return digits[L] |
| mid = (L + R) >> 1 |
| shift = (mid - L) * n |
| return (inner(mid, R) << shift) + inner(L, mid) |
| |
| return inner(0, len(digits)) if digits else 0 |
| |
| |
| def _divmod_pos(a, b): |
| """Divide a non-negative integer a by a positive integer b, giving |
| quotient and remainder.""" |
| # Use grade-school algorithm in base 2**n, n = nbits(b) |
| n = b.bit_length() |
| a_digits = _int2digits(a, n) |
| |
| r = 0 |
| q_digits = [] |
| for a_digit in reversed(a_digits): |
| q_digit, r = _div2n1n((r << n) + a_digit, b, n) |
| q_digits.append(q_digit) |
| q_digits.reverse() |
| q = _digits2int(q_digits, n) |
| return q, r |
| |
| |
| def int_divmod(a, b): |
| """Asymptotically fast replacement for divmod, for 'int'. |
| Its time complexity is O(n**1.58), where n = #bits(a) + #bits(b). |
| """ |
| if b == 0: |
| raise ZeroDivisionError('division by zero') |
| elif b < 0: |
| q, r = int_divmod(-a, -b) |
| return q, -r |
| elif a < 0: |
| q, r = int_divmod(~a, b) |
| return ~q, b + ~r |
| else: |
| return _divmod_pos(a, b) |
| |
| |
| # Notes on _dec_str_to_int_inner: |
| # |
| # Stefan Pochmann worked up a str->int function that used the decimal |
| # module to, in effect, convert from base 10 to base 256. This is |
| # "unnatural", in that it requires multiplying and dividing by large |
| # powers of 2, which `decimal` isn't naturally suited to. But |
| # `decimal`'s `*` and `/` are asymptotically superior to CPython's, so |
| # at _some_ point it could be expected to win. |
| # |
| # Alas, the crossover point was too high to be of much real interest. I |
| # (Tim) then worked on ways to replace its division with multiplication |
| # by a cached reciprocal approximation instead, fixing up errors |
| # afterwards. This reduced the crossover point significantly, |
| # |
| # I revisited the code, and found ways to improve and simplify it. The |
| # crossover point is at about 3.4 million digits now. |
| # |
| # About .adjusted() |
| # ----------------- |
| # Restrict to Decimal values x > 0. We don't use negative numbers in the |
| # code, and I don't want to have to keep typing, e.g., "absolute value". |
| # |
| # For convenience, I'll use `x.a` to mean `x.adjusted()`. x.a doesn't |
| # look at the digits of x, but instead returns an integer giving x's |
| # order of magnitude. These are equivalent: |
| # |
| # - x.a is the power-of-10 exponent of x's most significant digit. |
| # - x.a = the infinitely precise floor(log10(x)) |
| # - x can be written in this form, where f is a real with 1 <= f < 10: |
| # x = f * 10**x.a |
| # |
| # Observation; if x is an integer, len(str(x)) = x.a + 1. |
| # |
| # Lemma 1: (x * y).a = x.a + y.a, or one larger |
| # |
| # Proof: Write x = f * 10**x.a and y = g * 10**y.a, where f and g are in |
| # [1, 10). Then x*y = f*g * 10**(x.a + y.a), where 1 <= f*g < 100. If |
| # f*g < 10, (x*y).a is x.a+y.a. Else divide f*g by 10 to bring it back |
| # into [1, 10], and add 1 to the exponent to compensate. Then (x*y).a is |
| # x.a+y.a+1. |
| # |
| # Lemma 2: ceiling(log10(x/y)) <= x.a - y.a + 1 |
| # |
| # Proof: Express x and y as in Lemma 1. Then x/y = f/g * 10**(x.a - |
| # y.a), where 1/10 < f/g < 10. If 1 <= f/g, (x/y).a is x.a-y.a. Else |
| # multiply f/g by 10 to bring it back into [1, 10], and subtract 1 from |
| # the exponent to compensate. Then (x/y).a is x.a-y.a-1. So the largest |
| # (x/y).a can be is x.a-y.a. Since that's the floor of log10(x/y). the |
| # ceiling is at most 1 larger (with equality iff f/g = 1 exactly). |
| # |
| # GUARD digits |
| # ------------ |
| # We only want the integer part of divisions, so don't need to build |
| # the full multiplication tree. But using _just_ the number of |
| # digits expected in the integer part ignores too much. What's left |
| # out can have a very significant effect on the quotient. So we use |
| # GUARD additional digits. |
| # |
| # The default 8 is more than enough so no more than 1 correction step |
| # was ever needed for all inputs tried through 2.5 billion digits. In |
| # fact, I believe 3 guard digits are always enough - but the proof is |
| # very involved, so better safe than sorry. |
| # |
| # Short course: |
| # |
| # If prec is the decimal precision in effect, and we're rounding down, |
| # the result of an operation is exactly equal to the infinitely precise |
| # result times 1-e for some real e with 0 <= e < 10**(1-prec). In |
| # |
| # ctx.prec = max(n.adjusted() - p256.adjusted(), 0) + GUARD |
| # hi = +n * +recip # unary `+` chops to ctx.prec digits |
| # |
| # we have 3 visible chopped operations, but there's also a 4th: |
| # precomputing a truncated `recip` as part of setup. |
| # |
| # So the computed product is exactly equal to the true product times |
| # (1-e1)*(1-e2)*(1-e3)*(1-e4); since the e's are all very small, an |
| # excellent approximation to the second factor is 1-(e1+e2+e3+e4) (the |
| # 2nd and higher order terms in the expanded product are too tiny to |
| # matter). If they're all as large as possible, that's |
| # |
| # 1 - 4*10**(1-prec). This, BTW, is all bog-standard FP error analysis. |
| # |
| # That implies the computed product is within 1 of the true product |
| # provided prec >= log10(true_product) + 1.602. |
| # |
| # Here are telegraphic details, rephrasing the initial condition in |
| # equivalent ways, step by step: |
| # |
| # prod - prod * (1 - 4*10**(1-prec)) <= 1 |
| # prod - prod + prod * 4*10**(1-prec)) <= 1 |
| # prod * 4*10**(1-prec)) <= 1 |
| # 10**(log10(prod)) * 4*10**(1-prec)) <= 1 |
| # 4*10**(1-prec+log10(prod))) <= 1 |
| # 10**(1-prec+log10(prod))) <= 1/4 |
| # 1-prec+log10(prod) <= log10(1/4) = -0.602 |
| # -prec <= -1.602 - log10(prod) |
| # prec >= log10(prod) + 1.602 |
| # |
| # The true product is the same as the true ratio n/p256. By Lemma 2 |
| # above, n.a - p256.a + 1 is an upper bound on the ceiling of |
| # log10(prod). Then 2 is the ceiling of 1.602. so n.a - p256.a + 3 is an |
| # upper bound on the right hand side of the inequality. Any prec >= that |
| # will work. |
| # |
| # But since this is just a sketch of a proof ;-), the code uses the |
| # empirically tested 8 instead of 3. 5 digits more or less makes no |
| # practical difference to speed - these ints are huge. And while |
| # increasing GUARD above 3 may not be necessary, every increase cuts the |
| # percentage of cases that need a correction at all. |
| # |
| # On Computing Reciprocals |
| # ------------------------ |
| # In general, the exact reciprocals we compute have over twice as many |
| # significant digits as needed. 1/256**i has the same number of |
| # significant decimal digits as 5**i. It's a significant waste of RAM |
| # to store all those unneeded digits. |
| # |
| # So we cut exact reciprocals back to the least precision that can |
| # be needed so that the error analysis above is valid, |
| # |
| # [Note: turns out it's very significantly faster to do it this way than |
| # to compute 1 / 256**i directly to the desired precision, because the |
| # power method doesn't require division. It's also faster than computing |
| # (1/256)**i directly to the desired precision - no material division |
| # there, but `compute_powers()` is much smarter about _how_ to compute |
| # all the powers needed than repeated applications of `**` - that |
| # function invokes `**` for at most the few smallest powers needed.] |
| # |
| # The hard part is that chopping back to a shorter width occurs |
| # _outside_ of `inner`. We can't know then what `prec` `inner()` will |
| # need. We have to pick, for each value of `w2`, the largest possible |
| # value `prec` can become when `inner()` is working on `w2`. |
| # |
| # This is the `prec` inner() uses: |
| # max(n.a - p256.a, 0) + GUARD |
| # and what setup uses (renaming its `v` to `p256` - same thing): |
| # p256.a + GUARD + 1 |
| # |
| # We need that the second is always at least as large as the first, |
| # which is the same as requiring |
| # |
| # n.a - 2 * p256.a <= 1 |
| # |
| # What's the largest n can be? n < 255**w = 256**(w2 + (w - w2)). The |
| # worst case in this context is when w ix even. and then w = 2*w2, so |
| # n < 256**(2*w2) = (256**w2)**2 = p256**2. By Lemma 1, then, n.a |
| # is at most p256.a + p256.a + 1. |
| # |
| # So the most n.a - 2 * p256.a can be is |
| # p256.a + p256.a + 1 - 2 * p256.a = 1. QED |
| # |
| # Note: an earlier version of the code split on floor(e/2) instead of on |
| # the ceiling. The worst case then is odd `w`, and a more involved proof |
| # was needed to show that adding 4 (instead of 1) may be necessary. |
| # Basically because, in that case, n may be up to 256 times larger than |
| # p256**2. Curiously enough, by splitting on the ceiling instead, |
| # nothing in any proof here actually depends on the output base (256). |
| |
| # Enable for brute-force testing of compute_powers(). This takes about a |
| # minute, because it tries millions of cases. |
| if 0: |
| def consumer(w, limit, need_hi): |
| seen = set() |
| need = set() |
| def inner(w): |
| if w <= limit: |
| return |
| if w in seen: |
| return |
| seen.add(w) |
| lo = w >> 1 |
| hi = w - lo |
| need.add(hi if need_hi else lo) |
| inner(lo) |
| inner(hi) |
| inner(w) |
| exp = compute_powers(w, 1, limit, need_hi=need_hi) |
| assert exp.keys() == need |
| |
| from itertools import chain |
| for need_hi in (False, True): |
| for limit in (0, 1, 10, 100, 1_000, 10_000, 100_000): |
| for w in chain(range(1, 100_000), |
| (10**i for i in range(5, 30))): |
| consumer(w, limit, need_hi) |