Fastest semiprime factorization

28

17

Write a program to factorize a semi-prime number in the shortest amount of time.

For testing purposes, use this: 38!+1 (523022617466601111760007224100074291200000001)

It is equal to: 14029308060317546154181 × 37280713718589679646221

Soham Chowdhury

Posted 2012-10-07T05:16:00.663

Reputation: 1 449

2While I like the "fastest" bit, since it gives languages like C the advantage over typical codegolfing languages, I do wonder how you will test the results? – Mr Lister – 2012-10-07T06:37:14.003

1If you mean that 12259243 will be used to test how fast the programs are, the results will be so small that you won't get any statistically significant differences. – Peter Taylor – 2012-10-07T06:55:27.380

I added a larger number, thx for the heads up. – Soham Chowdhury – 2012-10-07T17:28:42.727

@Mr Lister, I'll test it on my own PC. – Soham Chowdhury – 2012-10-07T17:29:25.283

5inb4 someone uses preprocessor abuse to write a 400 exabyte lookup table. – Wug – 2012-10-09T19:47:26.613

Answers

59

Python (w/ PyPy JIT v1.9) ~1.9s

Using a Multiple Polynomial Quadratic Sieve. I took this to be a code challenge, so I opted not to use any external libraries (other than the standard log function, I suppose). When timing, the PyPy JIT should be used, as it results in timings 4-5 times faster than that of cPython.

Update (2013-07-29):
Since originally posting, I've made several minor, but significant changes which increase the overall speed by a factor of about 2.5x.

Update (2014-08-27):
As this post is still receiving attention, I've updated my_math.py correcting two errors, for anyone who may be using it:

  • isqrt was faulty, sometimes producing incorrect output for values very close to a perfect square. This has been corrected, and the performance increased by using a much better seed.
  • is_prime has been updated. My previous attempt to remove perfect square 2-sprps was half-hearted, at best. I've added a 3-sprp check - a technique used by Mathmatica - to ensure that the tested value is square-free.

Update (2014-11-24):
If at the end of the calculation no non-trivial congruencies are found, the progam now sieves additional polynomials. This was previously marked in the code as TODO.


mpqs.py

from my_math import *
from math import log
from time import clock
from argparse import ArgumentParser

# Multiple Polynomial Quadratic Sieve
def mpqs(n, verbose=False):
  if verbose:
    time1 = clock()

  root_n = isqrt(n)
  root_2n = isqrt(n+n)

  # formula chosen by experimentation
  # seems to be close to optimal for n < 10^50
  bound = int(5 * log(n, 10)**2)

  prime = []
  mod_root = []
  log_p = []
  num_prime = 0

  # find a number of small primes for which n is a quadratic residue
  p = 2
  while p < bound or num_prime < 3:

    # legendre (n|p) is only defined for odd p
    if p > 2:
      leg = legendre(n, p)
    else:
      leg = n & 1

    if leg == 1:
      prime += [p]
      mod_root += [int(mod_sqrt(n, p))]
      log_p += [log(p, 10)]
      num_prime += 1
    elif leg == 0:
      if verbose:
        print 'trial division found factors:'
        print p, 'x', n/p
      return p

    p = next_prime(p)

  # size of the sieve
  x_max = len(prime)*60

  # maximum value on the sieved range
  m_val = (x_max * root_2n) >> 1

  # fudging the threshold down a bit makes it easier to find powers of primes as factors
  # as well as partial-partial relationships, but it also makes the smoothness check slower.
  # there's a happy medium somewhere, depending on how efficient the smoothness check is
  thresh = log(m_val, 10) * 0.735

  # skip small primes. they contribute very little to the log sum
  # and add a lot of unnecessary entries to the table
  # instead, fudge the threshold down a bit, assuming ~1/4 of them pass
  min_prime = int(thresh*3)
  fudge = sum(log_p[i] for i,p in enumerate(prime) if p < min_prime)/4
  thresh -= fudge

  if verbose:
    print 'smoothness bound:', bound
    print 'sieve size:', x_max
    print 'log threshold:', thresh
    print 'skipping primes less than:', min_prime

  smooth = []
  used_prime = set()
  partial = {}
  num_smooth = 0
  num_used_prime = 0
  num_partial = 0
  num_poly = 0
  root_A = isqrt(root_2n / x_max)

  if verbose:
    print 'sieving for smooths...'
  while True:
    # find an integer value A such that:
    # A is =~ sqrt(2*n) / x_max
    # A is a perfect square
    # sqrt(A) is prime, and n is a quadratic residue mod sqrt(A)
    while True:
      root_A = next_prime(root_A)
      leg = legendre(n, root_A)
      if leg == 1:
        break
      elif leg == 0:
        if verbose:
          print 'dumb luck found factors:'
          print root_A, 'x', n/root_A
        return root_A

    A = root_A * root_A

    # solve for an adequate B
    # B*B is a quadratic residue mod n, such that B*B-A*C = n
    # this is unsolvable if n is not a quadratic residue mod sqrt(A)
    b = mod_sqrt(n, root_A)
    B = (b + (n - b*b) * mod_inv(b + b, root_A))%A

    # B*B-A*C = n <=> C = (B*B-n)/A
    C = (B*B - n) / A

    num_poly += 1

    # sieve for prime factors
    sums = [0.0]*(2*x_max)
    i = 0
    for p in prime:
      if p < min_prime:
        i += 1
        continue
      logp = log_p[i]

      inv_A = mod_inv(A, p)
      # modular root of the quadratic
      a = int(((mod_root[i] - B) * inv_A)%p)
      b = int(((p - mod_root[i] - B) * inv_A)%p)

      k = 0
      while k < x_max:
        if k+a < x_max:
          sums[k+a] += logp
        if k+b < x_max:
          sums[k+b] += logp
        if k:
          sums[k-a+x_max] += logp
          sums[k-b+x_max] += logp

        k += p
      i += 1

    # check for smooths
    i = 0
    for v in sums:
      if v > thresh:
        x = x_max-i if i > x_max else i
        vec = set()
        sqr = []
        # because B*B-n = A*C
        # (A*x+B)^2 - n = A*A*x*x+2*A*B*x + B*B - n
        #               = A*(A*x*x+2*B*x+C)
        # gives the congruency
        # (A*x+B)^2 = A*(A*x*x+2*B*x+C) (mod n)
        # because A is chosen to be square, it doesn't need to be sieved
        val = sieve_val = A*x*x + 2*B*x + C

        if sieve_val < 0:
          vec = set([-1])
          sieve_val = -sieve_val

        for p in prime:
          while sieve_val%p == 0:
            if p in vec:
              # keep track of perfect square factors
              # to avoid taking the sqrt of a gigantic number at the end
              sqr += [p]
            vec ^= set([p])
            sieve_val = int(sieve_val / p)

        if sieve_val == 1:
          # smooth
          smooth += [(vec, (sqr, (A*x+B), root_A))]
          used_prime |= vec
        elif sieve_val in partial:
          # combine two partials to make a (xor) smooth
          # that is, every prime factor with an odd power is in our factor base
          pair_vec, pair_vals = partial[sieve_val]
          sqr += list(vec & pair_vec) + [sieve_val]
          vec ^= pair_vec
          smooth += [(vec, (sqr + pair_vals[0], (A*x+B)*pair_vals[1], root_A*pair_vals[2]))]
          used_prime |= vec
          num_partial += 1
        else:
          # save partial for later pairing
          partial[sieve_val] = (vec, (sqr, A*x+B, root_A))
      i += 1

    num_smooth = len(smooth)
    num_used_prime = len(used_prime)

    if verbose:
      print 100 * num_smooth / num_prime, 'percent complete\r',

    if num_smooth > num_used_prime:
      if verbose:
        print '%d polynomials sieved (%d values)'%(num_poly, num_poly*x_max*2)
        print 'found %d smooths (%d from partials) in %f seconds'%(num_smooth, num_partial, clock()-time1)
        print 'solving for non-trivial congruencies...'

      used_prime_list = sorted(list(used_prime))

      # set up bit fields for gaussian elimination
      masks = []
      mask = 1
      bit_fields = [0]*num_used_prime
      for vec, vals in smooth:
        masks += [mask]
        i = 0
        for p in used_prime_list:
          if p in vec: bit_fields[i] |= mask
          i += 1
        mask <<= 1

      # row echelon form
      col_offset = 0
      null_cols = []
      for col in xrange(num_smooth):
        pivot = col-col_offset == num_used_prime or bit_fields[col-col_offset] & masks[col] == 0
        for row in xrange(col+1-col_offset, num_used_prime):
          if bit_fields[row] & masks[col]:
            if pivot:
              bit_fields[col-col_offset], bit_fields[row] = bit_fields[row], bit_fields[col-col_offset]
              pivot = False
            else:
              bit_fields[row] ^= bit_fields[col-col_offset]
        if pivot:
          null_cols += [col]
          col_offset += 1

      # reduced row echelon form
      for row in xrange(num_used_prime):
        # lowest set bit
        mask = bit_fields[row] & -bit_fields[row]
        for up_row in xrange(row):
          if bit_fields[up_row] & mask:
            bit_fields[up_row] ^= bit_fields[row]

      # check for non-trivial congruencies
      for col in null_cols:
        all_vec, (lh, rh, rA) = smooth[col]
        lhs = lh   # sieved values (left hand side)
        rhs = [rh] # sieved values - n (right hand side)
        rAs = [rA] # root_As (cofactor of lhs)
        i = 0
        for field in bit_fields:
          if field & masks[col]:
            vec, (lh, rh, rA) = smooth[i]
            lhs += list(all_vec & vec) + lh
            all_vec ^= vec
            rhs += [rh]
            rAs += [rA]
          i += 1

        factor = gcd(list_prod(rAs)*list_prod(lhs) - list_prod(rhs), n)
        if factor != 1 and factor != n:
          break
      else:
        if verbose:
          print 'none found.'
        continue
      break

  if verbose:
    print 'factors found:'
    print factor, 'x', n/factor
    print 'time elapsed: %f seconds'%(clock()-time1)
  return factor

if __name__ == "__main__":
  parser =ArgumentParser(description='Uses a MPQS to factor a composite number')
  parser.add_argument('composite', metavar='number_to_factor', type=long,
      help='the composite number to factor')
  parser.add_argument('--verbose', dest='verbose', action='store_true',
      help="enable verbose output")
  args = parser.parse_args()

  if args.verbose:
    mpqs(args.composite, args.verbose)
  else:
    time1 = clock()
    print mpqs(args.composite)
    print 'time elapsed: %f seconds'%(clock()-time1)

my_math.py

# divide and conquer list product
def list_prod(a):
  size = len(a)
  if size == 1:
    return a[0]
  return list_prod(a[:size>>1]) * list_prod(a[size>>1:])

# greatest common divisor of a and b
def gcd(a, b):
  while b:
    a, b = b, a%b
  return a

# modular inverse of a mod m
def mod_inv(a, m):
  a = int(a%m)
  x, u = 0, 1
  while a:
    x, u = u, x - (m/a)*u
    m, a = a, m%a
  return x

# legendre symbol (a|m)
# note: returns m-1 if a is a non-residue, instead of -1
def legendre(a, m):
  return pow(a, (m-1) >> 1, m)

# modular sqrt(n) mod p
# p must be prime
def mod_sqrt(n, p):
  a = n%p
  if p%4 == 3:
    return pow(a, (p+1) >> 2, p)
  elif p%8 == 5:
    v = pow(a << 1, (p-5) >> 3, p)
    i = ((a*v*v << 1) % p) - 1
    return (a*v*i)%p
  elif p%8 == 1:
    # Shank's method
    q = p-1
    e = 0
    while q&1 == 0:
      e += 1
      q >>= 1

    n = 2
    while legendre(n, p) != p-1:
      n += 1

    w = pow(a, q, p)
    x = pow(a, (q+1) >> 1, p)
    y = pow(n, q, p)
    r = e
    while True:
      if w == 1:
        return x

      v = w
      k = 0
      while v != 1 and k+1 < r:
        v = (v*v)%p
        k += 1

      if k == 0:
        return x

      d = pow(y, 1 << (r-k-1), p)
      x = (x*d)%p
      y = (d*d)%p
      w = (w*y)%p
      r = k
  else: # p == 2
    return a

#integer sqrt of n
def isqrt(n):
  c = n*4/3
  d = c.bit_length()

  a = d>>1
  if d&1:
    x = 1 << a
    y = (x + (n >> a)) >> 1
  else:
    x = (3 << a) >> 2
    y = (x + (c >> a)) >> 1

  if x != y:
    x = y
    y = (x + n/x) >> 1
    while y < x:
      x = y
      y = (x + n/x) >> 1
  return x

# strong probable prime
def is_sprp(n, b=2):
  if n < 2: return False
  d = n-1
  s = 0
  while d&1 == 0:
    s += 1
    d >>= 1

  x = pow(b, d, n)
  if x == 1 or x == n-1:
    return True

  for r in xrange(1, s):
    x = (x * x)%n
    if x == 1:
      return False
    elif x == n-1:
      return True

  return False

# lucas probable prime
# assumes D = 1 (mod 4), (D|n) = -1
def is_lucas_prp(n, D):
  P = 1
  Q = (1-D) >> 2

  # n+1 = 2**r*s where s is odd
  s = n+1
  r = 0
  while s&1 == 0:
    r += 1
    s >>= 1

  # calculate the bit reversal of (odd) s
  # e.g. 19 (10011) <=> 25 (11001)
  t = 0
  while s:
    if s&1:
      t += 1
      s -= 1
    else:
      t <<= 1
      s >>= 1

  # use the same bit reversal process to calculate the sth Lucas number
  # keep track of q = Q**n as we go
  U = 0
  V = 2
  q = 1
  # mod_inv(2, n)
  inv_2 = (n+1) >> 1
  while t:
    if t&1:
      # U, V of n+1
      U, V = ((U + V) * inv_2)%n, ((D*U + V) * inv_2)%n
      q = (q * Q)%n
      t -= 1
    else:
      # U, V of n*2
      U, V = (U * V)%n, (V * V - 2 * q)%n
      q = (q * q)%n
      t >>= 1

  # double s until we have the 2**r*sth Lucas number
  while r:
    U, V = (U * V)%n, (V * V - 2 * q)%n
    q = (q * q)%n
    r -= 1

  # primality check
  # if n is prime, n divides the n+1st Lucas number, given the assumptions
  return U == 0

# primes less than 212
small_primes = set([
    2,  3,  5,  7, 11, 13, 17, 19, 23, 29,
   31, 37, 41, 43, 47, 53, 59, 61, 67, 71,
   73, 79, 83, 89, 97,101,103,107,109,113,
  127,131,137,139,149,151,157,163,167,173,
  179,181,191,193,197,199,211])

# pre-calced sieve of eratosthenes for n = 2, 3, 5, 7
indices = [
    1, 11, 13, 17, 19, 23, 29, 31, 37, 41,
   43, 47, 53, 59, 61, 67, 71, 73, 79, 83,
   89, 97,101,103,107,109,113,121,127,131,
  137,139,143,149,151,157,163,167,169,173,
  179,181,187,191,193,197,199,209]

# distances between sieve values
offsets = [
  10, 2, 4, 2, 4, 6, 2, 6, 4, 2, 4, 6,
   6, 2, 6, 4, 2, 6, 4, 6, 8, 4, 2, 4,
   2, 4, 8, 6, 4, 6, 2, 4, 6, 2, 6, 6,
   4, 2, 4, 6, 2, 6, 4, 2, 4, 2,10, 2]

max_int = 2147483647

# an 'almost certain' primality check
def is_prime(n):
  if n < 212:
    return n in small_primes

  for p in small_primes:
    if n%p == 0:
      return False

  # if n is a 32-bit integer, perform full trial division
  if n <= max_int:
    i = 211
    while i*i < n:
      for o in offsets:
        i += o
        if n%i == 0:
          return False
    return True

  # Baillie-PSW
  # this is technically a probabalistic test, but there are no known pseudoprimes
  if not is_sprp(n, 2): return False

  # idea shamelessly stolen from Mathmatica
  # if n is a 2-sprp and a 3-sprp, n is necessarily square-free
  if not is_sprp(n, 3): return False

  a = 5
  s = 2
  # if n is a perfect square, this will never terminate
  while legendre(a, n) != n-1:
    s = -s
    a = s-a
  return is_lucas_prp(n, a)

# next prime strictly larger than n
def next_prime(n):
  if n < 2:
    return 2
  # first odd larger than n
  n = (n + 1) | 1
  if n < 212:
    while True:
      if n in small_primes:
        return n
      n += 2

  # find our position in the sieve rotation via binary search
  x = int(n%210)
  s = 0
  e = 47
  m = 24
  while m != e:
    if indices[m] < x:
      s = m
      m = (s + e + 1) >> 1
    else:
      e = m
      m = (s + e) >> 1

  i = int(n + (indices[m] - x))
  # adjust offsets
  offs = offsets[m:] + offsets[:m]
  while True:
    for o in offs:
      if is_prime(i):
        return i
      i += o

Sample I/O:

$ pypy mpqs.py --verbose 94968915845307373740134800567566911
smoothness bound: 6117
sieve size: 24360
log threshold: 14.3081031579
skipping primes less than: 47
sieving for smooths...
144 polynomials sieved (7015680 values)
found 405 smooths (168 from partials) in 0.513794 seconds
solving for non-trivial congruencies...
factors found:
216366620575959221 x 438925910071081891
time elapsed: 0.685765 seconds

$ pypy mpqs.py --verbose 523022617466601111760007224100074291200000001
smoothness bound: 9998
sieve size: 37440
log threshold: 15.2376302725
skipping primes less than: 59
sieving for smooths...
428 polynomials sieved (32048640 values)
found 617 smooths (272 from partials) in 1.912131 seconds
solving for non-trivial congruencies...
factors found:
14029308060317546154181 x 37280713718589679646221
time elapsed: 2.064387 seconds

Note: not using the --verbose option will give slightly better timings:

$ pypy mpqs.py 94968915845307373740134800567566911
216366620575959221
time elapsed: 0.630235 seconds

$ pypy mpqs.py 523022617466601111760007224100074291200000001
14029308060317546154181
time elapsed: 1.886068 seconds

Basic Concepts

In general, a quadratic sieve is based on the following observation: any odd composite n may be represented as:

This is not very difficult to confirm. Since n is odd, the distance between any two cofactors of n must be even 2d, where x is the mid point between them. Moreover, the same relation holds for any multiple of n

Note that if any such x and d can be found, it will immediately result in a (not necessarily prime) factor of n, since x + d and x - d both divide n by definition. This relation can be further weakened - at the consequence of allowing potential trivial congruencies - to the following form:

So in general, if we can find two perfect squares which are equivalent mod n, then it's fairly likely that we can directly produce a factor of n a la gcd(x ± d, n). Seems pretty simple, right?

Except it's not. If we intended to conduct an exhaustive search over all possible x, we would need to search the entire range from [n, √(2n)], which is marginally smaller than full trial division, but also requires an expensive is_square operation each iteration to confirm the value of d. Unless it is known beforehand that n has factors very near n, trial division is likely to be faster.

Perhaps we can weaken this relation even more. Suppose we chose an x, such that for

a full prime factorization of y is readily known. If we had enough such relations, we should be able to construct an adequate d, if we choose a number of y such that their product is a perfect square; that is, all prime factors are used an even number of times. In fact, if we have more such y than the total number of unique prime factors they contain, a solution is guaranteed to exist; It becomes a system of linear equations. The question now becomes, how do we chose such x? That's where sieving comes into play.

The Sieve

Consider the polynomial:

Then for any prime p and integer k, the following is true:

This means that after solving for the roots of the polynomial mod p - that is, you've found an x such that y(x) ≡ 0 (mod p), ergo y is divisible by p - then you have found an infinite number of such x. In this way, you can sieve over a range of x, identifying small prime factors of y, hopefully finding some for which all prime factors are small. Such numbers known as k-smooth, where k is the largest prime factor used.

There's a few problems with this approach, though. Not all values of x are adequate, in fact, there's only very few of them, centered around n. Smaller values will become largely negative (due to the -n term), and larger values will become too large, such that it is unlikely that their prime factorization consists only of small primes. There will be a number of such x, but unless the composite you're factoring is very small, it's highly unlikely that you'll find enough smooths to result in a factorization. And so, for larger n, it becomes necessary to sieve over multiple polynomials of a given form.

Multiple Polynomials

So we need more polynomials to sieve? How about this:

That'll work. Note that A and B could literally be any integer value, and the math still holds. All we need to do is choose a few random values, solve for the root of the polynomial, and sieve the values close to zero. At this point we could just call it good enough: if you throw enough stones in random directions, you're bound to break a window sooner or later.

Except, there's a problem with that too. If the slope of the polynomial is large at the x-intercept, which it will be if it isn't relatively flat, there'll only be a few suitable values to sieve per polynomial. It'll work, but you'll end up sieving a whole lot of polynomials before you get what you need. Can we do better?

We can do better. An observation, as a result of Montgomery is as follows: if A and B are chosen such that there exists some C satisfying

then the entire polynomial can be rewritten as

Furthermore, if A is chosen to be a perfect square, the leading A term can be neglected while sieving, resulting in much smaller values, and a much flatter curve. For such a solution to exist, n must be a quadratic residue modA, which can be known immediately by computing the Legendre symbol:
(n | √A) = 1. Note that in order to solve for B, a complete prime factorization of √A needs to be known (in order to take the modular square root √n (mod √A)), which is why √A is typically chosen to be prime.

It can then be shown that if , then for all values of x ∈ [-M, M]:

And now, finally, we have all the components necessary to implement our sieve. Or do we?

Powers of Primes as Factors

Our sieve, as described above, has one major flaw. It can identify which values of x will result in a y divisible by p, but it cannot identify whether or not this y is divisible by a power of p. In order to determine that, we would need to perform trial division on the value to be sieved, until it is no longer divisible by p. We seemed to have reached an impassé: the whole point of the sieve was so that we didn't have to do that. Time to check the playbook.

That looks pretty useful. If the sum of the ln of all of the small prime factors of y is close to the expected value of ln(y), then it's almost a given that y has no other factors. In addition, if we adjust the expected value down a little bit, we can also identify values as smooth which have several powers of primes as factors. In this way, we can use the sieve as a 'pre-screening' process, and only factor those values which are likely to be smooth.

This has a few other advantages as well. Note that small primes contribute very little to the ln sum, but yet they require the most sieve time. Sieving the value 3 requires more time than 11, 13, 17, 19, and 23 combined. Instead, we can just skip the first few primes, and adjust the threshold down accordingly, assuming a certain percentage of them would have passed.

Another result, is that a number of values will be allowed to 'slip through', which are mostly smooth, but contain a single large cofactor. We could just discard these values, but suppose we found another mostly smooth value, with exactly the same cofactor. We can then use these two values to construct a usable y; since their product will contain this large cofactor squared, it no longer needs to be considered.

Putting it all together

The last thing we need to do is to use these values of y construct an adequate x and d. Suppose we only consider the non-square factors of y, that is, the prime factors of an odd power. Then, each y can be expressed in the following manner:

which can be expressed in the matrix form:

The problem then becomes to find a vector v such that vM =(mod 2), where is the null vector. That is, to solve for the left null space of M. This can be done in a number of ways, the simplest of which is to perform Gaussian Elimination on MT, replacing the row addition operation with a row xor. This will result in a number of null space basis vectors, any combination of which will produce a valid solution.

The construction of x is fairly straight-forward. It is simply the product of Ax + B for each of the y used. The construction of d is slightly more complicated. If we were to take the product of all y, we will end up with a value with 10s of thousands, if not 100s of thousands of digits, for which we need to find the square root. This calcuation is impractically expensive. Instead, we can keep track of the even powers of primes during the sieving process, and then use and and xor operations on the vectors of non-square factors to reconstruct the square root.

I seem to have reached the 30000 character limit. Ahh well, I suppose that's good enough.

primo

Posted 2012-10-07T05:16:00.663

Reputation: 30 891

5Well, I never passed algebra in high school ( actually dropped out during the first semester of freshmen year ), but you make it simple to understand from a programmer's perspective. I won't pretend to fully understand it without putting it into practice, but I applaud you. You should consider expanding this post off-site and publishing it, seriously! – jdstankosky – 2012-11-27T14:08:09.377

2I agree. Excellent answer with a great explanation. +1 – Soham Chowdhury – 2012-11-28T07:45:45.557

1@primo Your answers to multiple questions on here have been incredibly thorough and interesting. Much appreciated! – Paul Walls – 2012-12-10T23:12:44.920

@PaulWalls I appreciate the comment :D This question I particularly enjoyed working on. Had I not run out of room, the last section would have been 'Implementation Details' - things that a serious implementation would need to contain, but this one does not (Block Lanczos instead of Gaussian Elimination, ECM factoring of probable smooths, fast modular multiplication, etc.). I also would have liked to mention the various functions in my_math.py, most especially is_prime and next_prime. – primo – 2012-12-11T04:12:30.463

@primo Well, it's unfortunate you did run out of room. You've got a knack for explaining topics in a reasonably straightforward manner. – Paul Walls – 2012-12-11T06:13:25.597

4

As a last remark, I would like to express my gratitude to Will Ness for awarding the +100 bounty on this question. It was literally his entire reputation.

– primo – 2012-12-11T17:59:28.577

@primo, first of all, many kudos!!! I've been working on my own QS implementation and can factor 38! + 1 in around 20 minutes (you can find it here). I've tried larger numbers, and I wanted to see how fast your algorithm could factor 10050120961360479179164300841596861740399588283187. It took my algo close to 13 hours.

– Joseph Wood – 2016-04-28T13:13:59.380

@JosephWood I time about 33s. I do have one idea for further improvement. Using a static log threshold unfortunately rules out most small values (which are actually most likely to be smooth!), whereas the actual curve looks like this: wolfram alpha plot. Using a rolling log threshold would likely find more smooths with less polynomials, although the individual threshold values would need to be fast to approximate to have any sort of speedup.

– primo – 2016-04-29T17:53:32.990

1@primo, 33s... very impressive!!! Your idea about implementing a rolling log threshold could be beneficial, but it could also be detrimental because as you say above "but it also makes the smoothness check slower." This is the problem I ran into (i.e. my smooth checker was extremely slow). Anywho, have you thought about implementing parallel programming? That would surely take your algo to the next level. With parallel programming implemented, you could probably factor 100 digit numbers within a day or so. Again, many kudos!! – Joseph Wood – 2016-04-29T18:22:07.933

1This is a wonderful answer. Pypy got faster so I think your code has too now :) 10050120961360479179164300841596861740399588283187 took 15 seconds on my 4 year old PC. – None – 2016-12-14T14:36:00.933

1This Python library used your function :) – Stephen – 2017-08-20T00:54:34.707

2@StepHen it does. Unfortunately, it uses the original version from 2012, without the speed improvements, and with a bug in in the gaussian elimination (errors when the final column is a pivot column). I attempted to contact the author some time ago, but received no response. – primo – 2017-08-21T06:30:57.267

It seems to be a bit faster now. By the way, but it doesn't work for even semiprimes. – Dennis – 2014-08-27T18:37:12.090

@Dennis Ahh yes, you're right. The Legendre symbol isn't properly defined for the (only) even prime, and thus can't double as trial division for that value. – primo – 2014-08-28T09:24:48.590

I just tried mpqs with 1414788616326148885598999113 = 2147483647 × 658812288653553079 but it returns the value 1. I will see if I can work out why it is doing this. – Arthur Vause – 2014-11-23T16:54:56.930

1@ArmchairCaver I've updated the code to sieve additional polynomials if no non-trivial congruencies are found on the first attempt. – primo – 2014-11-24T05:53:00.120

2

Well, your 38!+1 broke my php script, not sure why. In fact, any semi-prime over 16 digits long breaks my script.

However, using 8980935344490257 ( 86028157 * 104395301 ) my script managed a time of 25.963 seconds on my home computer (2.61GHz AMD Phenom 9950). A lot faster than my work computer which was nearly 31 seconds @ 2.93GHz Core 2 Duo.

php - 757 chars incl. new lines

<?php
function getTime() {
    $t = explode( ' ', microtime() );
    $t = $t[1] + $t[0];
    return $t;
}
function isDecimal($val){ return is_numeric($val) && floor($val) != $val;}
$start = getTime();
$semi_prime = 8980935344490257;
$slice      = round(strlen($semi_prime)/2);
$max        = (pow(10, ($slice))-1);
$i          = 3;
echo "\nFactoring the semi-prime:\n$semi_prime\n\n";

while ($i < $max) {
    $sec_factor = ($semi_prime/$i);
    if (isDecimal($sec_factor) != 1) {
        $mod_f = bcmod($i, 1);
        $mod_s = bcmod($sec_factor, 1);
        if ($mod_f == 0 && $mod_s == 0) {
            echo "First factor = $i\n";
            echo "Second factor = $sec_factor\n";
            $end=getTime();
            $xtime=round($end-$start,4).' seconds';
            echo "\n$xtime\n";
            exit();
        }
    }
    $i += 2;
}
?>

I'd be interested to see this same algorithm in c or some other compiled language.

jdstankosky

Posted 2012-10-07T05:16:00.663

Reputation: 1 474

PHP's numbers have only 53 bits precision, roughly 16 decimal digits – copy – 2012-11-03T21:31:02.273

3Implementing the same algorithm in C++ using 64 bit integers only took about 1.8 seconds to run on my computer. There are several problems with this approach though:

  1. It can't handle large enough numbers.

  2. Even if it could & assuming all numbers, no matter the length, used the same amount of time for trial division, every order of magnitude increase would result in an equivalent amount of time increase. Since your first factor is about 14 orders of magnitude smaller than the given first factor, this algorithm would take over 9 million years to factor the given semiprime.

  3. < – CasaDeRobison – 2012-11-04T06:21:04.293

I'm not the best at math, admittedly, but for very large numbers, standard methods of factoring semi-primes will simply not work (using an elipse, etc.), as far as I know. With that in mind, how could the algorithm itself be improved? – jdstankosky – 2012-11-05T14:27:41.337

If I knew how to improve the algorithm, I'd be rich! :) I was just observing that the algorithm as is would most likely not solve the stated problem in a measurable time period. I'm still trying to come up with a solution of my own. – CasaDeRobison – 2012-11-12T03:52:11.697

The efficiency can be greatly increased by using a pre-calced sieve of eratosthenes. See the is_prime implementation from my solution for details. It uses full trial division whenever the number to be tested can be represented as a 32-bit integer. – primo – 2012-11-27T08:26:21.703

@primo No idea what you just said, but I guarantee I'll Google the heck out of it! – jdstankosky – 2012-11-27T14:09:18.983

2

The Sieve of Eratosthenes starts with a list of number, then removes all multiples of 2, and then 3, and then 5, and then 7, etc. What remains after the sieve is complete are only prime numbers. This sieve can be 'pre-calced' for a certain number of factors. Because lcm(2, 3, 5, 7) == 210, the pattern of numbers eliminated by these factors will repeat every 210 numbers, and only 48 remain. In that way, you can eliminate 77% of all numbers from trial division, instead of the 50% by taking only odds.

– primo – 2012-11-27T15:26:33.103

1@primo Out of curiosity, how much time did you devote to this? It would have taken me ages to think of this stuff. At the time I wrote this, I was only thinking about how prime numbers were always odd. I didn't attempt to go beyond that and eliminate non-prime odds as well. It seems so simple in retrospect. – jdstankosky – 2012-11-27T17:39:36.753

@jdstankosky To be honest, I spent almost as much time writing the is_prime and next_prime functions as I did writing the sieve itself. – primo – 2012-11-27T17:50:14.830

Hmm, I think PHP must be pretty slow. In C, factoring the number 8980935344490257 by trial division of known primes up to its square root takes only 0.09 seconds on my 1.7 GHz Intel Core i5 laptop. (As a comparison with another algorithm, Shanks's Square Forms Factorization requires only 0.0003 seconds.) – Todd Lehman – 2014-08-11T08:46:02.177