Decomposing a factorial into large factors
What's new 2025-03-27
I’ve just uploaded to the arXiv the paper “Decomposing a factorial into large factors“. This paper studies the quantity , defined as the largest quantity such that it is possible to factorize
into
factors
, each of which is at least
. The first few values of this sequence are
This quantity was introduced by Erdös, who asked for upper and lower bounds on
; informally, this asks how equitably one can split up
into
factors. When factoring an arbitrary number, this is essentially a variant of the notorious knapsack problem (after taking logarithms), but one can hope that the specific structure of the factorial
can make this particular knapsack-type problem more tractable. Since
Some further exploration of was conducted by Guy and Selfridge. There is a simple construction that gives the lower bound
- (i) Is
for all
?
- (ii) Is
for all
? (At
, this conjecture barely fails:
.)
- (iii) Is
for all
?
In this note we establish the bounds
asThe upper bound argument for (3) is simple enough that it could also be modified to establish the first conjecture (i) of Guy and Selfridge; in principle, (ii) and (iii) are now also reducible to a finite computation, but unfortunately the implied constants in the lower bound of (3) are too weak to make this directly feasible. However, it may be possible to now crowdsource the verification of (ii) and (iii) by supplying a suitable set of factorizations to cover medium sized , combined with some effective version of the lower bound argument that can establish
for all
past a certain threshold. The value
singled out by Guy and Selfridge appears to be quite a suitable test case: the constructions I tried fell just a little short of the conjectured threshold of
, but it seems barely within reach that a sufficiently efficient rearrangement of factors can work here.
We now describe the proof of the upper and lower bound in (3). To improve upon the trivial upper bound (1), one can use the large prime factors of . Indeed, every prime
between
and
divides
at least once (and the ones between
and
divide it twice), and any factor
that contains such a factor therefore has to be significantly larger than the benchmark value of
. This observation already readily leads to some upper bound of the shape (4) for some
; if one also uses the primes
that are slightly less than
(noting that any multiple of
that exceeds
, must in fact exceed
) is what leads to the precise constant
.
For previous lower bound constructions, one started with the initial factorization and then tried to “improve” this factorization by moving around some of the prime factors. For the lower bound in (3), we start instead with an approximate factorization roughly of the shape
The general approach of first locating some approximate factorization of (where the approximation is in the “adelic” sense of having not just approximately the right magnitude, but also approximately the right number of factors of
for various primes
), and then moving factors around to get an exact factorization of
, looks promising for also resolving the conjectures (ii), (iii) mentioned above. For instance, I was numerically able to verify that
by the following procedure:
- Start with the approximate factorization of
,
by
. Thus
is the product of
odd numbers, each of which is at least
.
- Call an odd prime
-heavy if it divides
more often than
, and
-heavy if it divides
more often than
. It turns out that there are
more
-heavy primes than
-heavy primes (counting multiplicity). On the other hand,
contains
powers of
, while
has none. This represents the (multi-)set of primes one has to redistribute in order to convert a factorization of
to a factorization of
.
- Using a greedy algorithm, one can match a
-heavy prime
to each
-heavy prime
(counting multiplicity) in such a way that
for a small
(in most cases one can make
, and often one also has
). If we then replace
in the factorization of
by
for each
-heavy prime
, this increases
(and does not decrease any of the
factors of
), while eliminating all the
-heavy primes. With a somewhat crude matching algorithm, I was able to do this using
of the
powers of
dividing
, leaving
powers remaining at my disposal. (I don’t claim that this is the most efficient matching, in terms of powers of two required, but it sufficed.)
- There are still
-heavy primes left over in the factorization of (the modified version of)
. Replacing each of these primes with
, and then distributing the remaining
powers of two arbitrarily, this obtains a factorization of
into
terms, each of which are at least
.
However, I was not able to adjust parameters to reach in this manner. Perhaps some readers here who are adept with computers can come up with a more efficient construction to get closer to this bound? If one can find a way to reach this bound, most likely it can be adapted to then resolve conjectures (ii) and (iii) above after some additional numerical effort.