Lecture from: 20.05.2025 | Video: Homelab

In this lecture, we study a classic geometric optimization problem and a strikingly elegant randomized algorithm to solve it in expected linearithmic time. The algorithm, originating from work by Clarkson, illustrates powerful ideas in randomized geometric optimization and probabilistic analysis.

The SmallEnclDisk Problem

Given a finite point set , determine the closed disk of minimal radius that contains all points in .

We define:

  • A disk with center and radius as:

  • The disk is said to enclose if:

  • This disk is called the smallest enclosing circle if it has the smallest possible radius among all such enclosing disks.

Lemma: Uniqueness of the Smallest Enclosing Circle

For every finite point set , there exists a unique smallest enclosing circle .

Proof sketch:

  • Suppose two distinct disks both minimally enclose , with equal radius but centers .
  • Let be the midpoint.
  • Construct a disk centered at passing through the intersection points of .
  • This new disk strictly contains and has smaller radius than - contradiction.

Hence, the smallest enclosing disk is uniquely defined.

Lemma: 3-Point Certificate

For , there exists a subset , , such that:

That is, the smallest enclosing circle is determined by at most 3 boundary points. These 3 points form a minimal certificate of the optimal solution.

Intuitively:

  • If contains all of , and any point lies in the interior, it’s not necessary to determine the boundary.
  • Only up to 3 points (due to geometry of circles) can lie on the boundary and determine the unique minimal disk.

Basic Algorithms

Complete Enumeration

Enumerate all 3-subsets . For each:

  • Compute
  • Check whether

  • There are candidates.
  • Each check takes .
  • Total time:

Smarter Enumeration

Instead of checking containment, track the candidate with largest radius:

  • Still iterate over subsets.
  • Keep best disk seen so far.

  • Runtime:
  • Still inefficient for large

A First Randomized Algorithm

  • Expected trials: since probability of correctly selecting the points in one sample is and this is a geometric distribution.
  • Each trial: check
  • Total expected runtime:

This naive approach lacks adaptivity: it doesn’t learn from previous failures. But the insight that “bad points” (outside current disk) matter more leads to improvement.

Clarkson’s Randomized Algorithm

We maintain a multiset initialized as , and evolve it by amplifying the influence of points that prevent current samples from enclosing the set.

  • Sample is drawn from , which changes over time.
  • Violating points (outside current disk) become more likely in future samples.
  • Idea resembles a multiplicative weights update method - emphasizing “important” points.

Why It Works

Let be the set of at most 3 points that define the optimal disk: .

  • If current disk fails: at least one point in is outside.
  • Each time we fail, at least one such point is doubled.
  • After rounds, some point in must appear times.
  • Hence,

But we can also bound expected size of from above.

Growth Upper Bound

Let be the size of after rounds. Then:

For , this becomes

Since , these bounds must eventually conflict → the process must terminate in rounds.

Sampling Lemma

This key lemma bounds expected number of points outside disk defined by random sample.

Let be a multiset of size , and be a uniform random sample of size . Then:

Intuition:

  • A point implies that is “essential” for any superset.
  • But at most 3 such essential points per subset.

Proof sketch: Define indicator functions:

  • if
  • if

Connect expected number of outs to essentiality and count via combinatorics.

Look at the proof below, I didn’t fully get it…

Final Runtime Analysis

Let be the number of rounds until success.

We showed:

  • But if , then

Set . For , the probability that algorithm hasn’t terminated is very small.

Finally, the expected number of rounds is:

Total expected runtime:

Conclusion

Theorem

The randomized algorithm computes the smallest enclosing circle of in expected time .

Summary

  • The method generalizes to enclosing balls (or ellipsoids) in fixed higher dimensions with modified constants.
  • Other linear-time randomized and deterministic methods exist, but Clarkson’s approach is notable for its “simplicity” and analysis.
ConceptInsight
Smallest enclosing circleDetermined by ≤ 3 points on boundary
Randomized samplingLikely to hit defining points after boosting their multiplicity
Geometric sensitivityPoints outside a sampled disk are crucial - “they teach us something”
Sampling lemmaShows expected number of offending points is small
Multiplicative biasingDrives convergence by boosting important point probability

Continue here: 25 Convex Hulls - Geometry, Algorithms, and Complexity