*Michael Fowler
UVa Physics Department*

In this lecture, we shall show how the Greeks made the first real measurements of astronomical distances: the size of the earth and the distance to the moon, both determined quite accurately, and the distance to the sun, where their best estimate fell short by a factor of two.

The first reasonably good measurement of the earth’s
size was done by Eratosthenes,
a Greek who lived in

The distance from

Of course, Eratosthenes fully recognized that the Earth is
spherical in shape, and that “vertically downwards” anywhere on the
surface just means the direction towards the center from that point. Thus two
vertical sticks, one at *were* parallel. Therefore, if the sun’s
rays were parallel to a vertical stick at Syene (so it had no shadow) the angle
they made with the stick at

According to the Greek historian Cleomedes, Eratosthenes
measured the angle between the sunlight and the stick at midday in midsummer in

How do we begin to measure the distance from the earth to the moon? One obvious thought is to measure the angle to the moon from two cities far apart at the same time, and construct a similar triangle, like Thales measuring the distance of the ship at sea. Unfortunately, the angle difference from two points a few hundred miles apart was too small to be measurable by the techniques in use at the time, so that method wouldn’t work.

Nevertheless, Greek astronomers, beginning with Aristarchus of Samos (310-230 B.C., approximately) came up with a clever method of finding the moon’s distance, by careful observation of a lunar eclipse, which happens when the earth shields the moon from the sun’s light.

For a Flash movie of a lunar eclipse, click here!

To better visualize a lunar eclipse, just imagine holding up
a quarter (diameter one inch approximately) at the distance where it just
blocks out the sun’s rays from one eye. Of course you shouldn’t try
this---you’ll damage your eye! You *can* try it with the full moon,
which happens to be the same apparent size in the sky as the sun. It turns out
that the right distance is about nine feet away, or 108 inches. If the quarter
is further away than that, it is not big enough to block out all the sunlight.
If it is closer than 108 inches, it will totally block the sunlight from some
small circular area, which gradually increases in size moving towards the
quarter. Thus the part of space where the sunlight is *totally *blocked is
conical, like a long slowly tapering icecream cone, with the point 108 inches
behind the quarter. Of course, this is surrounded by a fuzzier area, called the
“penumbra”, where the sunlight is partially blocked. The fully
shaded area is called the “umbra”. (This is Latin for shadow.
Umbrella means little shadow in Italian.) If you tape a quarter to the end of a
thin stick, and hold it in the sun appropriately, you can see these different
shadow areas.

*Question: If you used a dime instead of a quarter, how
far from your eye would you have to hold it to just block the full moonlight
from that eye? How do the different distances relate to the relative sizes of
the dime and the quarter? Draw a diagram showing the two conical shadows.*

Now imagine you’re out in space, some distance from
the earth, looking at the earth’s shadow. (Of course, you could only
really *see* it if you shot out a cloud of tiny particles and watched
which of them glistened in the sunlight, and which were in the dark.) Clearly,
the earth’s shadow must be conical, just like that from the quarter. And
it must also be *similar* to the quarter’s in the technical
sense---it must be 108 earth diameters long! That is because the point of the
cone is the furthest point at which the earth can block all the sunlight, and
the ratio of that distance to the diameter is determined by the angular size of
the sun being blocked. This means the cone is 108 earth diameters long, the far
point 864,000 miles from earth.

Now, during a total lunar eclipse the moon moves into this
cone of darkness. Even when the moon is completely inside the shadow, it can
still be dimly seen, because of light scattered by the earth’s
atmosphere. By observing the moon carefully during the eclipse, and seeing how
the earth’s shadow fell on it, the Greeks found that *the diameter of
the earth’s conical shadow at the distance of the moon was about
two-and-a-half times the moon’s own diameter*.

*Note: It is possible to check this estimate either from a
photograph of the moon entering the earth’s shadow, or, better, by actual
observation of a lunar eclipse*.

*Question: At this point the Greeks knew the size of the
earth (approximately a sphere 8,000 miles in diameter) and therefore the size
of the earth’s conical shadow (length 108 times 8,000 miles). They knew
that when the moon passed through the shadow, the shadow diameter at that
distance was two and a half times the moon’s diameter. Was that enough
information to figure out how far away the moon was?*

Well, it did tell them the moon was no further away than
108x8,000 = 864,000 miles, otherwise the moon wouldn’t pass through the
earth’s shadow at all! But from what we’ve said so far, it could be
a tiny moon almost 864,000 miles away, passing through that last bit of shadow
near the point. However, such a tiny moon could never cause a *solar*
eclipse. In fact, as the Greeks well knew, *the moon is the same apparent
size in the sky as the sun*. This is the crucial extra fact they used to
nail down the moon’s distance from earth.

They solved the problem using geometry, constructing the figure below. In this figure, the fact that the moon and the sun have the same apparent size in the sky means that the angle ECD is the same as the angle EAF. Notice now that the length FE is the diameter of the earth’s shadow at the distance of the moon, and the length ED is the diameter of the moon. The Greeks found by observation of the lunar eclipse that the ratio of FE to ED was 2.5 to 1, so looking at the similar isosceles triangles FAE and DCE, we deduce that AE is 2.5 times as long as EC, from which AC is 3.5 times as long as EC. But they knew that AC must be 108 earth diameters in length, and taking the earth’s diameter to be 8,000 miles, the furthest point of the conical shadow, A, is 864,000 miles from earth. From the above argument, this is 3.5 times further away than the moon is, so the distance to the moon is 864,000/3.5 miles, about 240,000 miles. This is within a few percent of the right figure. The biggest source of error is likely the estimate of the ratio of the moon’s size to that of the earth’s shadow as it passes through.

This was an even more difficult question the Greek astronomers asked themselves, and they didn’t do so well. They did come up with a very ingenious method to measure the sun’s distance, but it proved too demanding in that they could not measure the important angle accurately enough. Still, they did learn from this approach that the sun was much further away than the moon, and consequently, since it has the same apparent size, it must be much bigger than either the moon or the earth.

Their idea for measuring the sun’s distance was very simple in principle. They knew, of course, that the moon shone by reflecting the sun’s light. Therefore, they reasoned, when the moon appears to be exactly half full, the line from the moon to the sun must be exactly perpendicular to the line from the moon to the observer (see the figure to convince yourself of this). So, if an observer on earth, on observing a half moon in daylight, measures carefully the angle between the direction of the moon and the direction of the sun, the angle a in the figure, he should be able to construct a long thin triangle, with its baseline the earth-moon line, having an angle of 90 degrees at one end and a at the other, and so find the ratio of the sun’s distance to the moon’s distance.

The problem with this approach is that the angle a turns out to differ from 90 degrees by
about a sixth of a degree, too small to measure accurately. The first attempt
was by Aristarchus, who estimated the angle to be 3 degrees. This would put the
sun only five million miles away. However, it would already suggest the sun to
be *much larger than the earth*. It was probably this realization that led
Aristarchus to suggest that the sun, rather than the earth, was at the center
of the universe. The best later Greek attempts found the sun’s distance
to be about half the correct value (92 million miles).

The presentation here is similar to that in Eric Rogers, *Physics
for the Inquiring Mind*,

Some exercises related to this material are presented in my notes for Physics 621.

Copyright © Michael Fowler 1996, 2007 except where otherwise noted.