Read online free
  • Home
  • Romance & Love
  • Fantasy
  • Science Fiction
  • Mystery & Detective
  • Thrillers & Crime
  • Actions & Adventure
  • History & Fiction
  • Horror
  • Western
  • Humor

    Lawrence Krauss - The Greatest Story Ever Told--So Far

    Prev Next


      is no going back.

      ͣ͞

      C h a p t e r 2

      S E E I N G I N T H E D A R K

      Let there be light: and there was light.

      —GENESIS 1:3

      In the beginning there was light.

      It is no coincidence that the ancients imagined in Genesis that

      light was created on the first day. Without light, there would be little

      awareness of the vast universe surrounding us. When we nod and

      say, “I see,” to a friend who is trying to explain something, we convey

      far more than just an observation, but rather a fundamental

      understanding.

      Plato’s allegory was appropriately centered on light—light from a

      fire to cast the shadows on the cave wall and light from the outside

      to temporarily blind the freed prisoner and then illuminate the real

      world for him. Like the prisoners in the cave, we too are prisoners of

      light—almost everything we learn about the world we learn from

      what we see.

      While the most significant words in the Western religious canon

      may be Let there be light, in the modern world this phrase now has a

      completely different significance from what it once did. Human

      beings may be prisoners of light, but so is the universe. What once

      appeared as a whim of a Judeo-Christian God, or other gods before

      that one, we now understand to be required by the very laws that

      allow both heaven, and more important, Earth, to exist. You cannot

      have one without the other. Earth, or matter, follows light.

      ͤ͞

      This change in perception underlies almost every development in

      the edifice we call modern science. I am writing these words as I

      stare out from a ship at one of the Galápagos Islands, which Charles

      Darwin made famous, and which made him famous in return, as he

      changed our perception of life and its diversity with a single brilliant

      realization: that all living species developed through the natural

      selection of small inherited variations that are passed along to future

      generations by survivors. As surely as the understanding of evolution

      changed everything about our understanding of biology, our

      changing understanding of light changed everything about our

      physical understanding of our place in the universe. As a useful

      fringe benefit, this change resulted in virtually all of the technology

      on which the modern world is based.

      The extent to which our observations of the world imprison our

      minds, and frame our description of the fabric of the universe,

      remained unappreciated for more than twenty centuries following

      Plato. Once serious minds began to investigate in detail the hidden

      nature of the universe, it took over four centuries for them to fully

      resolve the question What is light?

      Perhaps the most serious modern mind, although certainly not

      the first, to ask this question was also one of the most famous—and

      oddest—scientists in history: Isaac Newton. It is not inappropriate to

      classify Newton as a modern mind—after all, his seventeenth-

      century Principia: Mathematical Principles of Natural Philosophy

      uncovered the classical laws of motion and laid the basis for his

      theory of gravity, both of which form the foundation of much of

      modern physics. Nevertheless, as John Maynard Keynes pointed out:

      Newton was not the first of the age of reason, he was the last of the

      magicians, the last of the Babylonians and Sumerians, the last

      great mind that looked out on the visible and intellectual world

      ͥ͞

      with the same eyes as those who began to build our intellectual

      inheritance rather less than 10,000 years ago.

      The truth of this statement reflects the revolutionary importance

      of Newton’s work. After the Principia, no rational person could view

      the world the same way the ancients had viewed it. But it also

      reflects the character of Newton himself. He devoted far more time,

      and far more ink, to writing about the occult, alchemy, and

      searching for hidden meanings and codes in the Bible—focusing in

      particular on the Book of Revelation and mysteries associated with

      the ancient Temple of Solomon—than he did to writing about

      physics.

      Newton was also one in a long line of people, which extends

      before and after him, who felt that he had been specifically chosen

      by God to help reveal the true meaning of the Scriptures. To what

      extent his studies of the universe derived from his fascination with

      the Bible is not clear, but it does seem reasonable to conclude that

      his primary interest was in theology, and that natural philosophy

      came in well below that, and probably below alchemy as well.

      Many individuals point to Newton’s fascination with God as

      evidence of the compatibility between science and religion, and to

      assert that modern science owes its existence to Christianity. This

      confuses history with causality. It is undeniable that many of the

      early giants of modern Western natural philosophy, from Newton

      onward, were deeply religious, although Darwin lost much, if not all,

      of his religious belief later in life. But remember that during much of

      this period there were primarily two sources of education and

      wealth: the Church and the Crown. The Church was the National

      Science Foundation of the fifteenth, sixteenth, and seventeenth

      centuries. All institutions of higher learning were tied to various

      denominations, and it was unthinkable for any educated person to

      not be affiliated with the Church. And as Giordano Bruno and later

      ͟͜

      Galileo discovered, it was unpleasant at best to counter its doctrine.

      It would have been remarkable for any of these leading early

      scientific thinkers to have been anything but religious.

      The religiosity of the early scientific pioneers is also cited today by

      sophists who claim that science and religious doctrine are

      compatible, but who confuse science and scientists. In spite of

      frequent appearances to the contrary, scientists are people. And like

      all people they are capable of holding many potentially mutually

      contradictory notions in their head at the same time. No correlation

      between divergent views held by any individual is representative of

      anything but human foibles.

      To claim that some scientists are or were religious is like saying

      some scientists are Republicans or some are flat-earthers or some are

      creationists. It doesn’t imply causality or consistency. My friend

      Richard Dawkins has told me of a professor of astrophysics who,

      during the day, writes papers that are published in astronomical

      journals assuming that the universe is more than 13 billion years old,

      but then goes home and privately espouses the literal biblical claim

      that the universe is six thousand years old.

      What determines intellectual consistency or lack thereof in the

      sciences is a combination of rational arguments with subsequent

      evidence and continued testing. It is perfectly reasonable to claim

      that religion, in the Western world, may be the mother of science.

      But as any parent knows, children rarely grow up to be
    models of

      their parents.

      Newton may, following tradition, have been motivated to look at

      light because it was a gift from God. But we remember his work not

      because of his motivation, but because of what he discovered.

      Newton was convinced that light was made of particles, which he

      referred to as corpuscles, while Descartes, and later Newton’s

      nemesis Robert Hooke, and still later the Dutch scientist Christiaan

      ͟͝

      Huygens, all claimed that light was a wave. One of the key

      observations that appeared to support the wave theory was that

      white light, such as light from the Sun, could split into all the colors

      of the rainbow when passed through a prism.

      As was often the case during his life, Newton believed that he was

      correct and several of his most famous contemporaries (and

      competitors) were wrong. To demonstrate this, he devised a clever

      experiment using prisms that he first performed while at home in

      Woolsthorpe, to escape the bubonic plague ravaging Cambridge. As

      he reported at the Royal Society in 1672, on the forty-fourth try, he

      observed precisely what he hoped he would see.

      Advocates of the wave theory had argued that light waves were

      made of white light and that the light split into colors when it passed

      through a prism because of “corruption” of the rays as they traversed

      the glass. In this case, the more glass, the more splitting.

      Newton reasoned that this was not the case, but that light is made

      of colored particles that combine together to appear white. (With a

      nod to his occult fascination, Newton classified the colored particles

      of the spectrum—a term he coined—into seven different types: red,

      orange, yellow, green, blue, indigo, and violet. From the time of the

      Greeks, the number seven had been considered to possess mystical

      qualities.) To demonstrate that the wave/corruption picture was

      incorrect, Newton passed a beam of white light through two prisms

      held in opposite orientations. The first prism split the light into its

      spectrum, and the second recomposed it back into a single white

      light beam. This result would have been impossible if the glass had

      corrupted the light. A second prism would have simply made the

      situation worse and would not have caused the light to revert back to

      its original state.

      This result does not in fact disprove the wave theory of light (it

      actually supports it, because light slows down as it bends upon

      ͟͞

      entering the prism, just as waves would do). But since the advocates

      of that theory had argued (incorrectly) that the spectral splitting was

      due to corruption, Newton’s demonstration that this was not the

      case struck a significant blow in favor of his particle model.

      Newton went on to discover many other facets of light that we

      use today in our understanding of the wave nature of light. He

      showed that every color of light has a unique bend angle when

      passing through a glass prism. He also showed that all objects appear

      to be the same color as the color of the light beam illuminating

      them. And he showed that colored light will not change its color no

      matter how many times it is reflected by or passes through a prism.

      All of these results, including his original result, can be explained

      simply if white light is indeed composed of a collection of different

      colors—that much he got right. But they can’t be explained if light is

      made of different-colored particles. Rather, white light is composed

      of waves of many different wavelengths.

      Newton’s opponents did not give up easily, even in the face of

      Newton’s rising popularity and the death of his chief opponent,

      Hooke. They did not give up even after Newton’s election as

      president of the Royal Society in 1703, the year before he actually

      published his research on light in his epic Opticks. Indeed, the debate

      on the nature of light continued to rage on for over a century.

      Part of the problem with a wave picture of light was the question

      “What is it that light is a wave of, exactly?” And if it is a wave, then

      since all known waves require some medium, what medium does it

      travel in? These questions were sufficiently perplexing that

      practitioners of the wave theory had to resurrect a new invisible

      substance permeating all space, the ether.

      The resolution of this conundrum came, as such resolutions often

      do, from a totally unexpected corner of the physical world, one full

      of sparks, and spinning wheels.

      ͟͟

      When I was a young professor at Yale—in the ancient but huge

      office I was lucky enough to commandeer when an equally ancient

      colleague retired—there was left hanging for me a copy of a

      photograph of Michael Faraday taken in 1861. I have treasured it

      ever since.

      I don’t believe in hero worship, but if I did, Faraday would be up

      there with the best. Perhaps more than any other scientist of the

      nineteenth century, he is responsible for the technology that powers

      our current civilization. Yet he had little formal education and at age

      fourteen became a bookbinder’s apprentice. Later in his career, after

      achieving world recognition for his scientific contributions, he

      insisted on keeping to his humble roots, turning down a knighthood

      and twice turning down the presidency of the Royal Society. Later on

      he refused to advise the British government on the production of

      chemical weapons for use in the Crimean War, citing ethical

      reasons. And for more than thirty-three years he gave a series of

      Christmas lectures at the Royal Institution to excite young people

      about science. What’s not to like?

      Much as one might admire the man, it is the scientist who

      matters here for our story. Faraday’s first scientific lesson is one I tell

      my students: always suck up to your professors. At the age of twenty,

      after completing seven years of apprenticeship as a bookbinder,

      Faraday attended the lectures of the famous chemist Humphry Davy,

      then the head of the Royal Institution. Afterward Faraday presented

      Davy with a three-hundred-page, beautifully bound book containing

      the notes Faraday had taken during the lectures. Within a year,

      Faraday was appointed Davy’s secretary and shortly thereafter got an

      appointment as chemical assistant in the Royal Institution. Later on,

      Faraday learned the same lesson but with the opposite result.

      Following his excitement over some early, quite significant

      experiments that he performed, Faraday accidentally forgot to

      ͟͠

      acknowledge work with Davy in his published results. This

      accidental snub probably resulted in his being reassigned to other

      activities by Davy and delaying his world-changing research by

      several years.

      When reassigned, Faraday had been working on the “hot” area of

      scientific research, the newly discovered connections between

      electricity and magnetism, driven by results of the Danish physicist

      Hans Christian Oersted. These two forces seem quite different, yet

      have odd simi
    larities. Electric charges can attract or repel. So can

      magnets. Yet magnets always seem to have two poles, north and

      south, which cannot be isolated, while electric charges can

      individually be positive or negative.

      For some time, scientists and natural philosophers had wondered

      if the two forces might have some hidden connection, and the first

      empirical clue came to Oersted by accident. In 1820, while

      delivering a lecture, Oersted saw that a compass needle was

      deflected when an electric current from a battery was switched on. A

      few months later he followed up on this observation and discovered

      that a current of moving electric charges, which we now commonly

      call an electric current, produced a magnetic attraction that caused

      compass needles to point in a circle around the wire.

      He had blazed a new trail. Word spread quickly among scientists,

      through the Continent and across the English Channel. Moving

      electric charges produced a magnetic force. Could there be other

      connections? Could magnets in turn influence electric charges?

      Scientists searched for such a possibility, without success. Davy

      and another colleague tried to build an electric motor based on the

      connection discovered by Oersted, but failed. Faraday ultimately got

      a wire with a current in it to move around a magnet, which did form

      a crude sort of motor. It was this exciting development that he

      reported without citing Davy’s name.

      ͟͡

      Partly this was mere gamesmanship. No new fundamental

      phenomenon was being uncovered. Perhaps this was the rationale

      for one of my favorite (likely apocryphal) stories about Faraday. It is

      said that William Gladstone, later to be British prime minister, heard

      of Faraday’s laboratory, full of weird devices, and asked in 1850 what

      the practical value of all this study into electricity was. Faraday was

      purported to have replied, “Why, sir, there is every probability that

      you will soon be able to tax it.”

      Apocryphal or not, both great irony and truth are in that witty

      comeback. Curiosity-driven research may seem self-indulgent and

      far from the immediate public good. However, essentially all of our

      current quality of life, for people living in the first world, has arisen

      from the fruits of such research, including all the electric power that

     

    Prev Next
Read online free - Copyright 2016 - 2025