Recent puzzling observations of tiny variations in nuclear decay rates have led some to question the science of using decay rates to determine the relative ages of rocks and organic materials. Scientists from the National Institute of Standards and Technology (NIST), working with researchers from Purdue University, the University of Tennessee, Oak Ridge National Laboratory and Wabash College, tested the hypothesis that solar radiation might affect the rate at which radioactive elements decay and found no detectable effect.
Atoms of radioactive isotopes are unstable and decay over time by shooting off particles at a fixed rate, transmuting the material into a more stable substance. For instance, half the mass of carbon-14, an unstable isotope of carbon, will decay into nitrogen-14 over a period of 5,730 years. The unswerving regularity of this decay allows scientists to determine the age of extremely old organic materialssuch as remains of Paleolithic campfireswith a fair degree of precision. The decay of uranium-238, which has a half-life of nearly 4.5 billion years, enabled geologists to determine the age of the Earth.
Many scientists, including Marie and Pierre Curie, Ernest Rutherford and George de Hevesy, have attempted to influence the rate of radioactive decay by radically changing the pressure, temperature, magnetic field, acceleration, or radiation environment of the source. No experiment to date has detected any change in rates of decay.
Recently, however, researchers at Purdue University observed a small (a fraction of a percent), transitory deviation in radioactive decay at the time of a huge solar flare. Data from laboratories in New York and Germany also have shown similarly tiny deviations over the course of a year. This has led some to suggest that Earth's distance from the sun, which varies during the year and affects the planet's exposure to solar neutrinos, might be related to these anomalies.
Researchers from NI
|Contact: Mark Esser|
National Institute of Standards and Technology (NIST)