At funfairs, there are often merry-go-rounds or more dangerous variants on this. I went on one at the Easter Show callled the 'gravitron': a big cylinder that spun--we were 'pinned to the wall'. On these, a ball does not travel in a vertical plane: it seems to turn corners. Kinematics seems seriously weird in this frame. Then the guy running the gadget shouted at us to stop throwing the ball and he was probably right to do so, because it was hard to predict where it was going to go. Inside the 'gravitron', or another fairground ride, many such simple experiments will show that Newton's laws seem to fail, and so that you are in a non-inertial frame. See the principle of relativity.
Now I look at you inside the car. At first, as the car swerves to the left, you continue in your straight line. The car seat slides under you: the car is curving and you are not. But this happens only for a short time. The sliding of the seat under you brings the raised edge of the seat against your thigh, and perhaps the seat belt against your chest. These lateral forces exerted by the seat edge and the seat belt push you around the corner: they provide the centripital force.
The observer above the car is in an inertial frame. For this observer, Newton's laws work: with an icy road and no horizontal force, the car travels in a straight line. No force, no acceleration. Add frictional force and there is an acceleration. However, if you measure all positions, velocities and accelerations inside the car, you are in a non-inertial frame. In this frame of reference, you appear to accelerate sideways towards the outside of the car. To deal with this, we could introduce fictitious forces. This is sometimes done. For instance, the surface of the Earth is rotating, but it is convenient to measure motion with respect to the surface of the Earth. When we do so, we have to introduced fictitious forces: both centrifugal force and the Coriolis force.
The two airplane version of the twin paradox: is General Relativity involved?(See the twin paradox for an introduction.) "We were discussing proof of special relativity/time dilation in class and used, as an example, the idea of a clock being taken on a fast plane having time run slower than an identical clock left on the ground. The suggestion was made, however, that this plane would really be accelerating (circling the earth) and would therefore be in a non-inertial frame of reference and we would need to use general relativity!"
So for both reasons (gravity and accelerating frame, which are locally indistinguishable according to the Principle of GR), there is a GR correction. The gravitational term is of the same order for both planes. In fact, the gravitational and SR terms turn out to be of comparable size: both are hundreds of nanoseconds. The acceleration term is smaller than the gravitational term. (The acceleration and gravitational terms would be comparable for a satellite, but planes travel much more slowly than near-earth satellites. The SR and gravitational terms are comparable for an object on the Earth's surface.) So yes, an explanation of the time difference in the two clocks requires either an explicit calculation of the two terms. The original report is: J.C. Hafele and R. E. Keating, Science 177, 166 (1972). In fact, the GR terms, while of the same order as the SR terms, are fairly similar for the two planes. So the main effect is the SR effect, and it is in agreement with SR calculations.
|
Seen from a non-rotating frame of reference, above the South pole, the Eastward flying plane has its speed plus the speed of the ground (the atmosphere travels with the Earth, to a good approximation). The Earth turns below the Westward flying plane. At sufficiently high latitude, it can stay in the same time zone for the entire flight.
|
Much more impressive demonstrations of time dilation are in the life times of subatomic particles, particularly in cosmic rays, where the factors can be much greater than one.
But the round the world clock experiment is useful because, provided that one does the GR correction, it also answers the remaining supporters of the twin paradox.
Incidentally, I cannot resist recommending a lovely paper by Sam Drake: The equivalence principle as a stepping stone from special relativity to general relativity: A Socratic dialog.
The relationship between thought and reality is the province of philosophers and, enjoyable though it be, I'd rather not trespass.
To separate a helium atom into two atoms of deuterium (proton plus neutron plus electron), one must do work W against the attractive forces among the nucleons. Thus the mass of the two hydrogen atoms is greater than that of the helium atom, by W/c2.
To dissemble any nucleus into protons and neutrons similarly requires doing work against the force (called the strong nuclear force, a very good name!) that holds the nucleus together. So the mass of any nucleus is less than the sum of the masses of its components. This difference is called the mass defect.
There is a mass defect when one rearranges molecules, too. The mass output by a fossil fuel power station is slightly less than the mass input. Indeed, the difference would be the same for two stations of equal power output. However, the fuel and the wastes of the fossil fuel station are so much greater than those of the nuclear station that we do not notice the chemical mass defect. See the explanation of mass defect on Einstein Light.
Nevertheless, many people have the confused notion that mass defect applies only to the nucleus. Relativity is (so far as we know) true, so it applies to everything. So a (sealed) battery would have a larger mass when charged than when discharged. The reagents in an exothermic reaction are slightly more massive than the products. In these cases, the energy:mass ratio is so small, however, that the mass defect is hard to measure. Even in nuclear reactions, it's small, but relatively easy to measure. See the explanation of mass defect on Einstein Light.
Today, we talk only about the mass m of an object, and we regard m as constant. If you are travelling in the same frame as the object, and if you apply a force F, it accelerates at F/m. If it is travelling at speed v with respect to you, it has a momentum
The reason why mass dilation is no longer used in discussing relativity is because it is very confusing.
In the original papers, Einstein himself did use the idea of a longitudinal mass that increased with speed, but he seems not to have used it after 1906, and indeed recommended against it. Some old text books use it (some use it in one chapter and discard it in another), some distinguish between longitudinal mass and lateral mass, but happily its use is disappearing. There's a nice discussion of the educational consequences of using it or not in On the abuse and use of relativistic mass.
Now let's suppose that the spacecraft accelerates vertically upwards at acceleration a. Because she is strapped in, she accelerates with the space craft so the total force on her is ma. This force F is supplied by sum of the forces of the seat and her weight, so
If |3W| is the total force, then we have
Now we need to find out wheter the maximum accleration of a space shuttle is closer 2g or 3g. According to one NASA site, "the shuttle goes from standing still on the launch pad to more than 27,359 kilometers per hour (17,000 mph) in just over eight minutes". This gives an average acceleration that is a little under 2g. So over these eight minutes, the astronauts would feel two to three times heavier than normal. However, this is an average: the maximum may exceed this value. Also, the flight is not vertical for all of this time, and so the angles between the forces must be considered in adding them.
As you are reading this, you can probably feel your chair pushing upwards on you with a force of several hundred newtons. If your feet are not touching the ground, this is an upwards force equal in magnitude to your weight (a downwards force). My weight is 680 N downwards, so I know that the force from the chair is about this much, upwards. You can also feel your abdominal muscles holding your abdominal organs in place. These forces and some others give you the sensation of having weight. You do not really sense your weight directly very much, because it is applied homogeneously over your whole body. When the forces from the chair or on your abdominal wall are reduced or zero, you may feel 'weightless'--the feeling you get when a lift starts to accelerated rapidly downwards, or when you go quickly over a peak on a roller coaster. I have put 'weightless' in inverted commas because in these situations, and in an orbiting spacecraft, your weight is virtually normal. Since the moon flights stopped, no human has been far enough from the Earth for his/her weight to be substantially reduced.
The three diagrams below show two situations that produce free fall. In an orbiting spacecraft, the spacecraft and the cosmonaut are both accelerating towards the centre of the earth at the same rate (their centripital acceleration is ac = v2/r, where v is the orbital speed and r the radius). Their weight is what keeps them in orbit: W = mac. Because they are *both* accelerating towards the centre of the earth at the same rate, there is on average no force between the cosmonaut and the spacecraft. This absence of forces from seat, floor, abdominal wall etc is what is commonly but misleadingly called 'weightlessness': the cosmonauts in the space station are not without weight, in fact the have (almost) their usual weight. It's just that they don't feel the force of chairs on their bums and they don't feel their abdomens holding in their organs.
In figure at right, a NASA airplane (nicknamed the 'vomit comet') cuts the power in its engines and, for about 25 seconds, travels in a trajectory that is nearly parabolic. Both the plane and its occupants accelerate towards the Earth at g: all are in free fall. Astronauts are thus exposed to free fall and obtain brief periods of experience in working in this condition.
Physicists tend to use the word 'weightless' in scare quotes (as I have done here), to make it clear that they are not talking about a situation in which there is no weight. Many physicists prefer to avoid the word altogether and talk instead about free fall.
There are some similiarities between the passenger (mass m) in the lift (let's put it at the equator) and a cosmonaut (mass m) in low Earth orbit. The weight of each is about mg. Both accelerate towards the centre of the Earth at approximately g. The difference is that the spacecraft makes a circle around the Earth in about 90 minutes, whereas the lift makes a circle around the Earth in about 24 hours. The acceleration g is just enough to keep an object in low Earth orbit with a period of 90 minutes. It is far too great for the 'orbit' of the hapless passenger in the lift. If a satellite loses speed, it gradually spirals in towards the Earth. The horizontal speed of the passenger in the lift is so low that his 'spiral' towards the centre of the Earth is almost a straight line. (There have been a few 'approximately's and 'almost's in the above. If you are interested in the analysis of motion in the rotating frame of the Earth, have a look at the formal analysis of the motion of a pendulum at the Earth's surface.)
Gravitational PE at a distance r from the earth's centre is given by
where r > radius of the Earth, and U = 0 at r = infinity.
The sketch compares the usual astronomical version (GMm/r, the solid line) and the local version (mgh, dashed line). mgh is a poor approximation for altitudes that are not negligible in comparison with the radius of the Earth.
|
2) For an object acted on by a conservative force, we can define a potential energy due to that conservative force, as a function of position. The difference in potential energy (Ub - Ua) between points a and b is defined as the work done against that force to move the object from a to b. (You can now see why we can't do this for nonconservative forces: if we do a round trip from point b to point b, we do work against such forces and so the potential energy at b would have no unique definition.)
3) Gravity is a conservative force. So, for a body with mass m, we define the difference in gravitational potential energy between points a and b as the work as the work done against gravity to move mass m from a to b.
4) Note that the force may vary with position: the Earth's gravitational field gets stronger as we approach the surface of the Earth, from either direction. So we could say
"The main energy cost associated with space travel is currently fuel to reach Low Earth Orbit. Energy is also needed to leave this orbit, change direction/ accelerate and for communication. What else is there to say?"
But we actually give spend more energy than that, because we have to lift a lot of fuel. Most of the fuel doesn't go very high, but we still have to lift and to accelerate that fuel. So most of the energy goes into carrying fuel, and most of that fuel is there to carry more fuel, and most of that..... Which is why the Saturn V booster is a very big can of fuel.
If we fired the satellite out of a gun (as in HG Wells' "From the Earth to the Moon"), we would burn the fuel on the ground, and therefore not have to lift nor to accelerate it We would therefore need only a tiny fraction of the energy that is currently used. Unfortunately, the huge acceleration would damage the satellite and kill any passengers.
In order to communicate over a long distance, you need to confine the radiated power to a beam of small cross sectional area, in other words to send out a nearly parallel beam. Then you need to intercept it and focus it. For both of these, you use parabolic dish, a little like those uesd for satellite television. Now the whole idea of rays and focussing only works well for dishes much bigger than a wavelength. So, for microwaves, the dish need only be metres in size. For radio waves, larger dishes are required.
For the dish on the ground, this is not a big problem, and dishes like the one at Tidbinbilla near Canberra are used both for transmission and reception. (There is a fun (but scientifically silly) movie called The Dish, which is about the use of the radiotelescope at Parkes--one of the world's finest astronomical instruments--for communication with Apollo XI.)
For the dish on the spacecraft, there are limits to the possible size. So the solution is to keep the wavelength short, the dish on the spacecraft smallish, and the dish on the ground big.
On the Earth, falling small objects quickly reach a speed at which the force of drag equals their weight: their terminal speed. For a bullet, this would be several tens of metres per second, whereas a bullet is usually fired at hundreds of m/s.
For the names, wavelengths, frequencies and uses of the different bands, see the electromagnetic spectrum.
Sensitivity refers to the minimum signal required by an instrument. Larger instruments receive more light, and so (all else equal) are more sensitive. For example, under optimum, dark adapated conditions, your eye requires about 70 photons to form an image (only about 10% of these are captured by photoreceptors).
The sensitivity of telescopes is often limited by optical noise (stray light) or by electrical noise in the detectors. Electrical noise is often minimised by running the detectors at very low temperatures, where the thermal motion of electrons is reduced.
Parallax refers to the different views that you see from two different positions. Try this experiment. Hold the index finger of your left hand vertical, 20 cm in front of you. Hold the index finger of your right hand vertical, 40 cm in front of you. Now close your left eye and, using just your right eye, move the two fingers sideways until they line up. Now close your right eye and open the left. The closer finger has 'jumped' to the right of the further finger. Repeat a few times. Compared to a distant background, both fingers have both jumped to the right, but the closer one jumps father. If you measure the angles through which they jump and the distance between your eyes, you can work out how far away the fingers are.
For distant objects, the distance between our viewing positions must be greater than the distance between your eyes. Fortunately for astronomers, the Earth shifts our telescopes round the sun once a year, so we can get a separation equal to the diameter of the orbit of the Earth (16 light minutes) if we wait six months, as shown in this diagram.
In this sketch, which is not to scale, imagine an observer looking at objects A and B, standing at the pole of the Earth with his head towards us. Now he sees object A to be to the right of B. Six months ago, he saw it to be to the left of B. Now most stars are so far away from us that we cannot observe any relative motion in this way. However, for close stars it is possible. The next sketch shows the path of light from a close object and from a very distant star.
From trigonometry,
One limitation to the angular resolution of telescopes is due to a wave effect called diffraction. When parallel light passes is incident on a circular lens or circular telescope with an aperture a, it cannot be focussed onto a perfect point, but rather makes a small circular smudge called the Airy disc. Around this bright circle is a dark ring, then a series of bright and dark rings: the diffraction pattern of a circular aperture. The angular diameter of the central bright ring is of the order of λ/a, where λ is the wavelength of the light (or other waves). If the angular separation between two stars is smaller than the size of this disc (as is the case for the majority of double stars), then it is very difficult to resolve them as two different stars. (In practice, this theoretical limit is not always achieved in optical telescopes because of such effects as the bending of light in the atmosphere.) So an angle of λ/a is approximately the theoretical limit to the angle that can be resolved by a telescope (or camera, or eye*).
Radio telescopes, which use long wavelengths (eg 21 cm, the wavelength of the 'hydrogen line') have to be much bigger than optical telescopes (L ~ 0.0005 mm), but in both cases, the bigger the better. Optical telescopes may have a of several metres. Individual radio telescopes may have a of 10s of m (eg the dish at Parkes is 64 m), but separate radio telescopes may be connected to provide a bigger effective aperture. The Australia Telescope links radio telescopes across the country to provide an effective aperture of thousands of km.
Space based optical telescopes have the advantage that they have no atmospheric distortion and so they can measure smaller angles than ground based ones.
* λ/a is also one of the limits to the angle you can resolve with your eye. This limit is only achieved, however, when your pupil is almost closed (aperture a less than about 2 mm), in very bright light. In dim light, when you pupil is open wider, the angular resolution is typically a bit better than one minute (1/60 degrees), and is determined by the spacing of photoreceptors in your retina (which you can now work out).
Cepheid variables---measuring longer distances
Fortunately for astronomers and cosmoligists, there is a class of stars called Cepheid variables. These stars have brightness that varies periodically over time. Further, the period T of the oscillation in brightness is related to the total output power of the star. For a large, high power Cepheid variable, the period may be longer than a month. For a small, low power star, it may be days.
Once we know the relation between the period T and the light power P, we can determine how much light power P it puts out by measuring the variation in its brightness. If we know how much light it puts out and how bright it appears viewed from Earth, we can work out how far away it is. Here's how it works:
First, we look at Cepheid variables whose distances are known. Some of them are close enough to allow us to determine their distance r from parallax. From the intensity (I = power per unit area) of light received on Earth, we work out their power P from the inverse square law. Consider a sphere centred on the Cepheid variable. The area of the sphere is 4πr2. All of the power P passes through this area, so the power per unit area is:
Cepheid variables were first studied in the first decade of the twentieth century by Henrietta Leavitt, one of the first women to become famous in astronomy. She studied cepheid variables in one of the clouds of Magellan. There is little proportional difference in the distance from Earth to the stars in this galaxy, so she knew that the different apparent brightness was determined only by the power output. At the time, she could only use the cepheid variables for relative distances, because the parallax method was not sufficiently accurate.
We now know that the Cepheid variable cycle involves thermal feedback produced by the different ionization states of helium, which is relatively abundant in older stars. Doubly ionised helium is more opaque than singly ionised. If the star is hot enough to produce doubly ionised helium, this opaque layer insulates the star, making it hotter still. As the temperature rises it expands, but this expansion cools it, so the helium captures an electron and becomes less opaque, which continues the cooling. Cooling causes it to contract, which raises the temperature, and the cycle continues.
The Cepheid variable method works for distant stars in our own galaxy, and it also works for 'close' galaxies such as the Magellanic clouds. However, for distant galaxies, we can no longer distinguish individual stars. In these cases, various other methods are used. For instance, we can estimate the power of the whole galaxy (eg the brightest galaxy in a cluster) and use that to infer the distance from the inverse square law. One type of supernova (the exploding white dwarf star) provide another method, because the stage at which they explode depends on their size, and so they do not vary much in intrinsic brightness.
For localised objects, the oldest ones have to be very hot so that they are still visible with huge red shifts. And galaxies and stars didn't form for a while, so the oldest visible galaxies are a little younger than the universe.
On the left is a sketch of the experimental apparatus used to observe black body radiation. An object of controlled temperature T contains a cavity, joined to the outside by a small hole. If the hole is very small, the radiation in the cavity comes to equilibrium with the walls. The hole allows a small fraction of the radiation to pass to a spectrometer. On the right is a plot of Planck's radiation law for two temperatures. Note that the wavelength for maximum emission becomes shorter (higher frequency) for higher temperature. Note also the strong dependence on temperature of the total emission. The radiancy is the power emitted per unit area per increment of wavelength and so has units of W.m-3.
Note that the peak of the curve moves to the left as the temperature increases: hotter objects output a larger fraction of their electromagnetic radiation at shorter wavelengths. This displacement of the peak of the curve is called Wien's displacement law. After taking the derivative of Planck's radiation law and setting it to zero, one finds an expression for the wavelength λmax at which the radiation is a maximum. It is related to the temperature T of the black body by the simple equation
AT LOW FREQUENCIES, the wavelengths are long. There are relatively few ways in which standing waves can fit into the box if their wavelengths are long. So, if the energy is shared among the different possible standing waves, we should expect the radiation intensity to go to zero at low frequency and to increase with increasing frequency. This is the classical result.
But now let's add the quantum hypothesis, that radiation energy is quantised and comes in photons that have an energy hf. In thermal equilibrium, the atoms and electrons of the wall have thermal motion. At sufficiently high frequency, few atoms or electrons will have enough energy to emit a photon with energy hf. Note how this depends on the quantisation hypothesis: if you could have radiation with frequency f and an energy less than hf, then the walls would emit high frequencies like crazy. They don't because they have to emit only whole photons.
So, at low frequency, the long wavelength limits the number of possible photons. At high frequency, the high energy makes emission unlikely. So the distribution goes to zero at both f = 0 and f = infinity, and has a maximum in between.
Incidentally, the frequency for maximum energy is proportional to the temperature of the walls. Wien's Law. The typical energy of any motion is kT/2, where k is Boltzmann's constant and T is the temperature. (k = R/NA where R is the gas constant and NA is Avagadro's number.)
Here are some technical links from
Uni of Virginia and the
Heriot Watt University. There's a nice essay by by G. Pattison, a high school student, at Black body radiation. Some pictures taken with a thermal imaging camera are at Thermal physics and how clothes work.
See also The electromagnetic spectrum.
Max Planck used quantisation to explain black body radiation. His biography, courtesy of the Nobel Foundation, has some of the details. Planck didn't like statistical mechanics and Boltzmann's work. Reluctantly, however, he found it necessary to use it. If the energy could only be emitted and absorbed radiation with energy amounts that were inversely proportional to λ (E = hν = hc/λ), then Boltzmann's equal distribution of energy among different modes would give the required short wavelength behaviour.
The next step, the idea that energies of radiation were inherently quantised (a little like the way matter is quantised in atoms etc), came from Einstein's analysis of the photoelectric effect, which is our next question.
A clean metal surface, often in vacuum, is exposed to light whose wavelength or frequency may be controlled. A nearby electrode can receive the emitted electrons and so allow a current, which may be measured. A variable potential difference may be applied to stop the current. For a given metal, there is a minimum frequency fo (ie a maximum wavelength) of light that will cause electrons to be emitted. For light (or UV radiation) with higher frequency, the stopping voltage increases linearly with frequency f, as sketched. More reactive metals have lower fo.
This phenomenon was investigated experimentally by Philipp Lenard and then later and more precisely by Robert Millikan. Both received Nobel prizes for the work. The photoelectric effect was explained by Albert Einstein in work for which he received the Nobel prize in 1921.
Photocells come in different types. Some are photovoltaic cells, some are phototransistors. In a phototransistor, the base of the transistor is exposed to light. Because photons can produce electrons in this region, the input light effectively replaces the base current (the input current in a normal transistor). So the output current of the transistor is determined by the light input. (See the photovoltaic effect and transistors.)
One type of breathalyser uses a chemical reaction involving alcohol to change the colour in an indicator. (The only one that I've ever been able to examine worked that way.) Some use the infrared spectrum of alcohol. They work like this:
Light of known spectrum passes through the breath and then through optical filters that respond to different wavelengths onto a photocell. Thus (part of) the spectrum of chemicals in the breath, including that of alcohol, if present is measured.
The photovoltaic effect is involved in the photocell. Go to Solar cells and the photovoltaic effect.
Now in a string, the wave is in the displacement of the string. What is it that waves in an electron wave? The quantity is called ψ, the Greek letter psi. ψ is a complex quantity – it has real an imaginary components. If you take ψ at any place and multiply it by ψ* (which is like ψ, but has the opposite imaginary component, you get the probability of interacting with the electron at that point. So the atomic nucleus is pictured (especially in chemistry text books) with clouds of 'electron probability' around it in different orbitals.
The interpretation of ψ as a function of position and time is a subtle question. In the case of the atom, ψ is a standing wave. In other cases, like the ψ for electrons in a cathode ray tube, ψ has the form of a travelling wave.
Is an electron a particle or a wave? Another subtle question. It can have wave properties (eg a wavelength) and particle properties (eg a position). However, it cannot be a 'good' wave and a 'good' particle at the same time. Because of Heisenberg's Uncertainty Principle, a precise measurement of the wavelength implies a poor measurement of position, and vice versa. So an experiment in which wavelength is controlled precisely will give you wave effects (such as interference), but the electrons will not be localised in space. Conversely, if you constrain the position, you have an uncertainty in the momentum and wavelength. Some people say that little things like electrons are 'wavicles'. I prefer to say that they are neither wave nor particle, and that these macroscopic ideas are misleading when applied to electrons.
We usually talk about the resolution of microscopes, and the resolving power of diffraction gratings and spectrometers. The resolution of a microscope is limited by the wavelenght it uses. The wavelength of light is typically half a micron, so the optical microscope cannot do much better than this size (though clever techniques can improve it noticeably, eg confocal microscopes.)
The wavelength of the electron beam is inversely proportional to the momentum of the electrons, and therefore inversely proportional to the square root of the accelerating voltage. By using high voltages, one can make the wavelength of atomic size or even smaller. Electron microscopes use this principle. Here is a link to some cool pics taken with a scanning electron microscope. It also has links to how it works.
(time taken to measure f) times (error in f) is about one or greater.
Now in quantum mechanics and atomic physics, the energies of photons are hf, where h is Planck's constant. So in order to know the energy, you have to take a certain time to measure the frequency. Mutliplying our previous inequality by h on both sides gives us
(uncertainty in energy) times (uncertainty in time) is greater than about h.
Werner Heisenberg won the Nobel prize in 1932.
Because h is so small (6.63 10-34 Js), the consequences of the uncertainty principle are usually only important for photons, fundamental particles and phonons. There are, however, many physical processes whose evolution with time depends sensitively on the initial conditions. (Sensitivity to intial conditions is fashionably called chaos.) The uncertainty principle prohibits exact knowledge of initial conditions, and therefore repeated performances of such processes will diverge. (Physicists will also tell you that one cannot have exact knowledge anyway, for a variety of practical reasons, including the fact that you don't have enough memory to record the infinite number of significant figures required to record an exact measurement.)
Some philosophers regard the consequences of the uncertainty principle as having a more fundamental importance. The argument goes like this: if one could know exactly the position, velocity and other details, one could, in principle, compute the complete future of the universe. Since one cannot know the position and momentum of even one particle with complete precision, this calculation is impossible, even in principle. Most scientists find this a trivial argument. A memory capable of storing all this information would be as complex as the universe, and then the contents of that memory would have to be included in the calculation, and that would make the amount of information greater, and that information would have to be stored..... We rather point out that all of that information is actually contained in the universe which, as an analogue computer, is computing its own future already.
(See also Heisenberg's uncertainty principle and the musician's uncertainty principle, which has some demonstrations.)
The important consequence of this is that 'electron shells' in the atom become filled. Once the innermost orbital is filled by two electrons having the same quantum number except for spin, then no further electrons can have that same state. So higher energy orbitals become occupied. This gives rise to chemistry.
Wolfgang Pauli won the Nobel prize in 1945.
Imagine a charge in a circular orbit around a charge of the opposite sign. This moving charge can be considered as a current, so we now have a little electromagnet, with a magnetic dipole moment μ, as shown in the diagram. When we calculate its energy, we would put in an electrostatic term and a kinetic term, as in the Sommerfeld-Bohr picture. (See notes on Bohr and the hydrogen spectrum.) These terms do not depend upon the direction of the motion and, in the absence of an external magnetic field, there is no term for magnetic energy. However, if there is an external magnetic field, then there is another energy term: the magnetic potential energy will be lower if our electromagnetic is aligned with the field, and higher if antiparallel. So this last term has different sign for charges moving in clockwise or anti-clockwise direction. Physicists say that the energy level is *split*. New energy levels means a new spectrum. This splitting is explained by the simple classical model given here, and is also explained by the quantum mechanical model. However, the Rutherford atom, which did not have either electron orbits or wave functions, does not explain it. See the links on the Zeeman effect from CERN, from Thomson Learning, and from John Hopkins University.
Chadwick (see links below) is credited with discovering the neutron. Without neutrons, hydrogen would be the only stable element, which would simplify chemistry considerably, but simultaneously eliminate chemists. The neutron is not only important for holding the nucleus together, but also as a projectile for smashing into the nucleus (it is not deflected by electrostatic forces). It can be used as a probe of the nucleus. Neutron bombardment is important in chain reactions.
See Chadwick and Fermi courtesy of the Nobel foundation.
Fermi and Dirac devised the statistics to describe particles (like the electron) that are subject to Pauli's exclusion principle: they cannot have the same quantum numbers. Such particles are called fermions (cf bosons). In the atom, these statistics explain the distribution of electrons among different states or orbitals.
These statistics are also fundamental in solid state physics and thus to its applications in the electronics and computing industry.
Here is short explanation of the Bohr-de Broglie-Sommerfeld model and the expression for the lines of the hydrogen spectrum. I repeat my caveat: I don't know much about the history of who did what when and am hoping to obtain assistance from a historian. However, the derivation below is worth reading because it is a good example of physics. It uses a few physical postulates, which are based on generalisations of experiments. It is mathematical and quantitative. It delivers quantitative predictions that may be readily compared with new experiments. This derivation requires only simple algebra, so it can be followed by high school students.
In the 19th century, a purely empirical formula was known for the hydrogen spectrum: it gives the frequency f of wavelength λ of EM radiation either absorbed or emitted by the hydrogen atom. It is
Bohr's name is associated with the 'solar system' model: the proton is so massive that its motion is neglected (cf sun) and the electron 'orbits in circles' (cf planet). I have put the scare quotes on 'orbits in circles' because we now know that this phrase is not merely untrue, but doesn't have a meaning for an electron.
In this classical model, the electron energy E is kinetic plus electric potential
v = nh/(2.π.r.m) (3)
The sketch at left shows the relative radii of the 'orbits' for the first four values of the integer n. The diagram at right shows some of the possible transitions from higher energy level to lower.
Now if N = 1, the set of lines with different n are called the Lyman series. N = 2 gives the Balmer series and N = 3 the Paschen series. (Our historian will explain why.) For example, the lowest energy (longest wavelength) in the Paschen series has
So, when we find that an electron has angular momentum and a magnetic dipole, it is natural to talk of its spin. Natural, but somewhat misleading, because on the very small scale one must use quantum mechanics, rather than classical mechanics. Like the energy of electrons in an atom, the spin of a fundamental particle is quantised: only discrete values are allowed (+ and - 1/2 for the electron). Further, if one imagines the electron is a little ball of spinning charge and applies classical physics, one gets the wrong answer for the magnetic dipole.
So why is it a necessary concept? If we apply an external magnetic field, the energy of an electron will be increased or decreased depending on the direction of its magnetic dipole (and thus on the value of its spin). It also gives an extra quantum number. The Pauli exclusion principle forbids electrons to have the same quantum numbers so, for any energy level in the atom, there can be two electrons, with positive and negative spin. Thus spin allows twice as many electrons, which has very considerable consequences for the periodic table and chemistry!
To step back in history: Rutherford was able to probe the inside of the atom by using 'probe' particles smaller than the atom. From the angles of recoil, he was able to make an important conclusion about the atom: that nearly all the mass was localised in a very small region (the nucleus).
The second reason is to do with mass-energy equivalence. If you organise a collision that, relative to the centre of mass of the colliding particles, has a kinetic energy E, it is possible to create a particle-antiparticle pair, provided that 2mc2 is less than E. Smashing an accelerated particle into a target is therefore one way to study structure. One drawback is that the kinetic energy in centre of mass frame is not very high, particularly when relativistic effects are included.
A better way of studying subatomic particles is to smash a proton (mass mp) into an antiproton. Sometimes they destroy each other, and then you get the kinetic energy of the collision, plus 2mpc2. Which may be enough to create various different particle-antiparticle pairs. Note that you usually create (or destroy) particle-antiparticle pairs, so that the total charge and spin of the things you create or destroy is zero.
p-type semiconductors are 'doped' with a small percentage of atoms that have the capacity to accept an extra valence electron. This can be thought of as an electron hole. Further, the hole can move: if an electron from a neighbouring atom enters the hole, it leaves a hole next door, so the hole appears to have moved. And since it was a negatively charged electron that moved to fill it, the movement of the hole is effectively the movement of a positive charge. In p-type semiconductors, one can think of the current being carried by positively charged electron holes, moving in the direction of the electric field (see drift velocity).
This is convenient as a way of thinking, although one can say that it is really the electrons that are moving. Here is a good analogy: take a sealed bottle of water and invert it. You will see the bubble of air move upwards through the water. Now of course you know that what is really happening is that water is flowing down into the bubble and leaving a hole in the water where it has come from, so really what you are watching is water motion. But because there is lots of water and only a small bubble, it is easier to think of a moving 'hole' in the water than to consider the motion of the water.
A junction transistor consists of a thin layer of one type of semiconductor sandwiched between two layers of the other type, as shown in the schematic diagrams above.
Let's look at the npn transistor. An electron can travel from the emitter (n doped) to the collector (n doped) only if it can get through the base without 'colliding' with a hole in the base (p doped). If the layer is very thin, that is possible, but the chance of an electron getting through is a sensitive function of the potential difference between base and emitter (called the bias voltage). For typical silicon transistors, if you set the base-emitter voltage at 0.2 V, there is hardly any current between collector and emitter. If you set it at 0.6 or 0.7 V, you get close to maximum collector current (the size depends upon the size and packaging of the transistor, but tens or hundreds of mA is typical). If you set the base-emitter voltage above this, you have an ex-transistor.
The pnp transistor operates similarly, except that it is the 'holes' that migrate across the thin base region, and the electrons in this region that control the flow. It is convenient to have symmetrical transistors (npn and pnp) for circuits with positive and negative supplies.
The field effect transistor or FET is simpler than the junction transistor. We show a p-gate transistor. The current through the n-doped material passes through the narrow section where it passes the 'gate' of p-doped material. The effective width of this passage can be made thinner or thicker by varying the voltage of the gate, and so removing conduction electrons from the thin passage. An advantage of FETs is that the input resistance of the device is high, which is what one usually wants when amplifying a small signal. In junction transistors, the input resistance is low. A disadvantage of FETs is that they usually can handle only small currents.
Circuit diagrams for a simple amplifier (left) and a logic gate that performs the NOR operation.
Logic operations. In the ciruit at right, let's consider digital signals, ie voltages that we consider only as 'high' (1) or 'low' (0). If either A or B is high, there will be a high base current, so the transistor will be on, lots of current will flow through the load resistor at right, so the output voltage will be low. A and B both high gives the same result. The only way to turn the transistor off (and so obtain low collector current and high output voltage) is if neither A NOR B is on. For that reason, this circuit is a NOR gate. Its output is 1 if neither A NOR B is 1, and 0 otherwise.
One can also use transistors and resistors to construct simple gates for the logical operations NOT, AND, OR, NAND (= not and) and XOR (exclusive or). Alternatively, one can construct any of these from a combination of NOR gates. By putting together combinations of gates, one can easily construct memory, arithmetic and more complicated operations. Although this material is usually taught in a laboratory course normally taught in second year, it is possible to do some experimental work yourself, if you are keen. The School of Physics at UNSW has an interactive course in digital electronics set up as a series of panels in the corridor. By manipulating knobs and switches you can do a series of experiments, starting with resistors and transistors and ending with the elements of a computer--it is a small, self-contained experimental course in digital electronics. Small groups would be welcome to visit by arrangement (info@phys.unsw.edu.au).
Another possibility is the Nobel prize website. The 2000 prize in Physics was awarded to Alferov, Kroemer and Kilby for contributions to the foundations of much of modern
electronics and infomation technologies. The resources on this site range from the somewhat technical (each awardee gets to write a
scientific article for reviews of modern physics on their work), through to their acceptance speeches and so forth, right down to the
basic including graphics-heavy descriptions of their work, and even an online game for kids to learn about the various prizes (see
, which relates to Kirby's work on the integrated circuit).
The more recent the prize, the more public stuff they have on the website, but you can go right back to Bardeen, Brattain and
Shockleys prize in '56 for the transistor, and some of the greats like Einstein, Heisenberg and co.
Roughly speaking, molecules with lots of thermal energy escape from their neighbours and evaporate. Those will little energy stay close to the same neighbours (solid). Those with intermediate energy can exchange neighbours but not escape from them (liquid). Finally, if molecules or atoms have enough thermal energy, their collisions may remove electrons and form a plasma. So:
The ranking above neglects pressure (and only considers entropy implicitly). If we make the pressure low, we can get gases at very low temperature. Solids will sublime at low pressure (CO2 does at atmospheric pressure). Very wide ranges of pressure and temperature can deliver some exotic states of matter:
Bose-Einstein condensates. At very low temperatures and pressures, Bose-Einstein condensation is sometimes possible. This occurs when the wave properties of atoms or molecules become important. The Heisenberg's uncertainty principle imposes a constraint here: we need the thermal momentum so slow that the error in position of the atom/molecule is large enough for several of them to be superposed. What happens then depends on the overall spin of the atom or molecule. If they are bosons, then one may have several or many atoms all with the same quantum numbers and all confined to the same region in space. They are indistinguishable. There's a question about them below.
Now the wavelength of a massive particle (the de Broglie wavelength) is inversely proportional to its speed. And the speed of an atom goes as the square root of its temperature. At room temperature, the wavelengths of atoms are so small (smaller than the atom itself) that interference effects are negligible. When you cool them enough, however, their speeds become slower and their wavelengths longer. And, if you can confine a number of similar atoms in a small volume, they can begin to interfere.
This (roughly speaking) is what happens in a condensate. If you have a collection of particles with the same quantum numbers, and if they are 'in the same place', then they are indistinguishable. This indistinguishability leads to some peculiar statistics, first worked out by Bose and Einstein.
We don't notice this at normal temperatures, because atoms are not 'in the same place', which here means being close enough and having long enough de Broglie waves for interference to occur.
news item on BECs
Electrons in normal metals occupy a set of quantum states, up to some maximum energy (called the "Fermi energy"). The relatively free "conduction electrons" (those which come from the unpaired valence electrons of the atoms) interact strongly with the positively charged ion cores, and as an electron moves through the lattice it will cause the cores to be displaced from their equlilibrium positions. Electrons with energies near the Fermi energy are able to change their quantum state relatively easily, and thus any interaction, such as with the lattice, can result in a drastic change in the quantum states of these electrons. It happens that in superconductors the electrons near the Fermi energy become highly correlated, forming a macroscopic coherent quantum state with exotic properies. This state can be thought of as being made up of "electron pairs", but it is important to understand that these pairs are transitory things which change continuously in a dynamical way. At any instant a given electron is a member of many pairs.
Does superconductivity qualify as a new state of matter? Not according to the classification scheme proposed here. If we had a scheme that made metals a different state of matter from materials that don't share electrons, then under such a scheme superconductors might be a new state of matter. However, this is all taxonomy and semantics, and is not of great importance to physics.
(The answers in the phonon and lattice distortion section were supplied by Prof Jaan Oitmaa.)
When a magnet is brought near a superconductor, this exclusion of the magnetic field distorts the field lines, as shown in the diagram below.
There are some further subtleties in magnetic levitation discussed in www.calpoly.edu/~rbrown/levitation.html.
In order to be quantitative about it, we observe that the smaller the gap between the object and the superconductor, the larger the magnetic field in the gap: the field is compressed in this gap. We can calculate the magnetic pressure (a force per unit area), which at any point is equal to the energy density (energy per unit volume) of the magnetic field. This is given by:
To estimate the size of this effect, let us consider a cubic iron bar magnet with side a = 1 cm. Let the mass of the magnet be 0.01 kg, and let the field near the pole, in the absence of any superconductor (picture at left), be Bo ~ 0.01 Tesla. Now let us put the cube at distance x from the superconductor. To do the geometry properly is a little difficult, but we can make some approximations.
In the diagram at left, the field from the pole diverges over a distance comparable with the size of the magnet, ie over ~a. In the middle diagram, it diverges over the distance x, so the field lines have been concentrated by a factor of about a/x, so the field between the magnet and the superconductor has a field strength
So the obvious question is why you cannot levitate in the second case. Well, the problem is that the symmetry of the two magnets is unstable and you will need to supply horizonatl forces to keep them upright. With some ingenuity, you may be able to supply such horizontal forces without much vertical force. If you do, you can measure the distance at which the upper magnet is supported, and this is roughly twice the distance at which it would levitate over a superconductor. It is also possible to supply the horizontal forces using an array of magnets, and some 'executive' toys use this principle to levitate permanent magnets.
By the way, this trick of using two similar magnets in symmetry would also be an easy way to calculate the forces, rather than to estimate them as we have done above. We treat the superconductor as a 'magnetic mirror' and calculate the 'image forces' due to the mirror image of the magnet.
The Meissner effect is a complex quantum phenomenon. It is due to the fact that electrons in the superconductor are in a quantum condensate described by a collective wave function for all electrons. For an introduction to quantum condensates, see Bose Einstein condensates.
This wave function has a very special property of rigidity: it requires a lot of applied energy to change the 'shape' of the wave function. (To discuss this correctly, we really do need complex mathematical operations, so this simplified discussion is not quite correct. and the shape we are talking about is a shape in Hilbert space.) As a result of this rigidity, the curl of the electric current density is proportional to the magnetic field. (The curl is a mathematical way of representing properties of the shapes of lines in a vector field, the current density in this case. It is a measure of the amount of twist in the lines.) Let us assume now that the magnetic field penetrates inside volume of the superconductor. Hence, because of the property mentioned above, the field induces currents inside the volume of the superconductor. Any current is related to some internal movement and hence to the kinetic energy related to this movement. Therefore such a state would have a very high energy. To minimize the energy the superconductor develops currents on the surface
in such a way that they exactly compensate the magnetic field inside. In
this case current flows only within thin 10-8m surface layer
and therefore energy of the system is relatively low.
This explains the mechanism of the Meissner effect.
The Meissner effect is very important for condensed matter physics and it
is equally important for elementary particle physics. The masses of the
particles arise due to the effect that is very similar to the Meissner
one because the physical vacuum is in the state similar to the state of
a superconductor (Meissner state). However at early stages of the
history the Universe was very hot, the vacuum was in the "normal"
(non "superconducting") state, and all the particles were massless.
(The answers on levitation and the Meissner effect were provided by Prof Oleg Sushkov.)
First, if you have higher conductivity, you can make thinner conducting elements. That makes chips physically smaller, signals travel smaller distances and that improves speed.
Second, with higher conductivity, you produce less heat. Getting rid of heat limits the size and sometimes the clock speed of processors. If you reduce that problem, you can pack large arrays of circuits, again making distances shorter. Instead of having two dimensional circuits (all the elements located at or very near a single, plane surface of semiconductor), you could have three dimensional ones (many parallel interfaces in a block), and that would much reduce the distances and increase the maximum number of components.
Then there are RC charging times. The junctions of a transistor have a capacitance C that is charged via the conductors (with a resistance R) leading to them. To charge a capacitor takes typically a time RC (ohm*farad = second). So low R means a faster switching time. See RC filters for more information.
By the way, a current set up in a superconductor circuit keeps flowing until you do something to stop it. This could in principle be used as a memory element.
There are also potential applications to motors and generators. The stationary magnets or stators in these devices are often electromagnets, in order to save weight or initial expense. (See Motors and generators for details.) If we could easily make these electromagnets superconducting, then we would save on the electrical power required to keep current flowing in them to maintain the field. This would make them more efficient. However, the insulation required to retain the liquid helium necessary for high current super conductors makes such systems large, heavy and expensive.
Power engineers dream of using superconducting cables for the transmission of electricity through power grids. About two thirds of the power generated by power stations is lost in the distribution network, including ohmic heating of the transmission cables. However, the prospect of cooling the distribution network is daunting. 'High' temperature superconductors (those that superconduct at liquid nitrogen temperatures) are not (yet) suitable for transmission cables because they cannot transmit high current density, they are usually not ductile or flexible.
There are however some actual applications of superconductors. One relatively common example is in the large electromagnet used for the constant field in Magnetic Resonance Imaging (see MRI). Liquid helium cools the wire coils to allow the large currents required to maintain the large, uniform magnetic field efficiently. An interesting problem arises when one wishes to turn off the field and bring the magnet back to room temperature. The magnet is a large inductance (see AC circuits) with value L, carrying a large current i, and thefore storing an energy Li2/2. When the temperature starts to rise, resistance appears in the coils and the ohmic (i2R) heating would quickly dissipate all this magnetic energy as heat. Despite the presence of the liquid helium, there are difficulties in disposing of this energy safely, so the current must be gradually reduced to zero before the coils can be warmed.
A photon has (is?) an electromagnetic field. If it has sufficient energy, it can interact with an electron and remove it from its atom. This is called ionizing radiation.
Such radiation can, in sufficient doses, kill cells because, if it strikes DNA in enough places, it disrupts the molecule and prevents reproduction.
In industry, radioisotopes are sometimes used as 'tracers': label a chemical by making one of its atoms radioactive, and you can trace where that chemical goes. They are also sometimes used to measure the composition of materials by measuring the amount of radioactivity absorbed. Radioactive tracers are used to identify organs and pathways for different chemicals. Positron Emission Tomography is also used to identify the distribution of different substances. (See also Medical physics.)
Radioactive isotopes are used in some measurement devices. The domestic smoke detector is the most common: they sometimes use Americium. Radioactive sources are also used to measure the thickness and composition of thin films (suitably calibrated, a measure of transmitted radiation tells you how much material was present to absorbe the radiation.
13C has been widely used (ie sufficiently widely that I've heard of it) in plant physiology to study the carbon cycle and photosynthesis. If you put 13C into a particular sugar and you find it in starch, then you know that there is a pathway from that sugar to starch etc. Often when researchers study biochemistry they choose a radioactive isotope to 'label' the biochemical in which they are interested. Then you can measure the concentration by measuring the radiation, and even better you can trace where it has come from. (Biologists would have better examples.)
As to agriculture, the link below reports the use of 15N to study root biomass and uptake, 13C for carbon exchange and 137Cs for studying soil redistribution.
As the temperature gets higher, the typical energy due to thermal motion (kT, where T is the absolute temperature and k is Boltzmann's constant) gets greater. When it becomes comparable with the ionization energy of hydrogen (kT ~ 10 eV), nuclei can no longer hold on to the electrons and so there are no longer atoms, just a plasma.
Keep going back in time, hotter and hotter. Eventually the protons are colliding with each other with the same sorts of energies produced in the big atom smashers (kT ~ GeV), and so all of the weird particles recorded at e.g. CERN are now present in the tiny, dense universe. To understand cosmological evolution at this stage, one needs to understand about these particles.
Hotter and hotter, and the difference between the different forces ceases to be evident: first electromagnetism and the weak force, then the strong force all become one (in the language of cosmology, they have not yet 'frozen out'). Some theoreticians think that in the early universe quantum gravity was as strong as the other forces.
Finally one gets back to the Planck length ((Gh/c3)1/2 ~ 10−35m) and the Planck time ((Gh/c5)1/2 ~10−43s), over which the spontaneous creation of transient black holes dominates proceedings. On this scale time and space cease to have a separate meaning, or even a meaning at all. Without time and space, cause and effect cease to have a meaning and so one of the Big Questions (How did it all start?) disappears. (The last sentence may seem glib, but I think that it is actually quite important and profound.)
If neutrinos are massless and travel at the speed of light, then they can be classified as having positive helicity or negative helicity. The neutrino has the same helicity as a normal screw: you turn clockwise and it goes away from you. The anti-neutrino has the helicity of a left-handed thread. See http://230nsc1.phy-astr.gsu.edu/hbase/particles/neutrino3.html for diagrams.
Note that it is only possible to make this distinction if neutrinos travel at the speed of light. If they have mass and travel at less than c, then it is theoretically possible to overtake one. Viewed from a spacecraft travelling faster than the hypothetical sub-c neutrino, its helicity would be reversed.
Do neutrinos have mass? We don't know yet, although if they do, it must be VERY small. Nevertheless, there are so very many neutrinos in the universe, that even a tiny mass for each one could turn out to be a substantial fraction of the total mass of the universe, which has important implications for cosmology and the future of the expansion of the universe.
A neutron consists of "up" and "down" quarks: n = (udd).
CT
Computed Tomography (CT) provides a 2D image of a thin cross-sectional slice so that structures within the body are clearly separated. They are produced by combining 2D projections taken from a number of angles. X-ray CT gives good images of the anatomical shape and structure.
PET and MRI also give images of thin cross-sectional slices. The images in fact resemble X-ray CT images and employ similar computational methods to convert detected signals to images but that is where the similarity ends.
PET
Positron Emission Tomography (PET) requires an injection of positron-emitting radioactive tracer (eg 18F, 11C, 13N, 15O) into the patient. The positrons anilhilate with electrons in the body to produce pairs of gamma rays. So the gamma rays used to produce the image arise from inside the patient. PET gives images related to biochemistry, metabolism and function, not just anatomical structure.
MRI
Magnetic Resonance Imaging (MRI) is based on the way certain atomic nuclei respond to radio waves while in the presence of a magnetic field. MRI images can show great detail, even in soft tissue and is often used to image the brain and spinal chord. It is also able to show function and biochemistry. And while all the above imaging techniques are non-invasive and safe, MRI causes the least damage to tissue.
Measurement of BUA uses a broadband ultrasound pulse made up of waves of many frequencies between 200kHz and 600kHz. BUA is determined by sending the broadband pulse through the bone and measuring how much of each frequency is absorbed by the bone. The absorption spectrum depends on the structure within the bone.
*NB. The heel bone is used because(i) of the type of bone it contains (trabecular bone) (ii) it has flat parallel surfaces and (iii) it is a weight bearing site.
The fibre diameter is one property of the optical fibres, which can affect the endoscope image. Images are transmitted through a bundle of up to 300,000 fibres with their ends bound and polished. Each individual fibre transmits the intensity and colour of one point (pixel) in the image. The resolution of the image therefore depends on the diameter of the fibre. So smaller fibre diameters produce images with finer resolution.
4. Explain that large differences would occur in the relaxation time between tissue containing hydrogen bound water molecules and tissues containing other molecules.
5. Describe the changes that occur in the orientation of the magnetic axis of nuclei, identify data sources, gather, process and present information using available evidence to explain why MRIs can be used to
Return to top of page and menu
ii) AC allows the supply to be transformed to lower voltages, and therefore allows its safe use in a range of small appliances that operate from low voltage AC and DC. Without AC, these might be operated by batteries, which are inefficient and whose production and disposal pose environmental hazards.
iii) Without any sort of electricity generation, mechanical power might be distributed via shafts and pulleys, usually only within one site. This concentrates and limits the use of mechanical energy.
iv) AC and DC generators produce ozone in small quantities. Released into the atmosphere remote from people, this probably does little damage.
v) AC generators radiate low frequency EM radiation. The eddy currents thus generated may some effects, though these are poorly understood, except in obvious cases like their effect on cardiac pacemaker circuits. The very high voltage transmission lines, which are permitted by AC, can create large electric fields. In areas where there are naturally ocurring radioactive gases (radon), ions may be trapped in these fields, and this may have a health effect. Likewise, charged dust particles trapped in these fields might have an environmental health effect (via allergies?).
vi) The very high voltage transmission lines, which are permitted by AC, have much weaker magnetic fields than would low voltage DC lines transmitting the same power, so eddy current losses and other magnetic effects are much smaller.
Indirectly, one could say that the equation that you quote defines the ampere. So many electrical measurements on motors depend indirectly on this equation.
or Vout/Vin = Iin/Iout
Discuss why some electrical appliances in the home that are connected to the mains domestic power supply use a transformer
Many appliances require low voltage for solid state electronics (transistor circuits typically require 5 - 20 V DC) or for safety. (eg a low voltage electric toothbrush is safer than a 240 one).
Some appliances require high voltage (neon lights, the cathode ray tubes in TVs and computer monitors).
The usual approach is to use a transformer to get the voltage to the appropriate value, and then (if required) a rectifier, capacitors and regulator to convert AC to DC. (There are also DC voltage converters that 'chop' the AC on and off, allowing capacitors to charge to the desired voltage. These do not need transformers.)
Eddy current heating is never completely overcome, and transformers are usually at least a little warm. (In Sydney, cockroaches like to live on or near transformers because they are warm.)
One could consider the coil in the cooktop as a primary and the metal in the saucepan as the secondary of a transformer.
As the heat is produced directly in the saucepan itself, less heat is wasted in the cooktop or the air.
There is a further subtlety. Let's compare a pot made from aluminium , which is non-magnetic but has low resistivity, with one made from iron, which is magnetic but has rather higher resistivity. If the two were placed in the same magnetic field varying in the same way, the Faraday emf would be the same. The ohmic power loss in the two would be given by V2/R, where V is the Faraday emf and R the resistance. R would be lower in the Al, so the ohmic power loss would be less.
However, there are three complications to this argument. First, the magnetic field in the Al pan will be less, because it is non-magnetic. This makes it a less effective transformer, just as an air-cored transformer is less effective than an iron cored one in low frequency applications. (Technically, we say that the magnetic permeability Al is much lower than that of Fe. The ratio is a few thousand times, so the effect is large.) Second, a higher secondary current actually decreases the magnetic field, because it provides a back emf in the primary. And thirdly, ohmic losses are not the only losses. Energy is also lost in magnetising and demagnetising the iron each time (technically hysteresis losses). So for these three reasons, the heating in the iron pot will be greater than in the Al pot.
Eddy current braking is used on some trains: the train carries a powerful electromagnet positioned near the rails produces eddy currents whose magnetic field opposes the motion of the magnets on the train. It has the advantage over mechanical brakes of having no pads to replace and being silent and smooth. It is of course more wasteful than regenerative braking.
Eddy current braking has the feature that it is strong at high speeds and weak at low. This is sometimes a disadvantage (e.g. for parking!) so mechanical brakes are required as well. This feature makes eddy currents useful for damping: to stop unwanted oscillation in things like balances.
The switch contains an oscillator using a resonant circuit operating at radio frequencies. The coil sets up a high frequency magnetic field. When a conductor enters this field, eddy currents flow in the conductor. This can have two effects. First, it changes the impedance of the coil, and so changes the frequency* of the resonant circuit. Also, because of ohmic losses in the conductor, energy is lost from the resonant circuit into the conductor. This may be large enough to stop* the oscillations. One or other of these changes then activates the switching circuitry. The actual switching of a high current circuit might achieved by a solid state device (MOSFETs, TRIACs etc) but for high current applications a relay is often used.
* You can think of the coil and the conductor as the primary and secondary of a transformer. The resistive load of the secondary is 'reflected' into the primary circuit, changing its properties. Note that both the frequency and the Q value of a resonant circuit depend on the resistance of the coil.
The deceleration due to magnetic braking is limited by (i) the strength and size of the magnetic fields available, (ii)
The electron beam in a CRO can be deflected left-and-right or up-and-down by electric fields. These are produced between two metal plates which have a potential difference supplied by the horizontal and vertical amplifiers.
In the timebase mode, a voltage which increases linearly with time is generated and input to the horizontal amplifier: this sweeps the beam smoothly from left to right. (It then starts again from the left, at a time determined by the trigger, so the timebase waveform is actually a sawtooth. You never see this waveform on the screen, however: it's used to drive the beam.)
Without any voltage input, the timebase thus causes the oscilloscope to "draw" a horizontal line across the screen. The speed of 'drawing' the graph depends on the timebase settings, sometimes called the SWEEP SPEED.
This control knob (usually towards the right) sets the time axis so that that one division represents the time interval indicated around the dial from seconds (fully counterclockwise) down to microseconds (fully clockwise). If you set this knob fully counterclockwise, the beam will sweep so slowly across the screen that you can see it cross. Fully clockwise and the sweep will look like a continuous line, because your eyes are not fast enough to see the motion.
Let us now measure position with respect to this equilibrium position, using the new variable y. For instance, we might apply a steady force F vertically and achieve a new equilibrium at position y (fig c) where
= (1/2)kx2 - kxy + (1/2)ky2 + mgy
= (1/2)kx2 + (1/2)ky2
So, from both the Newtonian and Hamiltonian points of view, the mass on the spring in the gravitational field behaves exactly as it would in the absence of gravity, except for the altered equilibrium position.
Why v depends on E:
In steady state, the speed is (approx) proportional to the force moving it, which is Eq. If we increase the electric field (eg increase the voltage applied to a fixed length of conductor), then the force causing the charge carriers to move is greater, so they achieve a higher drift velocity before the 'driving force' is equal to the 'drag force' due to in the conductor. The driving force is (approx) proportional to the force moving it, which is Eq. If we increase the electric field (eg increase the voltage applied to a fixed length of conductor), then the force causing the charge carriers to move is greater, so they achieve a higher drift velocity before the 'driving force' is equal to the 'drag force' due to interactions with the medium. (More formally, we would normally do this in terms of average speeds and consider the acceleration in the field and the regular collisions with the atoms in the material.)
Notice that the drift velocity does not depend explictly on the geometry of the conductor in question (its area and length): if the cross sectional area were larger, and if we kept the electric field the same, then more charge would move at the same speed and we would get a larger current. If we made the conductor longer, but kept the field constant, then we would have a larger voltage applied to a higher resistance sample and the current would be constant, the same number of charge carriers would pass a given point per unit time at the same speed.
Given the drift velocity, the material and the geometry, we can then work out the current. This is proportional to the number of current carriers available per unit volume, their charge, their drift velocity and the cross sectional area
Is it possible to make v inversely proportional to A?
Yes, over a limited range of currents. One way would be to use a very large EMF in series with a resistance R that was very much greater than the resistance of your sample.
Question from the specimen paper (as quoted by one correspondent)
So it's one of those delightful rare things: an open question to which an answer just might possibly be found, particularly if you read German and have access to previously unpublished correspondence from the early twentieth century. Good luck! If you find something, please let us know and we shall make the answer available here.
The atomic model and quantum mechanics led physicists do develop Condensed Matter Physics into a sophisticated and powerful science (whose most important practical implications are in semicondutors and thus in electronics, computing etc).
They also led to molecular physics (with many applications in chemistry, biochemistry, pharmacy, medicine, materials science, engineering etc).
Knowledge of the nucleus led to the Standard Model, a comprehensive theory that explains what we know about nuclear interactions and which unifies the strong and weak nuclear forces with electromagnetism. This also led to a greater comprehension of the early stages of the universe.
This comes about because generator 1 produces both AC and DC at the same time (ie a current that has both an AC and a DC component.
The splits in the ring in generator 1 are in such a position that the circuit is reversed at two orientations when the flux is changing at a reasonably high rate, so it goes quickly from an emf in one direction to an equal emf in the other. The current produced would in the shape of sin(ωt) for half a cycle then −sin(ωt) for the other half cycle, but it doesn't change at π/4 or 3π/4 radians, so the output has some DC component, and a larger AC component. Exactly where it changes depends on how you estimate the angle in the drawing. The current will not be discontinuous because of the inductance of the coil, and there will be some spikes and sparking.
So the examiners should have marked both B and C correct.
Q28 The quantity should really be the 'specific acoustic impedance' or 'wave impedance', rather than the 'acoustic impedance', but this is unlikely to have confused anyone.
This confusion arises regularly in discussions on Special Relativity. The word "when", as used here, implies that simultaneity is absolute. One of the important conclusions of Special Relativity is that simultaneity is relative. This is explained in any introduction to SR, or on my web page Special Relativity. So yes, the question as written is meaningless. I am informed that it was not included in calculating marks.
This is another problem with mutliple guess questions: there is no room for the student to write an explanation of why the question does not have an answer. Rather the (good) student wastes a lot of time trying to decide which answer is least wrong. This question was probably intended as a "plug and chug" question to determine how well students could remember a formula and insert values. No-one seems to have stopped to wonder what it meant.
It's rather sad to think people across the state teaching and learning in this style in timeslots supposedly reserved for physics, but the reality is that, while this syllabus and exam style remain in force, it will be a good way to succeed. It is therefore feasible that we could use this FAQ or the HSC bulletin board to prepare and to polish a canonical set of the required historical, social, geographical, technical and physics facts in ways that are easy to remember and that will earn good marks. Suggestions are welcome.
How can one tell how far away a star is? For close stars, you can use parallax. But because the Earth's orbit is small compared to interstellar distances, this doesn't work for most stars, or for galaxies.
I = P/(4 π r2).
From these measurements on relatively close Cepheid variables, we construct the calibration curve of P as a function of T. Once we have this curve, then whenever we find a distant Cepheid, we use T to get P. Then, using the same inverse square law, we can work out the distance r.The most distant objects are reported to be about 13 billion light years away, and the universe is said to be 14 billion light years away. What stops us seeing further?
In a sense, the furthest visible thing one can see is exactly cT away, where T is the age of the universe. That is the big bang itself, which we see as the microwave background. This is microwave radiation coming almost uniformly from all directions in the sky - it is the bang, red shifted and cooled by the expansion of the universe, which makes its wavelengths longer just as it makes intergalactic distances longer. Those microwaves have taken all of the age of the universe to get here.
Return to top of page and menu
The atom, photoelectric effect, energy levels, quanta, black body radiation,.
How does the quantisation of emitted radiation explain the black body radiation curve? Why does it have a peak?What is Wien's law?
THE EXPERIMENT. Black body radiation is (by definition) radiation in thermal equilibrium with its container. So let us think of a box with hot walls giving off and receiving photons at an equal rate. If we open a small window in the box, the radiation coming out will have the same spectrum, and this is what we measure with our spectrometer to determine the black body radiation spectrum. The spectral radiancy is given by Planck's radiation law, which was initially empirical.
Return to top of page and menu
λmax = (2.9 x 10-3 mK)/T.
THE QUESTION. What we want to know is the distribution of energy among the photons in the box. To do this properly would require a bit of quantum mechanics and statistical mechanics. But your question is simpler: why does the distribution go from zero at very low frequency to a peak at some frequency and then back to zero at high frequency? Where exactly did E = hν come from? [snip] What exactly was the maths that Planck was trying to do that made him quantise?
The keys are black body radiation (see the preceding question) and Boltzmann's ideas in statistical mechanics. Along with others at the time, Planck was trying to explain black body radiation: the intensity, as a function of wavelength, of EMR emitted from a surface in equilibrium with its own radiation (as in the interior of a heated box: see above).
Classical physics gives an expression proportional to 1/λ4. This is approximately what is observed for very long wavelengths. But it fails spectacularly at medium and short wavelengths, a problem known as the ultraviolet catastrophe.
To write an equation that fits the data, Planck needed to put in the denominator a factor that we now write as (exp(hc/λkBT) − 1):
Return to top of page and menu
B(λ,T) = (2hc2λ5)/(exp(hc/λkBT) − 1).
We now recognise this as very much like the Boltzmann distribution of energies, exp(−E/kBT), which in turn comes from the equipartition of energy among different ways of vibrating.
Proviso: actually, the one in this factor comes from the Bose-Einstein distribution, which applies to photons and other bosons, and to which Boltzmann's distribution is only an approximation. The statistics of energy distribution depends on whether particles can share quantum numbers (bosons) or not (fermions). Boltzmann's distribution was extended for bosons (photons, gluons, the W, Z, Higgs and the graviton, if it exists) by Bose and Einstein, hence the name) and for fermions (leptons and quarks) by Fermi and Dirac, hence that name. To obtain Boltzmann's distribution, consider the case of small λ, or large energy, where the 1 in the denominator is negligible.
I sometimes wonder how close Boltzmann was to discovering quantum mechanics. If you think about the entropy of a single atom in a box, classical physics would give it infinite entropy. Boltzmann would have known this, and quantization of energy is only a couple of steps away... But sadly, Boltzmann was under attack from the philosophers for his insights in thermal physics and was probably not in a great position to pursue further radical implications of his ideas.
The photoelectric effect
The photoelectric effect refers to the emission of electrons from the surface of a conductor (usually a metal) by incoming electromagnetic radiation.
"How is the photoelectric effect used in the following: breathalysers, solar cells and photocells?"
This would be stretching the usual meaning of the phrase 'photoelectric effect', which usually refers to the interaction of a photon and an electron in a metal to produce an electron no longer bound by the metal. Probably it is a misprint for 'photovoltaic effect', which is used in breathalysers, solar cells and photocells.
Could you please tell me about the relationship in solar cells among the photoelectric effect, semiconductors, electric fields and current?
As explained above, this may be a confusion between 'photoelectric' and 'photovoltaic'. Solar cells are usually semiconductors. A photon of light interacts with an electron and transfers it to a state with a higher electric potential. This creates an emf: the electron with higher potential can flow back to its original state via the external circuit and thus do work. Go to Solar cells and the photovoltaic effect.
Return to top of page and menuBy thinking that electrons behave like waves, how does it help to explain that the accelerating particle does not give out energy?
In the 'solar system' model of the atom, a particle-like electron travels in a circular orbit around the atom. There are different circles for orbits with different energy. Travelling in a circle, it would be accelerating (centripital acceleration) and so would radiate. In quantum mechanics, the atom has an electron wave. The wave is a bit like a standing wave in a string (see waves and strings), except that it is three dimensional. It is going nowhere. No acceleration, so no radiation. Different energy orbits have different waves, most of them have nodes, just like the waves in a string, except that in three dimensions nodes are surfaces, not points.
Electron microscope
What is magnetic diffraction and focussing of electron beams? What are the differences in resolving power between optical and electron microscopes?
A magnetic field exerts a force on a moving charge. Hence, magnetic fields can be used to bend a beam of electrons. This bending is called electron refraction. Suitable geometry can be used to focus the beam.
Heisenberg's Uncertainty Principle
(See also the separate page on this topic.) Heisenberg's Uncertainty Principle follows from a classical result, which is at least as old as Fourier. I prefer to introduce it as the Musician's Uncertainty Principle. When musicians tune up, we listen to the note for a long time so that we can adjust the frequency precisely. We tune by removing beats. (See What are interference beats?) If the frequency difference is one Hz, then you hear an interference beat every second. So, roughly speaking, if the frequencies differ by Δf, then you need a time of 1/Δf to notice. In other words:
Δf.Δt > ~ 1
Musicians know this: if the chord is short, or if you are playing a percussive instrument, the tuning is not as critical. In a long sustained chord, you have to get the tuning accurate.
Δ(hf).Δt > ~ h
which is one example of the Heisenberg uncertainty principle. Applying it to spatial frequency (number of cycles per unit distance) rather than temporal frequency (number of cycles per unit time) gives
Δp.Δx > ~ h
(uncertainty in momentum) times (uncertainty in position) is greater than about h.Pauli's exclusion prinicple
Pauli's exclusion principle states that electrons (and other fermions, ie other particles having half integer spins) cannot have the same set of quantum numbers.
The Zeeman effect, and why the Rutherford atom doesn't account for it
To deal with the Zeeman effect properly requires a level of quantum physics that very few high school students will have. I'm saying this because what I shall do next is discuss it improperly, using a picture rather like Bohr's and Sommerfeld's. Strictly, it is not appropropriate to think of an electron orbiting a nucleus, because this involves us imagining it as localised in space and having a position that is a function of time. This ideas we now know to be misleading and false. However, they are a nice picture, and contain some of the essential ideas, as does the Bohr-Sommerfeld picture.
Return to top of page and menuHow did Chadwick and Fermi's work change our understanding of the atom?
I am hoping that a historian will visit this site and help out with this side of things. Until he does, here is some inexpert help from a physicist who is not a historian.
Return to top of page and menuThe Bohr-de Broglie-Sommerfeld atom and the Hydrogen spectrum
Can anyone could explain to me how Bohr's postulates led to the development of a mathematical model to account for the existence of the hydrogen spectrum.
f = c/λ = const(1/n2 − 1/N2)
where the constant was an empirical constant, and n and N are integers.
E = (1/2)mv2 − ke2/r
where m is the electron mass, v its (tangential) velocity, k is Coulomb's constant, and r the 'orbit' radius. But the centripital force F is provided by the electrostatic attraction
mv2/r = F = ke2/r2 (1)
so substitution gives the classical result (exactly analogous to planetary mechanics):
E = − (1/2)ke2/r. (2)
Now we introduce de Broglie's contribution. In the 19th century, classical electromagnetism (Maxwell's equations) gave the momentum p of light as
p = E/c.
Using Einstein's quantisation of energy E = hf, we get the momentum of a photon
p = h/λ
where λ is the wavelength of the light. de Broglie speculated that this formula could hold for an electron also. Now the electrons in the H atom have sufficiently low energy that we may neglect relativistic effects, so de Broglie's speculation gives us for the electron:
mv = h/λ.
Now if the electron wave is to give constructive interference in a circular orbit, one requires that an integral number n of wavelengths make up a circumference, so
2.π.r = nλ = nh/mv, whence
Let's now solve (1) and (3) for r, and substitute in (2) to get the energy of the electron. Combining (1) and (3) gives
1/r = (4.π2.k2e2m)/(n2h2)
and substituting in (2) gives
E = − (2.π2.k2e4m/h2)(1/n2)
All the messy first factor is now seen to be the empirical constant. So the energy E of an electron in an orbit that has n waves around the circumference is
E(n) = −constant/n2
Now in this model, orbits with non-integral values of n are forbidden, because they correspond to destructive interference of electrons. So energy can only be absorbed or given to photons whose energy is E(n) − E(N) where n and N are both integers. So we can write the wavelengths λ of the emitted or absorbed spectrum as
1/λ = ΔE/hc = − R(1/n2 − 1/N2)
where R, called Rydberg's constant, has the value of 100 fm−1.
1/λ = − R(1/42 − 1/32).
So, from simple postulates established under other circumstances, and with a bit of logic and mathematics, we have derived a key result that may be checked by the very precise science of spectroscopy. This is what physics is about: using physical and mathematical arguments to explain the world in a quantitative way.
The spin quantum number
What does "spin" refer to in particle physics? And why is this a necessary concept?
Let's start with some classical ideas. Angular momentum is the rotational analogue of (linear) momentum. If an everyday object is spinning, it has angular momentum. If we attach electric charge to that spinning object, the circulating charge acts like a loop of current, and produces a magnetic dipole, ie a little electromagnet.
Accelerators as probes of nuclear structure
Can you please explain why accelerators are used to probe into the structure of matter?
One reason is related to the de Broglie hypothesis: that the wavelength λ of any particle is λ = h/p, where h is Planck's constant and p its momentum. So, in order to probe the nucleus (size ~ 10-15m), we need a 'probe' particle smaller than this, which means one with large momentum. To obtain information about quarks (ie to look inside a nucleon), even higher momenta are required. Accelerators provide particles with very large momentum (travelling within a fraction of a percent of the speed of light), and the momentum is well known.
Return to top of page and menu
Semiconductors, transistors, solar cells etc
What are n-type and p-type semiconductors?
n-type semiconductors are 'doped' with a small percentage of atoms that have an extra valence electron. These extra electrons can be considered to provide most of the charges available for conduction. The electrons, being negatively charged, move in the direction opposite that of the electric field (see drift velocity).
How do diodes and transistors work?
When you take some p material and n material and put them together, you get a diode (see the schematic diagram below). If you make the p side positive and the n side negative, then holes move from the p side to the junction, while electrons move from the n side to the junction. At the junction, the electrons 'fill' the holes, thereby destroying both the free electron and the hole. So the process can continue indefinitely: the diode conducts in this direction. If you reverse the polarity, holes and electrons both move away from the junction. This leaves no charge carriers near the junction, so there is no conduction. Thus a diode conducts current in only one direction---the direction of the arrow in its circuit symbol. Diodes are useful in rectification (turning AC to DC), in logic circuits and in many other applications in electronics.
Transistors as amplifiers and logic gates
Amplification. The circuit below left allows you to apply a small, variable voltage or current to the base of npn transistor via the (large) resistor Rbase. Raising this voltage or current (by decreasing Rbase) turns the transistor 'on', i.e. it allows more current to flow from collector to emitter (= more electrons from emitter to collector). A bigger current in the collector means a bigger voltage drop across the (smaller) resistance Rcollector. So a small increase in the input voltage (voltmeter at left) causes a large decrease in the output voltage (voltmeter at right). Thus this is a (very simple) inverting amplifier.
History of the invention of the transistor.
For versions of the history of the invention of the transistor, see American Institute of Physics version or the Bell Labs' version and Time Magazine article on Shockley.
What was the impact of the invention of transistors, microchips and microprocessors on society?
Simple logic circuits can be made with resistors and diodes, but for complicated circuits one needs an amplifying component. Which generally means valves or transistors.
More about transistors, computers
Can you provide more resources for teaching the Age of Silicon (NSW syllabus topic)?
The US PBS has a website associated with their documentary Transistorized. It has a good basic review of the science involved, but much historical detail
(including a lot of interesting things about the various personalities involved).
Solar cells and the photovoltaic effect
Could you please tell me about the relationship in solar cells between the photoelectric effect, semiconductors, electric fields and current?
A photon of light interacts with an electron via the photovoltaic effect and transfers it to a state with a higher electric potential. This creates an emf: the electron with higher potential can flow back to its original state via the external circuit and thus do work.
Solar cells are usually semiconductors. In solar cells, a photon interacts with an electron, but does not remove it from the material (as it does in the photoelectric effect). Rather, the energy hf of the photon increases the potential energy of the electron by eV, and so produces a potential difference which has an upper limit of hf/e (which cannot be reached in practice because of the thermal effects). The electron with higher potential energy (lower voltage) can then flow through the external circuit and do electrical work (e.g. charge a battery). This is explained in more detail on an introductory page on the photovoltaic effect from the UNSW Centre for Photovoltaics, which produces the world's most efficient silicon solar cells.
Return to top of page and menu
States of matter
The syllabus asks us to recall the states of matter and their properties and debate whether superconductivity is a new "state"
This question is mainly about the taxonomy of matter. Taxonomy is sometimes rather arbitrary, but let's put in some physical insight. We begin with the four commonly defined states of matter, which, ranked by increasing temperature, are:
solid, liquid, gas, plasma.
We can differentiate these by considering the thermal energy---a measure of the typical energy of a molecule or mole due to its thermal motion. In molecular terms, the thermal energy is kT (where k is Boltzmanns constant and T is the aboslute temperature) and in molar terms it is RT (where R is the gas constant). Let's compare this with the energies of intermolecular interaction (U) and the energy of ionisation or work function (W).
solid: kT <~U
One could further subdivide solids into crystals (solids with regular structure), glasses (solids with much less regular structure) etc. Do these count as different states of matter, or are they different forms of solids? It's a semantic question.
liquid: kT ~ U
gas: kT >~U
plasma: kT >~WBose-Einstein condensates
Why is large wavelength so imporant? How does this tie in with quantum mechanics?
I think that this is best introduced by an analogous concept: it's what surfers might call a 'set' of waves. Suppose a surfer's 'set' has 5 waves, with wavelength of 30 metres. This means that that particular 'set' will not interfere with another 'set' that is more than 150 metres away. (Physicists would refer to a wave train, or wave packet and its coherence length--and they would object to the very simplified arm-waving explanation I'm giving here!)
Are Bose-Einstein condensates a new state of matter.
NIST site on BECs
They certainly have properties that are very different from those of solids, liquids, gases & plasmas. Is that enough to be called a new state of matter? This is a semantic question.
Return to top of page and menu
Neutron stars. At exceedingly high pressures and temperatures, electrons and protons can combine to form neutrons. This happens in some massive stars when they cool enough for their gravity to squeeze the atoms into ultra high density. A whole, star bigger than the sun, becomes something rather like a gigantic atomic nucleus, several km in diameter, containing only neutrons.
Superconductors
How do superconductors work?
It is difficult to give a qualitative/conceptual description without missing something, but here is a simplistic explanation:
What is a phonon?
The atoms of a crystalline solid form a regular lattice structure, which can be likened to an array of balls connected by springs, the springs representing forces between neighbouring atoms. Such structures can vibrate mechanically in various ways, either because of thermal motion or some external force. A sound wave travelling through a solid is an example of the latter. Quantum mechanics tells us that these vibrations can only gain or lose energy in discrete amounts and these are called 'phonons', by analogy with photons for light. Under ordinary conditions there is an enormous number of phonons or photons present and we do not see this 'graininess'. However, careful experiments confirm the picture.
What causes the lattice distortions in a superconductor?
A conducting solid, such as a metal, contains electrons which can move relatively freely through the background of positively charged ion cores. As an electron moves it exerts a Coulomb attraction on neighbouring ions and will distort the lattice structure locally. This is referred to as 'electron-phonon interaction'. In some materials, at sufficiently low temperatures, this effect can lead to a dynamic pairing between electrons. This is believed to the mechanism behind superconductivity in most, but not all, superconductors.
Levitation of a magnet by a superconductor
The magnetic field does not penetrate into a superconductor. This is called the Meissner effect. This effect together with conservation of the magnetic flux provides the levitation
of a magnetic object above a superconductor or a superconductor above the magnet.
p = B2/2μo.
where B is the field strength and μo is the magnetic permeability of space. As we bring the magnet closer to the superconductor (middle picture), the field lines must be closer together, so the field is more intense, so its energy density and magnetic pressure p rise. Mechanical equilibrium (levitation) is achieved when the force due to this magnetic pressure is equal to the weight of the object.
B ~ Boa/x
For mechanical equilibrium, we want the force of the magnetic pressure acting on the pole of the magnet to equal the weight, so
mg = pa2 = a2B2/2μo.
Rearranging gives an estimate for x, the levitation height:
You probably don't have liquid nitrogen to do the experiment, but you can make a similar measurement using two similar magnets and by taking advantage of symmetry. If you keep the magnets symmetrical while bringing similar poles towards each other, then (because the magnets are equally strong) no field lines cross the plane of symmetry. So the top half of the field diagram looks just like that of the levitation case.The Meissner effect
The two following properties are specific for superconductors:
Actually a normal metal (non superconductor) also has zero resistance at zero absolute temperature. However a normal metal never manifests the Meissner effect. Thus the Meissner effect is the most important property of a superconductor.Superconductivity and computers
I've read several books now all describing superconductors, if used in computers, will allow them to operate at higher speeds. None of them describe how this actually happens.
There are a few different effects.
Applications to magnetic fields, motors, power distributions, MRI
I'm finding it really difficult to find any info of "the effects of those applications [of superconductivity] on computers, generators, motors and transmission of electricity through power grids".
Currently, these are almost entirely potential applications, so do not expect to find much information. For potential applications to computers, see the previous section.
Return to top of page and menu(A little parenthesis about comparing electric cars with petrol cars. Although fossil fuelled power station generators are much more efficient than motor car engines, the distribution of electrical energy is much less efficient than the distribution of petrol, which tends to cancel out the efficiency of generation. However, power stations are less polluting than cars, and electric cars are still much more efficient than petrol cars. This is because electric cars are designed rationally for intelligent driving. Petrol cars are constrained by the need to protect the fragile egos of some male drivers and so are almost always vastly overpowered. In principle one could design efficient petrol cars but, because of the temporarily low price of oil, there is little incentive to do so.)
Nuclear physics, radioisotopes, neutrinos etc
What are some of the industrial and medical applications of radioactivity and nuclear physics?
Radioactive materials give off charged particles (electrons, antielectrons, helium nuclei) and (uncharged) neutrons, and also high energy photons (packets of light). The charged particles collide with the charged particles in ordinary matter and give off more high energy photons.
and how is this property utilised in medicine and industries?
In medicine, hard (high energy) X-rays are used to treat cancer. Because the cancerous cells divide more quickly than normal ones, they are more vulnerable, so the radiation kills the cancer cells preferentially. Sometimes the photons are delivered directly from an X-ray beam (e.g. breast cancer). In other cases, radioactive sources may be used.
How are isotopes used in engineering or agriculture?
An important example is in nuclear engineering: In countries such as France, the construction of power stations involves lots of nuclear engineers. One of the more important isotopes is 235U. There are also nuclear reactors used for producing diverse isotopes for medical and other uses. Some bad and good news: in various countries, engineers are involved both in the building and the dismantling of nuclear weapons.
How are isotopes used in agriculture?
I don't know much about this, but I'll put down these suggestions and perhaps others can add to them.
Return to top of page and menuWhat are the links between high energy particle physics and cosmology?
Much astronomical evidence (particularly the galactic red shifts for which Hubble is famous and the ubiquitous 3 Kelvin background microwave radiation in the sky (Penzias and Wilson, more recently the cosmic background explorer COBE)) point to a 'Big Bang' about 13 billion years ago. At that time, the universe was extremely dense and hot. Let's imagine going backwards in time towards t =0.
What is a neutrino, what is an anti-neutrino?
Neutrinos (or perhaps neutrini?--Pauli was an Italo-American) have anglular momentum--and not much else. The sign of the angular momentum (+ve or −ve) is used to assign a spin, although spin in this sense is abstract, and not very similar to the spin on a cricket ball. If something has spin and is travelling, you can picture it tracing out a helix, as a spot on a ball bowled by Shane Warne would do. One of the particles produced by beta decay is an antineutrino--it has left-handed helicity.
How is beta decay described in terms of quarks?
A neutron can decay into a proton, an electron and an antineutrino. A neutron consists of an 'up' quark and two 'down' quarks. A proton consists of two 'up' quarks and one 'down' quark. The electron and the antineutrino are both leptons. How is it that one of the quarks appears to have changed from a 'down' quark to an 'up' quark?
Return to top of page and menu
Similarly, a proton is p = (uud)
The electric charge of the u quark is 2/3 (in units of the elementary charge) and the electric charge of the d quark is -1/3. As a result the proton has charge 1 and the neutron is neutral. The neutron decay goes via the following mechanism:
The d quark decays into u quark and a virtual W- boson. As a result of the decay, the neutron is converted to proton. The virtual W- boson lives for 10-26 seconds and than decays to electron and antineutrino. Blink and you miss it!
Medical Physics
Scanning: what are the features and advantages of MRI, PET, CT, X-Rays?
All four are useful for non-invasive medical imaging.
X-rays
Conventional X-ray images are usually on photographic film or a fluorescent screen. X-rays from an X-ray tube pass through the body and the difference in absorption by different structures in the body is what produces contrast in the images. The images are 2D projections of a 3D object so they contain overlapping structures. They are good for imaging structures which have a density very different from that of the material around them (eg bones) but are less effective for differentiating between different types of soft tissue.
Medical physics: ultrasound, optical fibres and MRI
1. Describe how ultrasound is used to measure bone density.
This set of medical physics answers are provided by Wendy Tsui, one of the recent graduates in medical physics at UNSW. And now a message from our sponsors: There is a shortage of medical physicists. This not only means that our graduates have no trouble finding jobs, it also means that all of our current (2002) fourth year students already have jobs in hospitals and are finishing their degrees part-time. If you want to keep it in mind as a career option, see medical physics courses.
Ultrasound measurements of bone are usually done on the calcaneous (heel bone*). Two quantities are usually measured:
2. Explain why different types of optical fibres will affect the image produced by an endoscope.
vs is determined by sending an ultrasound wave pulse through the heel and measuring the time it takes for the pulse to pass through. The speed of the wave through the heel bone depends on the stiffness or flexibility of the bone as well as the density of the bone. In general, sound waves travel faster through relatively stiff and dense objects than through more flexible objects.
In endoscopes, optical fibres are used to transmit light into the human body to illuminate tissue and they are also used to transmit the image of that tissue back out of the body to be viewed.
3. Explain that the amplitude of the signal given out when precessing nuclei relax is related to the number of nuclei present.
MRI is a rather broad subject, and we shaln't try to develop the necessary physics here. Instead we give links to dedicated sites.
Magnetic focusing
This section refers to a question (#30 part C) from a trial exam. It raises the important features of magnetic focusing. The question begins "An electron microscope has a magnetic lens for focusing a wide beam of electrons to a point. Theoretically, a beam of electrons can be focused to a point by a magnetic field with the shape shown in Diagram 1, on condition that all hte electrons are moving at the same speed." The full question and two answers at different levels are given in this .pdf file.
Motors and Generators
See Motors and generators for a background to this topic. There are descriptions and diagrams of the main classes of motors.Analyse the environmental impact of the development of AC generators.
i) AC generators have the advantage over DC generators that they allow transformers. Therefore long distance transmission may be at high voltage and low current, so there is less I2*R losses in the wires. This is a very important advantage: even with supply lines at hundreds of kV, we still lose much of the electric power generated as heat generated in the distribution system. AC therefore also allows power stations to be relatively remote from users, so users are isolated from environmental affects of the stations. This remote delivery may save energy elsewhere (e.g. goods transport and commuting).
Return to top of page and menuForce between two wires
The HSC physics syllabus, in the Motors and Generators topic, asks us to "solve problems and analyse information about simple motors using F/L = k.I1I2/d". I have never seen a problem relating this equation to motors. Can you suggest any? What relevance does this equation have to the operation of a simple motor?
Sorry, I cannot see any important direct relevance. The equation refers to infinite wires, ie wires whose separation is much smaller than their lengths. So the forces that make a motor work don't relate to this equation. The forces between say elements of wire in a stator coil and a rotor coil that are parallel are calculated using the Biot-Savart law rather than this equation. The only places where lengths of wires in motors are much closer than their length are between adjacent wires in the coil winding. And yes it's true that there would be a force tending to push the wires closer together, but as the wires (or their lacquer) are touching already, this doesn't have much effect.
Does the force between current carrying wires on either the same or opposite side of the coil effect the torque?
Internal forces in a rigid object have no direct effect on its motion. If the force were large enough to change the geometry of the coil, then there could be an effect, but this would be very small for any normal coil.
What is the implication of the force between current carrying wires for power distribution networks?
The wires must be very close together for the force to be important. In power distribution, the wires are usually separated by distances that make the force tiny. In transformers, wires are wound closely together and they exert larger forces on each other. I expect that transformers are tightly wound so that they don't rattle.
Return to top of page and menu
Transformers
This section includes transformers, power lines, induction cooktops, eddy current switching and regenerative braking. See Transformers for background. Explain why voltage transformations are related to conservation of energy
In normal operation, with a load, the rate of electrical energy supplied to the primary coil of a transformer (power in) is approximately equal to the rate at which the secondary coil supplies power to the load (power out).
Vin*Iin = Vout*Iout
These equations neglect some losses of energy (hysteresis losses in the core, heat, sound and low frequency EMR). Strictly, they should either include a phase factor or should be considered as complex quantities, because the voltage V and the current I are not exactly in phase. See Transformers.Transformers in electricity supply and the home
Why does one need to have transformers in the transfer of electrical energy from a power station to its point of use?
Much of the energy generated by a power station is lost in ohmic losses in the distribution network. The power lost in a wire of resistance R is R*i2. To make R small requires thicker wire, which is expensive. So i is made smaller. To deliver the same power P = Vi, V must be made bigger. So, near the power station, a 'step-up' transformer steps the voltage up to a high value (e.g. 110 kV). At local substations and surburban transformers, this is successively stepped down to 240 V.
How is heating caused by eddy currents in transformers overcome?
Laminating the core reduces the heat loss. The core is made out of strips of metal, separated by lacquer or other insulator, so the area of the eddy currents is reduced. Consequently, the flux and therefore the Faraday emf are smaller, so smaller eddy currents flow. You may notice transformers buzzing at 100 Hz (near G at the bottom of the bass clef) as the laminations all become electromagnets 100 times per second.
How does the principle of induction apply to cooktops in electric ranges?
Some electric ranges have a coil of wire instead of a hot plate. This carries AC and produces a strongly varying magnetic field. When a conducting saucepan is placed upon it, eddy currents are induced in the metal in the base of the saucepan. These produce heat via ohmic losses (and hysteresis in magnetisation).
Return to top of page and menuHow have eddy currents been utilised in electromagnetic braking?
The most useful sort of electromagnetic braking is what is called regenerative braking. This is used in trains and some other electric vehicles. To decelerate, the electric motor is used as a generator and converts the train's kinetic energy back to electrical energy which is fed back in to the grid. Great for energy conservation, and engineers try to maximise this. However, this is not really eddy current braking, because 'eddy currents' is usually used only when the currents waste energy.
How are eddy currents used in switching devices?
Proximity switches often involve eddy currents. These switches have the advantage that they have no (internal) moving parts or electrical contacts. This is an advantage because springs and bearings can wear out and because electrical contacts can become corroded. Proximity switches are activated by the proximity of a conductor, such as a human finger. You will have noticed them in the control panels of lifts, for example. If the proximity of your finger throws the switch but the proximity of a non-conductor does not, the switch may work on eddy currents (although it could also work by capacitance). Proximity switches of one type work like this:
Eddy current braking
Electromagnetic current braking is smoother, but why is it an advantage over conventional braking? Does it brake smoother in less time and less distance then conventional braking?
The deceleration due to friction braking is limited (i) by how much normal force you can apply to the brake pad or brake shoe (often not a problem with servo-assisted hydraulic cylinders application), (ii) by the mechanics and size of the pad: it may shear or break up if the forces are too great and (iii) by what to do with the heat generated. The last is quite important: think of how much kinetic energy a train has at speed: if you turn that into heat in a series of small brake pads they may melt, vaporise, weld to the wheels etc.
Return to top of page and menu
Oscilloscope
Could someone please explain to me the timebase properties in CRO. Does it relate to horizontal or vertical movements?
The timebase affects horizontal motion of the beam. (The voltage inputs affect vertical motion of the beam when in 'timebase' mode, but in X-Y mode, the two different input voltages deflect in the X and Y directions.)
Simple Harmonic Motion under gravity
We attach a mass m below a vertical spring with constant k (fig b). At equilibrium, the spring is extended a distance x, where
kx = mg.
F + k(x - y) = mg.
Substituting for x gives us
F + mg - ky = mg, or
Put in words: the weight of the mass and the force due to the equilibrium displacement of the mass cancel out. Displacement from this equilibrium position requires a force ky, and so the potential energy at position y (with respect to y = 0) is
F = ky.Utotal = integral of ky dy = (1/2)ky2
This potential energy could be said to include a compent due to the spring and another due to gravity. The choice of a reference for potential energy is arbitrary. Let's take it instead with respect to the unstretched position, and signify this difference with a prime.
U'total = U'grav + U'spring = (1/2)k(x - y)2 + mgy
where the first term is a constant (the spring energy at equilibrium) and the second term is the energy required to displace it from that position.
Drift velocity
There is a page about drift velocity and Ohm's law that introduces this topic. The text below is a summary of several questions about drift velocity of charge carriers in a conductor. It arises from a peculiar and confusing statement about drift velocity in the physics syllabus in New South Wales, Australia, and two multiple choice quesions in specimen papers in which insufficient information was given to allow one to answer. Briefly, if you are not studying in a NSW high school, you don't need to read this section.
What affects the drift velocity v of charge carriers in a conductor?
We assume that steady state conditions are achieved, that the temperature is not changing and that the medium is behaving in a linear way. These conditions are well approximated in many experiments.
Advice to students trying to answer a question to which there is no correct answer
v = constant.E.q
where E is the applied electric field and q is the charge of the carrier. The constant depends on the material and the charge carrier being considered: it is high for a good conductor and a mobile carrier.
I = nqvA
Does this mean that v and A are inversely proportional? Only in the special and rather peculiar case where you keep the current constant. But to keep I constant, you would have to change the field. Far more common and easy to do would be to keep E constant (e.g. keep the voltage constant and change A).
"In a certain experimental arrangement, the effect of the variation of the cross-sectional area of a conductor on the drift velocity of the electrons was investigated. If the original drift velocity was v m/s, then after doubling the cross-sectional area the new drift velocity would be expected to be:
Answer to the specimen question
a) v m/s b) v/2 m/s c) 2v m/s d) 4v m/s"
This question cannot be answered without more information about the experimental arrangement. If the potential difference were held constant (a fairly common experiment), then the answer closest to correct would be (a). If the current were held constant (a less common but possible experiment), then the answer closest to correct would be (b).
Question from another specimen paper (as quoted by another correspondent)Q8 Which of the following does the drift speed of electrons in a conducting wire depend on?
Answer to the specimen question
(A) The length of the wire
(B) The cross-sectional area of the wire
(C) The insulating material around the wire
(D) The straightness of the wire
This question cannot be answered without more infomation. If one connected the wire to an ideal EMF, the answer would be (A). If the internal resistance of the battery were much larger than the resistance of the wire, the answer would be approximately (B).
First, the real HSC paper is expected to contain few mistakes. When/if you encounter them, don't waste too much time. Give the answer that is least wrong---here (a)---and then quickly write that the question is incomplete. Then move on to the next question. After the exam is over, raise the point with your teacher (and on the bulletin board if you like) and then start making a fuss. If a question has been set and the correct answer is not given, or if the question is confusing, then the markers would be under considerable pressure to omit that question from the result.
Return to top of page and menu
Miscelleaneous questions in history and social studies
Michael Frayn's play "Copenhagen" treats the relationship between Niels Bohr and Werner Heisenberg. Bohr's wife Marguerite is the only other character, but many of the great theoretical physicists of the 1920's are mentioned. It is a great piece of theatre, and a painless way of learning a bit of history. (Physicists usually don't know much history--many of us went into physics because we didn't like rote learning. Be aware that physics syllabi in other states and in universities have much less history and rote learning than does the NSW high school syllabus.)
The Einstein-Planck debate
What debate? Einstein and Planck met (for instance at the Solvay Conference in 1927), they presumably spoke together, and they presumably must have disagreed on at least something, even if only on where to go for a beer after a hard day's physics. So in that sense, there probably has been a debate. The trouble is that there seems to be no evidence of it. This question was posted to a newsgroup for historians of science, and no-one responded with evidence. Because the peculiar syllabus taught in high schools in New South Wales devotes a segment to a debate between these two people, students in that state have been searching for evidence of such a debate. So far, no-one has posted any evidence on this site, although Bob Emery reported juxtaposed quotations of Einstein and Planck in the prologue of the book Planck M. (1931) Where is Science Going? Ox Bow Press, Woodbridge, CT (1981 reprint).
The Nobel prizes in physics
Some of the history questions in the new syllabus involve winners of the Nobel prize in physics. This site is maintained by the Nobel foundation.
What has been the impact of advances in understanding of matter on work of physicists ?
This question is pretty vague, and many answers are possible. Here are just a few.
Return to top of page and menu
2001 HSC Paper on Physics
Question 4 caused a lot of feedback. Was it B or C?
Both answers are correct. B is true: both generators produce AC. C is also true: Generator 1 produces DC and Generator 2 produces AC.
Other problems with the 2001 paper
Q3. 'Resistance' should be 'Resistivity', or else there should be some words between 'resistance of' and 'mercury'. The question as asked is dimensionally incorrect. This error is unlikely to have confused anyone.
2003 HSC Paper on Physics
Question 5
An astronaut set out in a spaceship from Earth orbit to travel to a distant star in our galaxy. The spaceship travelled at a speed of 0.8 c. When the spaceship reached the star the on-board clock showed the astronaut that the journey took 10 years. An identical clock remained on Earth. What time in years had elapsed on this clock when seen from the astronaut¹s spaceship?
Return to top of page and menu
(A) 3.6
(B) 6.0
(C) 10.0
(D) 16.7Short answers suitable for the high school syllabus in NSW
Many of the questions posed on the bulletin board, both by teachers and students, sought a fact or sentences about a particular fact. Many 'syllabus bullet points' and many questions on the exams set on that syllabus are about factual recall: some physics facts, some history facts and suppositions, some social studies facts and speculations. The practice of learning such facts etc (or even whole paragraphs about them) by rote is apparently widespread. It may also be effective, because the new syllabus (especially since the 2001 exam) puts a heavier emphasis on learning such facts etc and a lighter emphasis on understanding physics.
Glossary and skill list
What exactly do they mean by 'appreciate', 'discuss', 'recommend', 'recount', 'synthesise' etc? The NSW Board of Studies has a glossary of such verbs and a skills module. We reproduce this list on a separate page for the benefit of HSC students and teachers, and for the amusement of others.
Hints on doing tests
[ Search
| School
Information |
Physics
Courses |
Research
| Graduate
]
[ Resources
| Physics
! |
Physics
Main Page |
UNSW
Main Page
]
Happy birthday, theory of relativity!
As of June 2005, relativity is 100 years old. Our contribution is Einstein Light: relativity in 10 minutes... or 10 hours. It explains the key ideas in a short multimedia presentation, which is supported by links to broader and deeper explanations.