Download Document
Showing page : 1 of 183
This preview has blurred sections. Sign up to view the full version! View Full Document
This is the end of the preview. Sign up to access the rest of the document.

Unformatted text preview: Advanced Physics Laboratory Physics 370 Fall 2008 2 Instructor: Name: Dr. Tom Kirkman Office: PEngel 111 Phone: 3633811 email: Office Hour: 1:00 p.m. Day 2 Informal Office Hours: 8:00 a.m. 5:00 p.m. Texts: An Introduction to Error Analysis by John R Taylor (University Science, 1997) Grading: Your grade will be determined by averaging five scores: 3 lab scores, electronics workshop score and poster score. Lab grades are based on what is recorded in your Ampad (or equivalent) lab notebook. Please be complete and legible! You will probably need at least 3 of these notebooks (which may be "used") as I'll be grading old labs while you'll working on current labs. Assigned work is generally due at the beginning of the following class period. In particular your lab notebook must be turned in before the lab lecture for the following experiment. (If your lab work is incomplete, you may be offered an improved grade for a complete report, nevertheless turn in what you have!) Note that lab may include pre-lab problems which must be turned in before you start the lab. Late work is generally not accepted. All work contributing to your grade must be turned in by our last meeting day: Friday 12 December. Questions: There is no such thing as a dumb question. Questions asked during lecture or lab do not "interrupt", rather they indicate your interests or misunderstandings. The aim of lab is to do things you've never done before; it's no surprise if you've got questions. Remember: you are almost never alone in your interests, your misunderstandings, or your problems. Please help your classmates and yourself by asking any question vaguely related to physics lab. If you don't want to ask your question during class, that's fine too: I can be found almost any time on the 100-level floor of Engel Science Center. Ask if you don't find me, as I spend just as much time in the nearby labs as I do in my office. Times/Locations: Half of this course will be self-scheduled. I hope many of you will still choose to do that work in the scheduled slot, because you can be then sure to find me (i.e., help) at those times and it will help you avoid the crime of procrastination. However, because of limited 4 lab equipment, in fact you cannot all perform the data collection simultaneously. Of course, data analysis (which usually takes longer than data collection) can be done simultaneously. Three cycles are scheduled for each lab. Most of the actual data collection and analysis will take place in the suite of labs across from my office. Half of this course will meet at the scheduled time: lab lectures, workshops and the poster sessions. We will be using the astronomy lab room (PEngel 319) for lab lectures. If you cannot attend at those times, the responsibility of mastering the material falls on you. (An alternative class time--agreed to by all--would also be fine.) Note that lab lectures typically run a bit more than an hour, which leaves plenty of time to start the lab immediately following the lab lecture. Groups will also need to schedule a night lab at the observatory for photometry data collection. Lab Notebook: Your lab notebook is the primary, graded work-product for this course. It should represent a detailed record of what you have done in the laboratory--complete enough so that you could look back after a year or two and reconstruct your work just using your notebook and this manual. Your notebook should include your preparation for lab, sketches and diagrams to explain the experiment, data collected, initial graphs (done as data is collected), comments on difficulties, sample calculations, data analysis, final graphs, final results, answers to questions asked in the lab manual, and a critique of the lab. A list of suggested sections can be found in the 191 lab manual. DO NOT collect data on scratch paper and then transfer to your notebook. Your notebook is to be a running record of what you have done, not a formal (all errors eliminated) report. There will be no formal lab reports in this course. Do not delete, erase, or tear out sections of your notebook that you want to change. Instead, indicate in the notebook what you want to change and why (such information can be valuable later on). Then lightly draw a line through the unwanted section and proceed with the new work. Be Prepared! In this "Advanced Lab" you will typically be combining some fairly advanced physics concepts with equally advanced instruments. The 10 minute pre-lab talk from 191/200 is now stretched into an hour "lab lecture"; in a four hour "workshop" you will demonstrate your ability to use the electrical instrumentation you spent a whole semester developing in Physics 200. It will be quite easy to be overwhelmed by the theory and the instrumentation. Your main defense against this tsunami of information is to read and understand the material before the lecture/lab. I know that this is difficult: technical readings never seems to make sense the first time through. But frankly, one of the prime skills you should be developing (i.e., the prime skill employers seek) is being able to read, understand, and act on technical documents. What I told you in 211 was: Read aggressively! Read with a pencil in hand so you can jot down questions, complete missing steps of algebra, and argue with the author. (In this case you can actually take your complaints, comments, and arguments to the author, rather than imagining how the author would respond.) A significant problem is 5 that readings (in contrast to lectures) generally aim at getting the details right. But details obscure the big picture and misdirect attention. This leads to the suggestion of "skimming" the material. . . which is OK as long as that's just the first step to understanding. I usually instead start by reading for detail, but bit-by-bit my confusion grows and I switch to skimming. But then I repeat the process from the start. After several repeats, I usually reach a point where I'm not making progress, and I find I must do something more active like: talk to somebody about the material, or try to solve a problem--perhaps one of my own design. The aim is to try to find out why the author thinks his points are the important ones. Topics: The following schedule is based on Day 4 labs. Equivalent labs occur on the following Day 5. Day 1/4 2/4 3/4 4/4 5/4 6/4 7/4 8/4 9/4 10/4 11/4 12/4 a b M T W F W R F M T W T W Date Sep 1 Sep 9 Sep 17 Sep 26 Oct 8 Oct 16 Oct 24 Nov 1 Nov 11 Nov 19 Dec 2 Dec 10 Topics Electrical Measurements Laba Lab Lecture: Photometry & Thermionic Emission, Fortran, GPIBb Thermionic Emission Lab Lab Lecture: Bubble Chamber Lab Lecture: Langmuir Probe?c Brief Poster Workshop Poster Conference Read "Electrical Measurement Review" before the lab! Prepare for night lab at observatory. Read "Systematic Error" and skim lab chapters before the lab lecture! c Only if observatory observations were impossible due to weather. Posters: A stitch in time saves nine. Presentation of a lab project as a poster is the final component of this course. While I know procrastination always seems like the easiest course, in fact, putting together a poster months after you've completed the lab is time consuming. The easy course is actually to start your poster (particularly the figures) soon after you've completed the lab. While you can delay final construction, preparation of poster-quality figures immediately following the lab will save you a lot of time just when you most need it (at the end of the semester). See page 125 for basic poster information and that section's references for much more detailed information. Poster topics will be assigned on a first come first served basis, so there is no reason to delay selecting your poster topic. 6 Contents Systematic Error . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Electrical Measurements Review . . . . . . . . . . . . . . . . . . . . . . . . . . . Electrical Measurements Lab . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Thermionic Emission . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Photometry . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Bubble Chamber . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Poster Presentation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Langmuir's Probe . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Appendix A--plot . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Appendix B--fit . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Appendix C--spj . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Appendix D--Linux Commands . . . . . . . . . . . . . . . . . . . . . . . . . . . Appendix E--Error Formulae . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9 19 35 41 71 105 125 129 161 167 173 179 181 7 8 CONTENTS 0: Systematic Error Physical scientists. . . know that measurements are never perfect and thus want to know how true a given measurement is. This is a good practice, for it keeps everyone honest and prevents research reports from degenerating into fish stories. Robert Laughlin (1998 Physics Nobel Laureate) p.10 A Different Universe A hypothesis or theory is clear, decisive, and positive, but it is believed by no one but the man who created it. Experimental findings, on the other hand, are messy, inexact things, which are believed by everyone except the man who did the work. Harlow Shapley, Through Rugged Ways to the Stars. 1969 Perhaps the dullest possible presentation of progress1 in physics is displayed in Figure 1: the march of improved experimental precision with time. The expected behavior is displayed in Figure 1(d): improved apparatus and better statistics (more measurements to average) results in steady uncertainty reduction with apparent convergence to a value consistent with any earlier measurement. However frequently (Figs. 1(a)1(c)) the behavior shows a `final' value inconsistent with the early measurements. Setting aside the possibility of experimental blunders, systematic error is almost certainly behind this `odd' behavior. Uncertainties that produce different results on repeated measurement (sometimes called random errors) are easy to detect (just repeat the measurement) and can perhaps be eliminated (the standard deviation of the mean 1/N 1/2 which as N , gets arbitrarily small). But systematic errors do not telegraph their existence by producing varying results. Without any telltale signs, systematic errors can go undetected, much to the future embarrassment of the experimenter. This semester you will be completing labs which display many of the problems of non-random errors. Experiment: Measuring Resistance I Consider the case of the digital multimeter (DMM). Typically repeated measurement with a DMM produces exactly the same value--its random error is quite small. Of course, the absence of random error does not imply a perfect measurement; Calibration errors are 1 Great advancements is physics (Newton, Maxwell, Einstein) were not much influenced by the quest for more sigfigs. Nevertheless, the ability to precisely control experiments is a measure of science's reach and history clearly shows that discrepant experiments are a goad for improved theory. 9 10 Systematic Error (a) Neutron lifetime vs. Publication Date (b) B + lifetime vs. Publication Date (c) width vs. Publication Date (d) W mass vs. Publication Date Figure 1: Measured values of particle properties `improve' with time, but `progress' is often irregular. The error bars (x) are intended to be `1': the actual value should be in the range x x 68.3% of the time (if the distribution were normal) and in the range x 2x 95.4% of the time. These figures are from the Particle Data Group, Systematic Error 11 2 1 0 1 2 10 5 0 5 Voltage (V) A power supply 10 R V V (V) 0.9964 2.984 4.973 6.963 8.953 10.942 -0.9962 -2.980 -4.969 -6.959 -8.948 -10.938 V (V) .005 .007 .009 .011 .013 .015 .005 .007 .009 .011 .013 .015 I (mA) 0.2002 0.6005 1.0007 1.4010 1.8009 2.211 -0.1996 -0.6000 -1.0002 -1.4004 -1.8001 -2.206 I (mA) .001 .003 .005 .007 .009 .012 .001 .003 .005 .007 .009 .012 Figure 2: A pair DM-441B DMMs were used to measure the voltage across (V ) and the current through (I) a 4.99 k resistor expected and reported in the device's specifications. Using a pair of DM-441B multimeters, I measured the current through and the voltage across a resistor. (The circuit and results are displayed in Figure 2.) Fitting the expected linear relationship (I = V /R), Linfit reported R = 4.9696 .0016 k (i.e., a relative error of 0.03%) with a reduced 2 of .11. (A graphical display showing all the following resistance measurements appears in Figure 3. It looks quite similar to the results reported in Figs. 1.) This result is wrong and/or misleading. The small reduced 2 correctly flags the fact that the observed deviation of the data from the fit is much less than what should have resulted from the supplied uncertainties in V and I (which were calculated from the manufacturer's specifications). Apparently the deviation between the actual voltage and the measured voltage does not fluctuate irregularly, rather there is a high degree of consistency of the form: Vactual = a + bVmeasured (0.1) where a is small and b 1. This is exactly the sort of behavior expected with calibration errors. Using the manufacturer's specifications (essentially V /V .001 and I/I .005) we would expect any resistance calculated by V /I to have a relative error of .12 + .52 = .51% (i.e., an absolute error of .025 k for this resistor) whereas Linfit reported an error 17 times smaller. (If the errors were unbiased and random, Linfit could properly report some error reduction due to "averaging:" using all N = 12 data points--perhaps an error reduction by a factor of N 1/2 3.5--but not by a factor of 17.) Linfit has ignored the systematic error that was entered and is basing its error estimate just on the deviation between data and fit. (Do notice that Linfit warned of this problem when it noted the small reduced 2 .) Current (mA) 12 Systematic Error 4.985 4.99 4.980 4.98 4.97 4.975 4.96 4.970 A B C 4.95 A B C Figure 3: Three different experiments are used to determine resistance: (A) a pair of DM441B: V /I, (B) a pair of Keithley 6-digit DMM: V /I, (C) a Keithley 6-digit DMM direct R. The left plot displays the results with error bars determined from Linfit; the right plot displays errors calculated using each device's specifications. Note that according to Linfit errors the measurements are inconsistent whereas they are consistent using the error directly calculated using each device's specifications. When the experiment was repeated with 6-digit meters, the result was R = 4.9828 .0001 k with a reduced 2 of .03. (So calibration errors were again a problem and the two measurements of R are inconsistent.) Direct application of the manufacturer's specifications to a V /I calculation produced a 30 larger error: .003 k A direct measurement of R with a third 6-digit DMM, resulted in R = 4.9845 .0006 k. Notice that if Linfit errors are reported as accurate I will be embarrassed by future measurements which will point out the inconsistency. On the other hand direct use of calibration errors produces no inconsistency. (The graphical display in Figure 3 of these numerical results is clearly the best way to appreciate the problem.) How can we know in advance which errors to report? Reduced 2 much greater or much less than one is always a signal that there is a problem with the fit (and particularly with any reported error). Lesson: Fitting programs are designed with random error in mind and hence do not properly include systematic errors. When systematic errors dominate random errors, computer reported `errors' are some sort of nonsense. Comment: If a high precision resistance measurement is required there is no substitute for making sure that when the DMM reads 1.00000 V the actual voltage is also 1.00000 V. Calibration services exist to periodically (typically annually) check that the meters read true. (However, our SJU DMMs are not calibrated periodically.) Warning: Statistics seems to suggest that arbitrarily small uncertainties can be obtained simply by taking more data. (Parameter uncertainties, like the standard deviation of the Resistance (k) Systematic Error 13 mean, will approach zero in proportion to the inverse square-root of the number of data points.) This promise of asymptotic perfection is based on the assumption that errors are exactly unbiased -- so that with a large number of data points the errors will cancel and the underlying actual mean behavior will be revealed. However, in real experiments the errors are almost never unbiased; systematic errors cannot generally be removed by averaging. Care is always required in interpreting computer reported uncertainties. You must always use your judgment to decide if your equipment really has the ability to determine the parameters to accuracy suggested by computer analysis. You should particularly be on your guard when large datasets have resulted in errors much smaller than those reported for the individual data points. Measure Twice: Systematic Error's Bane In the thermionic emission lab you will measure how various properties of a hot tungsten wire are affected by its temperature. The presence of some problem with the thermionic lab measurements is revealed by the odd reduced 2 in fits, but how can we determine which measurements are the source of the problem? Systematic errors are most commonly found by measuring the same quantity using two different methods and not getting the same result. (And this will be the approach in this course: you will often be asked to measure a quantity (e.g., path length, temperature, plasma number density) using two different methods, and find different answers.) Under these circumstances we can use the deviation between the two different measurements as an estimate for the systematic error. (Of course, the error could also be even larger than this estimate!) Problem of Definition Often experiments require judgment. The required judgments often seem insignificant: Is this the peak of the resonance curve? Is A now lined up with B? Is the image now best in focus? Is this the start and end of one fringe? While it may seem that anyone would make the same judgments, history has shown that often such judgments contain small observer biases. "Problem of definition errors" are errors associated with such judgments. Historical Aside: The "personal equation" and the standard deviation of the mean. Historically the first attempts at precision measurement were in astrometry (accurate measurement of positions in the sky) and geodesy (accurate measurement of positions on Earth). In both cases the simplest possible measurement was required: lining up an object of interest with a crosshair and recording the data point. By repeatedly making these measurements, the mean position was very accurately determined. (The standard deviation of the mean is the standard deviation of the measurements divided by the square root of the number of measurements. So averaging 100 measurements allowed the error to be reduced by a factor of 10.) It was slowly (and painfully: people were fired for being `poor' observers) determined that even as simple an observation as lining up A and B was seen differently by different people. Astronomers call this the "personal equation": an extra adjustment to be made to an observer's measurements to be consistent with other observers' measurements. This small bias would never have been noticed without the error-reduction produced by 14 Systematic Error averaging. Do notice that in this case the mean value was not the `correct' value: the personal equation was needed to remove unconscious biases. Any time you use the standard deviation of the mean to substantially reduce error, you must be sure that the random component you seek to remove is exactly unbiased, that is the mean answer is the correct answer. In the bubble chamber lab, you will make path-length measurements from which you will determine a particle's mass. Length measurements (like any measurement) are subject to error, say 0.1 mm. A computer will actually calculate the distance, but you have to judge (and mark) the beginning and end of the paths. The resulting error is a combination of instrument errors and judgment errors (problem of definition errors). Both of these errors have a random component and a systematic component (calibration errors for the machine, unconscious bias in your judgments). A relatively unsophisticated statistical treatment of these length measurements produces a rather large uncertainty in the average path length (and hence in the particle's mass calculated from this length). However, a more sophisticated treatment of the same length data produces an incredibly small estimated length error much less than 0.1 mm. Of course it's the aim of fancy methods to give `more bang for the buck' (i.e., smaller errors for the same inputs), however no amount of statistical manipulation can remove built in biases, which act just like systematic (non-fluctuating) calibration errors. Personal choices about the exact location of path-beginning and path-end will bias length measurements, so while random length errors can be reduced by averaging (or fancy statistical methods), the silent systematic errors will remain. Experiment: Measuring Resistance II If the maximum applied voltage in the resistance experiment is increased from 10 V to 40 V a new problem arises. The reduced 2 for a linear fit balloons by a factor of about 50. The problem here is that our simple model for the resistor I = V /R (where R is a constant) ignores the dependence of resistance on temperature. At the extremes of voltage (40 V) 1 about 3 W of heat is being dumped into the resistor: it will not remain at room temperature. If we modify the model of a resistor to include power's influence on temperature and hence on resistance, say: V I= (0.2) k1 (1 + k2 V 2 ) (where fitting constant k1 represents the room temperature resistance and k2 is a factor allowing the electrical power dissipated in the resistor to influence that resistance), we return to the (too small) value of reduced 2 seen with linear fits to lower voltage data. However even with this fix it is found that the fit parameters depend on the order the data is taken. Because of `thermal inertia' the temperature (and hence the resistance) of the resistor will lag the t temperature: T will be a bit low if the resistor heating up during data collection or a bit high if the resistor is cooling down. The amount of this lag will depend on the amount of time the resistor is allowed to equilibrate to a new applied voltage. Dependence of data on history (order of data collection) is called hysteresis. You might guess that the solution to this `problem' is to always use the most accurate model of the system under study. However it is known that that resistance of resistors depends on pressure, magnetic field, ambient radiation, and its history of exposure to these quantities. Very commonly we simply don't care about things at this level of detail and seek the fewest Systematic Error 15 possible parameters to `adequately' describe the system. A resistor subjected to extremes of voltage does not actually have a resistance. Nevertheless that single number does go a long way in describing the resistor. With luck, the fit parameters of a too-simple model have some resemblance to reality. In the case of our Ohm's law resistance experiment, the resulting value is something of an average of the high and low temperature resistances. However, it is unlikely that the computer-reported error in a fit parameter has any significant connection to reality (like the difference between the high and low temperature resistances) since the error will depend on the number of data points used. The quote often attributed to Einstein: "things should be made as simple as possible, but not simpler" I hope makes clear that part of art of physics is to recognize the fruitful simplifications. Lesson: We are always fitting less-than-perfect theories to less-than-perfect data. The meaning of of the resulting parameters (and certainly the error in those parameters) is never immediately clear: judgment is almost always required. The Spherical Cow I conceive that the chief aim of the physicist in discussing a theoretical problem is to obtain `insight' -- to see which of the numerous factors are particularly concerned in any effect and how they work together to give it. For this purpose a legitimate approximation is not just an unavoidable evil; it is a discernment that certain factors -- certain complications of the problem -- do not contribute appreciably to the result. We satisfy ourselves that they may be left aside; and the mechanism stands out more clearly freed from these irrelevancies. This discernment is only a continuation of a task begun by the physicist before the mathematical premises of the problem could even be stated; for in any natural problem the actual conditions are of extreme complexity and the first step is to select those which have an essential influence on the result -- in short, to get hold of the right end of the stick. A. S. Eddington, The Internal Constitution of the Stars, 1926, pp 1012 As Eddington states above, the real world is filled with an infinity of details which a priori might affect an experimental outcome (e.g., the phase of the Moon). If the infinity of details are all equally important, science cannot proceed. Science's hope is that a beginning may be made by striping out as much of that detail as possible (`simple as possible'). If the resulting model behaves --at least a little bit-- like the real world, we may have a hold on the right end of the stick. The short hand name for a too-simple model is a "spherical cow" (yes there is even a book with that title: Clemens QH541.15.M34 1985). The name comes from a joke that every physicist is required to learn: Ever lower milk prices force a Wisconsin dairy farmer to try desperate--even crazy--methods to improve milk production. At he end of his rope, he drives to Madison to consult with the greatest seer available: a theoretical physicist. The physicist listens to him, asks a few questions, and then says he'll take the 16 Systematic Error assignment, and that it will take only a few hours to solve the problem. A few weeks later, the physicist phones the farmer, and says "I've got the answer. The solution turned out to be a bit more complicated than I thought and I'm presenting it at this afternoon's Theory Seminar". At the seminar the farmer finds a handful of people drinking tea and munching on cookies--none of whom looks like a farmer. As the talk begins the physicist approaches the blackboard and draws a big circle. "First, we assume a spherical cow..." (Yes that is the punch line) One hopes (as in the spherical cow story) that approximations are clearly reported in derivations. Indeed, many of the `problems' you'll face this semester stem from using high accuracy test equipment to test an approximate theory. (It may be helpful to recall the 191 lab on measuring the kinetic coefficient of friction in which you found that accurate measurement invalidated F = k N where k was a constant. Nevertheless `coefficient of friction' is a useful approximation.) For example, in the Langmuir's probe lab we assume that the plasma is in thermal equilibrium, i.e., that the electrons follow the Maxwell-Boltzmann speed distribution and make a host of additional approximations that, when tested, turn out to be not exactly true. In that lab, you will find an explicit discussion of the error (20% !) in the theoretical equation Eq. 7.53. kTe 1 (0.3) Ji en 2 Mi Again this `error' is not a result of a measurement, but simply a report that if the theory is done with slightly different simplifications, different equations result. Only rarely are errors reported in theoretical results, but they almost always have them! (Use of flawed or approximate parameters is actually quite common, particularly in engineering and process control--where consistent conditions rather than fundamental parameters are the main concern.) What can be done when the model seems to produce a useful, but statistically invalid fit to the data? 0. Use it! Perhaps the deviations are insignificant for the engineering problem at hand, in which case you may not care to expore the reasons for the `small' (compared to what matters) deviations, and instead use the model as a `good enough' approximation to reality. 1. Find a model that works. This obvious solution is always the best solution, but often (as in these labs) not a practical solution, given the constraints. 2. Monte Carlo simulation of the experiment. If you fully understand the processes going on in the experiment, you can perhaps simulate the entire process on a computer: the computer simulates the experimental apparatus, producing simulated data sets which can be analyzed using the flawed model. One can detect differences (biases and/or random fluctuation) between the fit parameters and the `actual' values (which are known because they are set inside the computer program). 3. Repeat the experiment and report the fluctuation of the fit parameters. In some sense the reporting of parameter errors is damage control: you can only be Systematic Error 17 labeled a fraud and a cheat if, when reproducing your work, folks find results outside of the ballpark you specify. You can play it safe by redoing the experiment yourself and finding the likely range (standard deviation) of variation in fit parameters. In this case one wants to be careful to state that parameter values are being reported not physical parameters (e.g., `indicated temperature' rather than actual temperature). Again, since systematic errors do not result in fluctuation, the likely deviation between the physical parameters and the fit parameters is not known. This was the approach used in the 191 k experiment. 4. Use bootstrapping2 to simulate multiple actual experiments. Bootstrapping `resamples' (i.e., takes subsets) from the one in-hand data set, and subjects these subsets to the same fitting procedure. Variation in the fit parameters can then be reported as bootstrap estimates of parameter variation. The program fit can bootstrap. (Again: report that an unknown amount of systematic error is likely to be present.) 5. Fudge the data. In dire circumstances, you might try scaling all your x and y error bars by a constant factor until the probability is acceptable (0.5, say), to get plausible values for A and B . Numerical Recipes by Press, et al., 3rd ed. p. 787 Increase the size of your error bars so you get reduced 2 = 1, and then calculate errors as in the usual approach. Clearly this is the least legitimate procedure (but it is what LINFIT does). One must warn readers of the dicey nature of the resulting error estimates. The program fit can fudge. Special Problem: Temperature Measuring temperature is a particular problem. (Here, the first two labs involve measuring temperatures above 1000 K in situations a bit removed from the experimenter.) You may remember from 211 that while temperature is a common part of human experience, it has a strikingly abstruse definition: ln 1 (0.4) kT E While the usual properties of Newtonian physics (mass, position, velocity, etc.) exist at any time, temperature is a property that exists contingent on a situation: `thermal equilibrium'. And thermal equilibrium is an idealization only approximately achieved--never exactly achieved--in real life. Furthermore in these experiments, thermal equilibrium is not even closely approximated, so the resulting temperatures have somewhat restricted meanings. In the photometry lab `the temperature of stars' is measured. In fact stars do not have a temperature and are not in thermal equilibrium. Nevertheless, astronomers find it useful to define an `effective temperature' which is really just a fit parameter that is adjusted for the best match between the light produced by the star and the light predicted by the model. 2 wiki Bootstrapping (statistics) 18 Systematic Error Special Problem: Assuming Away Variation In the 191 k lab, you assumed the sliding motion was characterized by one value of k , whereas a little experimentation finds usually slippery and sticky locations (handprints?). In the thermionic emission lab you will measure how various properties of a hot wire depend on temperature, however the hot wire does not actually have a temperature: near the supports the wire is cooled by those supports and hence is at a lower temperature. Our spherical cow models have simplified away actual variation. The hope is that the fit model will thread between the extrems and find something like the typical value. Of course, real variations will result in deviations-from-fit which will be detected if sufficiently accurate measurements are made. Special Problem: Derive in Idealized Geometry, Measure in Real Geometry Often results are derived in simplified geometry: perfect spheres, infinite cylinders, flat planes, whereas measurements are made in this imperfect world. In these labs (and often in real life) these complications are set aside; instead of waiting for perfect theory, experiment can test if we have "the right end of the stick". Thus a Spherical Cow is born. The theory should of course be re-done using the actual geometry, but often such calculations are extremely complex. Engineering can often proceed perfectly adequately with such a first approximation (with due allowance for a safety factor) and, practically speaking, we simply may not need accuracy beyond a certain number of sigfigs. Indeed it takes a special breed of physicist to push for the next sigfig; such folks are often found in national standards labs like Purpose: In all your physics labs we have stressed the importance of `error analysis'. However, in this course you will have little use for that form of error analysis (because it was based on computer reports of random errors). Instead, my aim in this course is to introduce you to the problems of non-random error. In the bubble chamber lab you will see how increasingly sophisticated analysis can reveal systematic error not important or evident in more elementary analysis. In the other labs you will see how systematic error can be revealed by measuring the same quantity using different methods. In all of these labs you will use too simple theory to extract characterizing parameters, which are not exactly the same quantity as might occur in a perfect version of the problem. Comment: The lesson: "measure twice using different methods" is often impractical in real life. The real message is to be constantly aware that the numbers displayed on the meter may not be the truth. Be vigilant; check calibrations and assumptions whenever you can. But the opening Shapley quotation tells the truth: "Experimental findings. . . are messy, inexact things, which are believed by everyone except the man who did the work". 1: Electrical Measurements Review Pre-Lab: Read and understand this chapter before coming to lab! Complete the following problems: 1a, 39, 11, 1419 Record the results in your notebook. See page 36. Work as individuals! Begin lab work: 1 September; Due: 17 September Horowitz & Hill, The Art of Electronics Chapter 1 & Appendix A Diefenderfer & Holton, Principles of Electronic Instrumentation Chapter 6 Paul Scherz, Practical Electronics for Inventors, 2nd edition Lab: Reference: Purpose This review aims to re-familiarize you with the electronic instrumentation you used in Physics 200 lab: components like resistors and capacitors, measuring devices like digital multimeters and oscilloscopes, and electrical sources like function generators and d.c. power supplies. In addition the usual schematic symbols for these devices are presented. Components In Physics 200 you learned about three passive, linear components: resistors (R), capacitors (a.k.a., condensers, C) and inductors (a.k.a., chokes or coils, L). These devices are called passive because they require no outside power source to operate (and thus circuits involving just these components cannot amplify: at best power in = power out). These devices are called linear because the current through these devices is linearly proportional to the voltage across them1 : 1 V Z I= (1.1) 1 Unless otherwise stated, you should always assume that "voltage" and "current" refers to the root-meansquare (rms) value of that quantity. That is what meters always report. Of course, this equation would also apply to peak or peak-to-peak values as long as they are consistently used. 19 20 Electrical Measurements Review R C L ground diode anode cathode XC = 1 2 f C X L = 2fL Figure 1.1: The schematic symbols for common components including resistors (R), capacitors (C), and inductors (L). For these three "linear" devices there is a linear relationship between the current through the device (I) and the voltage across the device (V ). For resistors, the resistance, R = V /I is constant. For capacitors and inductors the reactance X = V /I depends on frequency (f ) as shown above. The two symbols for ground (zero volts) are, respectfully, chassis and earth ground. The impedance Z (unit: ) determines the proportionality constant. Large impedances (think M) mean small currents (think A) flow from moderate driving voltages. Small impedances (think 1 ) mean large currents (1 A) flow from moderate driving voltages. Impedance2 is an inclusive term: for resistors the impedance is called resistance; for inductors and capacitors the impedance is called reactance. Inductors and capacitors are useful only in circuits with changing voltages and currents. (Note: changing voltage and/or current = alternating current = a.c.; unchanging current/voltage = direct current = d.c..) The reactance (or impedance) of inductors and capacitors depends on the frequency f of the current. A capacitor's impedance is inversely proportional to frequency, so it impedes low frequency signals and passes high frequency signals. An inductor's impedance is proportional to frequency, so it impedes high frequency currents but passes low frequency currents. Recall that current and voltage do not rise and fall simultaneously in capacitors and inductors as they do in resistors. In the inductors the voltage peaks before the current peaks (voltage leads current, ELI). In capacitors the current peaks before the voltage peaks (current leads voltage, ICE). Diodes are non-linear passive devices. Positive voltages on one terminal (the anode) results in large current flows; positive voltages on the other terminal (the cathode) results in essentially no current flow. Thus the defining characteristic of diodes is easy current flow in only one direction. The arrow on the schematic symbol for a diode shows the easy direction for current flow. On a diode component a white line often marks which terminal allows easy outward flow. Light Emitting Diodes (LED) are specialized diodes in which part of the electrical power (IV ) is dissipated as light rather than heat. The color of the emitted light (from IR to UV) depends on the material and V . Impedance is often distinguished as being a complex quantity (as in Z = a + bi, where i = -1 and a, b R). Resistors then have real impedances whereas Z is purely imaginary for inductors and capacitors. This advanced approach is followed in Physics 338. In Physics 200 this reality is hidden behind `phasers'. 2 Electrical Measurements Review current source ac voltage transformer source 21 + - real battery battery r + - POWER ON DC VOLTS METER VOLTAGE DC AMPS OFF CURRENT CURRENT ADJ FREQ MULT (Hz) SWEEP WIDTH SWEEP RATE DC OFFSET 50 or 1.2 VOLTAGE ADJ COARSE FINE lab power supply function generator Figure 1.2: The schematic symbols for common electric sources. Real sources can be modeled as ideal sources with hidden resistors. Lab power supplies are fairly close to ideal sources, if operated within specified limits. For example, the Lambda LL-901 specifications report an internal resistance less than 4 m. Sources D.C. Current and Voltage Sources An ideal voltage source produces a constant voltage, independent of the current drawn from the source. In a simple circuit consisting of a voltage source and a resistor, the power dissipated in the resistor (which is the same as the power produced by the voltage source) is V 2 /R. Thus as R 0 infinite power is required. I hope it comes as no surprise that infinite power is not possible, so ideal voltage sources do not exist. Every real voltage source has some sort of current limit built in. (Unfortunately it is not uncommon that the current limiting feature is the destruction the device -- beware!!!) Batteries can be thought of as an ideal voltage source in series with a small resistor3 , r, (the internal resistance). The maximum battery current flow (achieved if the external circuit is a "short" i.e., R 0) is V /r. Laboratory power supplies ("battery eliminators") usually have an adjustable maximum current limit that can be achieved without damaging the device. When this current limit is reached the supplied voltage will be automatically reduced so no additional current will flow. When operating in this mode (current pegged at the upper limit, with actual output voltage varying so that current is not exceeded) the power source is acting as a nearly ideal current source. An ideal current source would produce a constant current, arbitrarily increasing the voltage if that currents meets a big resistance. In a simple circuit consisting of a current source and a resistor, the power dissipated in the resistor (which is the same as the power produced by the current source) is I 2 R. Thus as R infinite power is required. No surprise: infinite power is not available, so ideal current sources do not exist. Every real current source has some sort of voltage limit built in. Real current sources can be modeled as ideal current sources in parallel with a (large) internal resistance4 . 3 4 Thvenin's Theorem claims most any two terminal device can be thought of this way! e Norton's Theorem claims most any two terminal device can be thought of this way! 1.4 1.6 1. 8 + - .2 02 .0 .4 10 100 1K 10K 100K 1M DC .6 1 .8 PWR OFF OFF OFF AMPLITUDE LO 50 OUT HI 1.0 2.0 VCG IN GCV OUT SWEEP OUT PULSE OUT (TTL) 22 Electrical Measurements Review A.C. Voltage Sources A function generator is a common source of a.c. signals. A function generator can produce a variety of wave shapes (sinusoidal, square, triangle, . . . ) at a range of frequencies, and can even `sweep' the frequency (i.e., vary the frequency through a specified range during a specified period). Usually the signals are balanced (i.e., produces as much positive voltage as negative), but a d.c. offset can be added to the signal, for example, producing a voltage of the form A cos(2f t) + B (1.2) (In this case the d.c. offset would be B, the amplitude would be A, and the peak-to-peak voltage would be 2A.) Most function generators are designed to have an internal resistance of 50 and maximum voltage amplitude of around 10 V. Generally they have a power output of at most a few watts. Certainly the most common a.c. source is the wall receptacle: 120 V at a frequency of 60 Hz. Transformers can be used to reduce this voltage to less dangerous levels for lab use. A `variac' (a variable transformer) allows you to continuously vary the voltage: 0120 V. Relatively large power (> 100 W) and current (> 1 A) can be obtained in this way. Of course the frequency is not changed by a transformer; it would remain 60 Hz. Electrical Measurement Digital Multimeter (DMM) The most common measurement device is the digital multimeter (DMM). Feel free to call these devices `voltmeters', but in fact they can measure much more than just volts. For example, the Keithley 169 is fairly generic, measuring volts (a.c. and d.c.), amps (a.c. and d.c.), and ohms. The hand-held Metex M-3800 measures the above and adds transistor hF E and diode test. The bench-top DM-441B measures all of the above and frequency too. The ease of switching measurement options should not blind you to the fact that these measurement options put the DMM into radically different modes. If, for example, the DMM is properly hooked up to measure voltage and -- without changing anything else -- you switch the DMM to measure amps, most likely something will be destroyed: either the DMM or the device it is connected to. Please be careful! Recall that voltage (or more properly potential difference) is a measurement of the electrical `push' applied to an electron as it moves through a section of the circuit. It is analogous to water pressure in that the difference in the quantity determines the driving force. (No pressure difference; no net force.) Note that the presence of big `push', in no way guarantees that there will be a large resulting flow (current). A large resistance (or for a.c. circuits, impedance) can restrict the flow even in the presence of a big push. In fact, large current flows are often driven by very small voltage differences as a very fat (small resistance) wire is provided for the flow. Wires work by having very small resistance; an ideal wire has zero resistance, and hence nearly zero voltage difference between its two ends. Electrical Measurements Review an ammeter acts like a short circuit A a voltmeter acts like a open circuit V 23 Figure 1.3: The schematic symbols for basic meters. An ammeter must substitute for an existing wire to work properly, whereas a voltmeter can be attached most anywhere. A voltmeter measures the potential difference across or between the two selected points. A good voltmeter is designed to draw only a small current so it must be equivalent to a large resistance. Modern DDMs typically have input impedances greater than 1 M. Voltmeters with even larger resistance (T) are called electrometers. An ideal voltmeter would draw no current; it would be equivalent to an `open circuit'. (An open circuit [R ] is the opposite of `short circuit' [R 0] which is obtained if the two points are connected by an ideal wire.) Since voltmeters draw only a small current, they should not affect the circuit being measured. This makes the voltmeter an excellent diagnostic tool. Measurement of the current flow through a wire, necessarily requires modification of the circuit. The flow normally going through the wire must be redirected so it goes through the ammeter. This requires breaking the path that the current normally uses, i.e., cutting the wire and letting the ammeter bridge the two now disconnected ends. (With luck, the wire may not need to be literately cut, perhaps just disconnected at one end.) Because current measurements require this modification of the circuit under study, one generally tries to avoid current measurements, and instead substitute a voltage measurement across a device through which the current is flowing. Knowledge of the impedance of the device will allow you to calculate the current from the voltage. Because an ammeter substitutes for a wire, it should have a very small resistance. An ideal ammeter would have zero resistance, i.e., be a short circuit between its two leads. (Note that this is the opposite of a voltmeter, which ideally has an infinite resistance between its two leads.) Real ammeters require a small voltage drop (typically a fraction of a volt for a full scale reading) to operate. This small V is called the voltage burden. I say again: converting a DMM from voltmeter to ammeter makes a drastic change from open circuit to short circuit. Making such a switch in a DMM connected to a circuit usually results in damaging something. Poking around in a circuit with a voltmeter is unlikely to cause damage, because the voltmeter acts like a huge resistor (not that different from the air itself). Poking around in a circuit with an ammeter is quite likely to cause damage, as it is linking two points with a wire, i.e., adding short circuits between points in the circuit. A DMM measures resistance by forcing a current through the leads and, at the same time, measuring the potential difference between the leads. The resistance can then be calculated by the DMM from R = V /I. Note that since a DMM in resistance mode is sending a current through its leads (`sourcing current') and assuming that this current is the only current flowing through the device, you cannot measure the resistance of a powered device. Furthermore, you can almost never use the DMM to measure the resistance of a device attached to an existing circuit because the current injected by the DMM may end looping back through the circuit rather than through the device. (In addition injecting current into 24 Electrical Measurements Review a circuit at random places may damage some components in the circuit.) Thus to measure the resistance of something, you almost always have to disconnect at least one end of it from its circuit. Lab Reminder: For accurate measurement you must use the appropriate scale: the smallest possible without producing an overscale. DMM's may report an overscale condition by a flashing display or a nonsense display like: . Similarly, for significant DMM measurements you should record in your notebook every digit displayed by the DMM. A.C. DMM Measurements Some special considerations are needed when using a DMM to measure a.c. currents or voltages. First, DMMs give accurate readings only for frequencies in a limited range. DMMs fail at low frequencies because DMMs report several readings per second and, in order to be properly measured, the signal needs to complete at least one cycle per reading frame. Thus f > 20 Hz or so for accurate readings. At the high frequencies, the input capacitance ( 100 pF) of the DMM tends to short out the measurement (recall the impedance of a capacitor at high frequency is small). No SJU DMM operates accurately above 0.3 MHz; some DMMs have trouble above 1 kHz. The DMM's manual, of course, reports these specifications. Recall that a.c. signals are time-variable signals . . . there is no steady voltage to report as "the" voltage. The solution is to report root-mean-square (`rms') quantities. (The square root of the average of the square of the voltage.) Since this is a complex thing to calculate, most cheap DMMs assume that the signal is sinusoidal so that there is a relationship between the rms value and the peak value: Vrms = Vpeak / 2 (1.3) These cheap DMMs find the peak voltage, divide it by 2 and report the result as if it were an rms voltage. This of course means the meter reports faulty values if non-sinusoidal signals are applied. "True rms" meters properly calculate the rms quantities. Oscilloscope Generally speaking DMMs work in the `audio' frequency range: 20 Hz 20 kHz. `Radio frequency' (rf, say frequencies above 1 MHz) require an alternative measuring device: the oscilloscope (a.k.a., o'scope or scope). Unlike the DMM, the oscilloscope usually measures voltage (not current). Also unlike the DMM, the scope effectively has only one lead: the `black' lead of the scope is internally connected to ground; the voltage on the `red' lead is displayed on the screen. (Note: with a DMM you can directly measure the 1 V potential difference between two terminals at 100 V and 101 V. You cannot do this with a scope--its `black' lead is internally connected to ground so if you connect it's `black' lead to the 100 V terminal you will cause a short circuit [the 100 V terminal connected to ground through the scope, which will probably damage either the device or the scope].) While a DMM takes a complex waveform and reduces it to a single number: Vrms , a scope displays the graph of voltage vs. time on its screen. Electrical Measurements Review 25 Multipurpose Knob Tektronix TDS 1002B TWO CHANNEL DIGITAL STORAGE OSCILLOSCOPE 60 MHz 1 GS/s Menu Select AUTORANGE SAVE/RECALL MEASURE ACQUIRE Trigger AUTOSET RUN/ STOP SINGLE SEQ HELP UTILITY REF MENU CURSOR DISPLAY DEFAULT SETUP SAVE VERTICAL POSITION HORIZONTAL POSITION TRIGGER LEVEL Option Buttons PRINT HORIZ MENU CH 1 MENU MATH MENU CH 2 MENU SET TO ZERO TRIG MENU SET TO 50% VOLTS/DIV SEC/DIV FORCE TRIG TRIG VIEW PROBE CHECK CH 1 PROBE COMP ~5V@1kHz CH 2 EXT TRIG 300 V CAT II USB Flash Drive ! Vertical Horizontal Figure 1.4: The Tektronix TDS 1002B is a two channel digital storage oscilloscope (DSO). Pushing a button in the Menu Select region displays a corresponding menu of the items to be re-configured adjacent to the option buttons. (The multifunction knob allows a continuous variable to be modified.) The vertical region has knobs that control the size (volts/div) and position of vertical (y) scales for channel 1 (CH 1) and channel 2 (CH 2). Push buttons in this region control the display of menus for those channels and combinations of those channels (math). The horizontal region has knobs that control the size (sec/div) and position of horizontal (x) scales. In addition to the Main time-base, this dual time-base scope can blow up a selected portion ("Window") of the display. The controls to do this are in the horiz menu. The trigger region has a knob that controls the voltage level for the triggering and the trig menu button allows the display of configurable options for the trigger. Note particularly the autoset, default setup, and help buttons to the right of the Menu Select region. 26 Electrical Measurements Review horizontal position trigger display trigger level holdoff Figure 1.5: An oscilloscope displays one wave-section after another making an apparently steady display. Determining when to start a new wave-section is called triggering. The level and slope of the signal determine a trigger point. The trigger point is placed in the center of the display, but it can be moved using the horizontal position knob. The holdoff is an adjustable dead time following a triggered wave-section. Oscilloscopes are generally used to display periodic signals. Every fraction of a second, a new section of the wave is displayed. If these successively displayed wave-sections match, the display will show an apparently unchanging trace. Thus the triggering of successive wave-sections is critical for a stable display. And an unsteady display (or a display with `ghosts') is a sign of a triggering problem. In addition, the scales used for the horizontal and vertical axes should be in accord with the signal. (That is you want the signal to neither be off-scale large or indistinguishable from zero. A too large time (horizontal) scale will result in displaying hundreds of cycles as a big blur; a too small time scale will result in just a fraction of a cycle being displayed.) Oscilloscope Controls and How To Use Them Pre-lab Exercise The knob-filled face of an oscilloscope may appear intimidating, but the controls are organized in a logical and convenient way to help you recall their functions. The class web site contains a line drawing of an oscilloscope (TDS1002Bscope.pdf). Print out this diagram and have it in hand as you read this section. As each control is discussed m below a circled number (e.g., 1 ) appears. Find this control on the line drawing and label it with that number. Attach your diagram in your notebook. The name of each control or feature will be printed in SmallCaps Text. Display Section The left hand side (lhs) of the scope is dominated by the display or m m screen 1 . Note that there is a USB port 2 below the display--this allows you to save scope data and display images on a thumb drive! There is an additional USB port in the back. The power switch is on top of the scope, lhs. m m Vertical Sections Right of the option buttons 25 29 are the knobs and buttons that control the vertical portions of the graph. Typically the display shows a graph of voltage (y or vertical) vs. time (x or horizontal). This scope has two BNC5 inputs so the vertical section is divided into two sections with identical controls for each input. The inputs are 5 According to Wiki, this denotes "bayonet Neill-Concelman" connector. This coaxial cable connector is very commonly used when signals below 1 GHz are being transmitted. You can also use Wiki to learn about banana connectors and alligator clips (a.k.a. crocodile clips). Electrical Measurements Review 27 (a) Channel 1 Menu (b) Cursor Menu (c) Measure Menu (d) Trigger Menu Figure 1.6: Buttons (often in the Menu Select region) control which menu appears on the rhs of the display. Here are four examples. 28 Electrical Measurements Review m m called channel 1 (CH 1) 5 and channel 2 (CH 2) 10 . The scale factor for each input is m 11 ; the vertical location of zero determined by the corresponding volts/div knob 6 & m m m volts on the screen is determined by the corresponding position knobs 8 & 13 . Note that the scale factors and the zero location for the channel traces are independently set. Therefore the graph axes can not show values in physical units (volts), rather the graph is displayed on a 8 10 grid with units called divisions. (Note that a division is about a cm.) The volts/div (or sensitivity) knobs are similar to the range switch on a multimeter. To display a 2 volt signal, set the volts/div knob to 0.5 volts/div. A trace 4 div high will then be obtained, since 4 div 0.5 V/div = 2 V. The current settings of these sensitivity knobs is displayed in the lower lhs of the display. You should always adjust the sensitivity so that the signal displayed is at least 3 div peak-to-peak. The traces from the two channels look identical on the screen; the symbols 1- and 2- on the lhs of the display show the position of zero volts for each trace. If you are unsure which trace is which, moving a position knob will immediately identify the corresponding trace. m m The ch 1 and ch 2 menu buttons 7 & 12 produce menus controlling how the corresponding input is modified and displayed. In addition pushing these menu buttons toggles the display/non-display of the corresponding signal. The top menu item in ch 1 and ch 2 menus, is Coupling, with options: DC, AC, Ground. These options control how the signal supplied to the BNC is connected (coupled) to the scope's voltage measuring circuits. The Ground option means the scope ignores the input and instead the display will graph a horizontal line at zero volts (ground) -- this allows you to precisely locate and position the zero volt reference line for that channel on the grid. When AC is selected a capacitor is connected between the inputted signal and scope's electronics. This capacitor prevents any d.c. offset voltage from entering the scope's circuits. Thus any d.c. offset will be subtracted from the signal and the display will show the remaining signal oscillating around a mean of zero volts. The usual selection for Coupling is DC, which means the signal is displayed unmodified. Proper Practice: Use CouplingDC almost always. Exceptional circumstances like a d.c. offset larger than an interesting a.c. signal (i.e., B A in Eq. 1.2) or a requirement to measure an rms voltage in the usual way, i.e., with any d.c. offset removed, may force occasional, brief uses of CouplingAC, but don't forget to switch back to CouplingDC ASAP. m In between the ch 1 and ch 2 menu buttons, find the math menu 9 button. which is used to display combinations of CH 1 and CH 2 (+, -, ) and Fourier transforms (FFT) of either signal. Electrical Measurements Review Ground 29 Probe The second item from the bottom in the ch 1 and ch 2 menus is Probe. "Probe" is the name for the device/wire that connects the scope to the signal source. In this course most often your probe will be nothing more complex than a wire, so the choice should be 1X Voltage. Note that this is not the factory default choice (which is 10X Voltage). So one of the first things you should do on turning on a scope, is check that the the probe actually attached to the scope matches what the scope thinks is attached to the scope. (If there is a mis-match, all scope voltage measurements will be wrong.) There m is a probe check button 3 on the scope to help you establish the attenuation of an unlabeled probe, but usually probes are labeled and it is faster just to immediately set the probe type yourself in the corresponding channel menu. Note: most probes used in this class have a switch to select either 10 or 1 attenuation. All SJU probes are voltage probes. Input FYI: The name "10" on a probe is quite confusing: "10" would be a better name as the voltage that reaches the scope has been reduced by a factor of 10. Why reduce a signal before measuring it? Because it reduces the probe's impact on the circuit it is connected to. A 10 probe has a larger impedance (smaller capacitance and larger resistance) than a 1 probe and hence affects the circuit less. This is particularly important for high frequency measurements (which are not the focus of this class). Horizontal Section In the center-right of the scope face, find the horizontal section. Just as in the vertical sections, there are knobs that control the horizontal scale (sec/div m m 15 ) and horizontal position 18 . In a single time-base scope, all the input channels must be displayed with the same horizontal scale (unlike the vertical scale). In this dual time base scope, a portion of the display can be expanded in a Window. The window controls are m found in the horiz menu 17 . When using the window feature the Main sec/div setting is labeled M and the Window sec/div setting is labeled W. Trigger Section As you might guess, the process of determining when to trigger and display the next wave-section is the most complex part of a scope. Luckily most often the default settings will work OK. Generally you will want to trigger when the wave has m m reached a particular level 23 . But which wave? The trig menu 22 allows you to m set the Source: CH1, CH2, Ext (the signal connected to the ext trig BNC 14 in the horizontal section), Ext/5 (the same signal, but first attenuated by a factor of 5--useful for larger triggering signals), or AC Line (which uses the 60 Hz, 120 V receptacle power line as the triggering signal--useful for circuits that work synchronously with the line voltage). Just as in the vertical section, the Coupling of this source to the triggering electronics can occur in a variety of ways: subtract the dc offset (AC), filter out (attenuate or remove) high frequency (HF Reject, "high" means > 80 kHz), filter out low frequency (LF Reject, "low" means < 300 kHz), use hysteresis to reduce the effects of noise (Noise Reject), or directly connected (DC). Note: triggering with CouplingAC is a common choice as then a level of zero is sure to match the wave at some point. Similarly Noise Reject is not an uncommon choice. The above options go with TypeEdge. There are additional sophisticated and useful triggering modes for TypePulse and TypeVideo. 1X 1 or 10 Switch 10X 30 Electrical Measurements Review m Measure Menu The measure menu 36 allows up to five measurements to be continuously updated and displayed. Push on one of the option buttons and a new menu is displayed allowing you to set the Source: CH1, CH2, MATH, and the Type: Freq, Period, Mean (voltage), Pk-Pk (peak-to-peak, i.e., the full range of voltage from the lowest valley to the highest peak), Cyc RMS (the root-mean-square voltage of the first complete cycle), Min (minimum voltage), Max (maximum voltage), Rise Time, Fall Time (10% to 90% transitions), Pos(itive) Width, Neg(itive) Width (using the 50% level). Warning: Unlike a DMM on AC, the option Cyc RMS does not subtract the d.c. offset before calculating rms voltage. You can assure yourself of a DMM-like rms result only if the channel is switched to CouplingAC. If a measurement is displayed with a question mark, try switching scales. (Generally the scope wants signals that are several divisions high and complete at least one--but not too many--cycles in the display.) m Cursor Menu The cursor menu 37 enables a pair of TypeAmplitude or TypeTime measuring lines. With Amplitude cursors, a pair of horizontal lines ("cursors") appears. Hitting the appropriate option button allows the multifunction knob to move each cursor up or down to the required place. The voltage for each cursor is displayed along with the difference (V ). With Time cursors, a pair of vertical lines appears. Hitting the appropriate option button allows the multifunction knob to move each line (cursor) right or left to the required place. The voltage and time for each cursor is displayed along with the differences (t, V ), and frequency 1/t. Display Values The bottom of the display is used to report key numerical values like scale settings. A typical example: CH1 500mV CH2 2.00V M 1.00ms 23-Nov-07 13:03 CH1 0.00V 1.01407kHz The first two numbers of the first line are the volts/div for channels CH1 and CH2, M refers to the main time-base of 1 ms/div, and the final sequence reports that positive edge triggering at a level of 0.00 V is being used with channel 1 as the source. The second line shows the date/time and the frequency of the triggering signal. Run/Stop In normal operation, the scope is constantly updating the display. It is possible to freeze the display (i.e., take a snapshot of the voltage vs. time graph) using the m m run/stop 44 or single seq 43 buttons. Hit Me First The scope remembers option settings between uses. Thus unless you are the sole user of the scope it is wise to set it to a well defined initial state before proceeding. The m default setup 41 achieves this goal (but it sets the Probe10X Voltage, in the channel m menus--which is not usually desired in this class). Similarly the autoset 42 button will attempt to make rational choices for scale factors, etc. given the signals connected to the Electrical Measurements Review 31 m scope. If you want you can save commonly used setups using the save/recall 34 menu button. Problems 1. (a) A 1.5 V battery can be modeled as an ideal 1.5 V voltage source in series with a 1 resistor. If this battery is connected to a 10 resistor, what voltage is actually across the 10 resistor? (b) A 1 mA current source can be modeled as an ideal 1 mA current source in parallel with a 1 M resistor. If this source is connected to a 10 k resistor, how much current actually flows through the 10 k resistor. 1.5 V battery 1 M 1 mA 1 10 + - 10 k 1.5 V power supply current source 2. (a) A voltage source produces a voltage V when its terminals are disconnected (open circuit). When a device that draws a current I is connected across its terminals, the voltage decreases to V - V . What is the internal resistance? (b) A current source produces a current of I when a wire connects the terminals. When a device is instead connected to the terminals the current drops to I - I and the voltage across the terminals is V . What is the internal resistance? 3. The manual for a stereo amplifier warns that it can be damaged if its "outputs are too heavily loaded". What sort of resistor would constitute a "heavy load": (A) R = 1 M or (B) R = 1 ? Explain! 4. (a) A current of 3 mA flows into a circuit consisting of 3 resistors, and 10 mA flows out. Report the readings on the three voltmeters. Draw a schematic diagram showing which lead on each voltmeter is the `red' lead. (b) An unknown device is connected in a circuit with a 9 V battery and a 15 k resistor. The ammeter reads 0.5 mA. What does the voltmeter read? V1 3 mA 5.6 k 1.8 k 10 mA V2 2.7 k A V3 device + - 15 k 9V V 5. What is the (warm) resistance of a 100 W light bulb operating from a 120 V source? 6. An ideal ammeter should act like a wire and hence have zero volts between its terminals. However real ammeters are less than perfect. The specifications for Keithley 196 in d.c. amps mode reports it has a voltage burden of about .15 V when measuring 100 mA on the proper scale. If you use a 196 to measure the current through a 15 resistor powered by a 1.5 V battery, what current does it read? 32 Electrical Measurements Review 7. Which light bulb in the below left circuit shines the brightest? Why? Which light bulb shines the dimmest? Why? A B parallel D C series V A + - A V + - 8. (a) An ammeter and a voltmeter are connected is series. Are either likely to be damaged? Why? Will either read the current or voltage of the battery? Why? (b) An ammeter and a voltmeter are connected in parallel. Are either likely to be damaged? Why? Will either read the current or voltage of the battery? Why? 9. The below left picture shows a circuit in which a battery powers a light bulb. (a) Make a careful drawing showing how the voltage produced by the battery could be measured. Include details like exactly where the red and black leads on the voltmeter would be attached. (b) Make a careful drawing showing how the current produced by the battery could be measured. Include details like exactly where the red and black leads on the ammeter would be attached. Keithley 169 a.c. volts mode D Cell 1.5 V Battery 100 pF 1 M V 10. The specifications for a Keithley 169 say that when operating in a.c. volts mode, the inputs look like 1 M in parallel with 100 pF. At what frequency is the current equally shared by the capacitor and the resistor? 11. Oscilloscope -- True or False: (a) When the vertical input coupling is set to dc mode, the voltage of an a.c. waveform cannot be measured. (b) When the vertical input coupling is set to ac mode, the voltage of a battery cannot be measured. 12. A typical lab power supply has knobs labeled Voltage Adjust and Current Adjust. If you turn the voltage knob the output voltage changes, but if you turn the current knob nothing seems to change and the current meter continues to read zero. Explain! 13. A function generator has an output impedance of 50 and, when unloaded and adjusted to produce its maximum output, produces a voltage amplitude of 10 V. What is the maximum power that can be transferred to an external device attached to the function generator. Electrical Measurements Review 33 14. A circuit consists of an inductor (inductance L) connected directly to a 120 V, 60 Hz wall receptacle. What is the smallest L you could use and avoid blowing the 20 A fuse? A similar circuit consists of a capacitor connected directly to a wall receptacle. What is the largest C you could use and avoid blowing the fuse? 15. Manufacturers typically report DMM errors as a percentage of the reading plus a certain number of "digits". In this context, one digit means a 1 in the rightmost displayed digit and zeros everywhere else. Consider a DMM display: 0.707. Find the absolute error in this reading if the device is: (a) MeTex-3800 DC current, 2 mA range. (b) MeTex-3800 AC current at 500 Hz, 2 A range. (c) Keithley 169 resistance, 2 k range. (d) Keithley 169 AC volts at 2 kHz, 2 V range. The specification sheets can be found posted in the lab or in the manuals in the physics library. Please note that errors should be rounded to one or two significant figures. 16. Work the previous problem assuming the display reads: 0.007 17. The section describing oscilloscope controls identified controls with a circled number m like: 1 . On the class web site, find and print the file TDS1024Bscope.pdf which is a line drawing of an oscilloscope. On this hardcopy, locate every control and label each with the appropriate number. 18. The below left diagram shows a single sinusoidal scope trace. Determine: the peak-topeak voltage, the voltage amplitude, the rms voltage, the wave period and frequency. Assume that the bottom of the scope display reads: CH1 500mV M 5.00ms Ext 0.00V 19. The above right diagram shows a pair sinusoidal scope traces. Assume that the scope controls are set as in the previous problem with CH 2 (dotted) and CH 1 (solid) identical. Which trace is lagging: dotted or solid? What is the phase shift in degrees? 34 Electrical Measurements Review 2: Electrical Measurements Lab 35 Electrical Measurement Lab DC & AC Measurements Impedance & Filters Work individually please! 0. Basic DC measurements. 1. Select a resistor from the "370 Resistors" cup. Sketch your resistor carefully recording the color of the bands on it. Decode the color bands to find the manufacturer's reported resistance. 2. Using a Keithley 169 DMM, measure the resistance of your resistor. Record the result with an error. Using a Metex 3800 DMM, measure the resistance of your resistor. Record the result with an error. Are your results consistent? 3. Find one of the dual battery packs with black and red leads attached. Using a Keithley 169 DMM, measure the voltage of the pack. Record the result with an error. Using a Metex 3800 DMM, measure the voltage of the pack. Record the result with an error. Are your results consistent? 4. Calculation: R = V /I Measure the voltage across your resistor with a Keithley 169 DMM (record the result with an error). Simultaneously measure the current through the resistor with a Metex 3800 DMM (record the result with an error). Calculate the resistance of the resistor along with its error. Is your calculated resistance consistent with those measured in #2? (If it isn't consult your lab instructor.) I bet the voltage measured here is less than that in #3 above. This is a consequence of the internal resistance of the battery pack that I call voltage droop and also related to the voltage burden of the ammeter discussed in problem 6 and measured in #8 below. You are now done with the battery pack. A power supply R V 5. Find a lambda power supply. Record its model number. Turn all three of its knobs to mid-range values; leave the terminal plugs unconnected. Notice that the terminal (white), - (black). Which terminals should you use plugs are labeled: + (red), to have the power supply function like the battery pack? What is the purpose of the third terminal? 6. Plug-in/turn on the power supply. Switch the meter switch to voltage and adjust the voltage adj knobs (both course and fine) until the meter reads approximately the same as #3 above. Now attach a DMM to the power supply. Using the fine voltage adj knob try again to match the output of the battery pack. (It need not be perfect.) 7. Reform the circuit of #4 now using the power supply adjusted to match the battery pack. Measure of the voltage across and current through your powered resistor and compare to those obtained in #4 above. Why the difference? 8. Voltage Burden of Ammeter. Move the voltmeter in the above circuit so as to measure the voltage drop across the ammeter (rather than your resistor). Report the result. Electrical Measurements Lab 37 9. Disconnect the DMMs, and make a circuit where the power supply is directly attached to the resistor. Holding the resistor between your fingers carefully and slowly raise the voltage produced by the power supply. At some point (typically > 20 V) the resistor should start to warm up. Using the power supply meter, record the voltage that produced a noticeable heating effect. Calculate the power (watts) required for this heating effect. (No need to calculate error.) 10. Return the power supply to the approximate voltage level produced by the battery pack and disconnect your resistor. Switch the meter switch to current. Use a banana plug wire to short circuit the output of the power supply. (Note that short circuiting most devices will result in damage. These lab power supplies are [I hope] an exception to this rule.) Notice that the current adj knob now allows you to set the current through the wire (and of course the voltage across the wire is nearly zero). As long as the output voltage is less than the set voltage limit, the current adj knob is in control. If (as before) the current is below the set current limit, the voltage adj knob is in control. Thus if the voltage limit is set high, you have an adjustable current source; if the current limit is set high, you have an adjustable voltage source. 11. Voltage divider. Set up the potentiometer circuit shown and power it with the lambda power supply. Measure Vout with a DMM. This is a classic voltage divider. Notice that by adjusting the pot, any fraction of the supplied voltage can be produced. + - Vout 0. Basic AC measurements. For this and following sections you need not report errors in measurements, but always record every digit displayed on the DMM or scope and use devices on proper scales/ranges. 1. Wavetek 180 as ac power source Set up the circuit shown using your resistor and your Wavetek 180 (as the ac voltage source). Start with all the Wavetek knobs fully counter-clockwise. (This is basically everything off and sine wave function selected.) Using the big knob and the freq mult knob, set the frequency to 1 kHz. This will also turn on the Wavetek. V 2. Adjusting the amplitude knob will change the output voltage. What range of voltages can be obtained using the hi BNC output? How about the lo BNC output? 3. Set an output voltage of approximately 1 V. Increase the frequency to find the highest frequency for which the reported voltage remains in the range 0.951.05 V (i.e., within 5% of the set value). Report the voltmeter you're using, its reported voltage, and the frequency. Further increase the frequency until the voltmeter reports a voltage under 0.5 V. The apparent change in voltage you are seeing is really mostly due to the failure of the voltmeter to work at high frequency. As you'll see below the Wavetek's output really is fairly constant as the frequency is varied. 38 Electrical Measurements Lab 4. With the output voltage still set at approximately 1 V, set the frequency to 100 Hz, and measure the ac current through and the ac voltage across your resistor. (Careful! The required circuit is analogous to #4 above. You might want to have your instructor check your circuit before powering up.) Calculate the resistance (no error calculation required). Your result should be consistent with previous resistance measurements. Measure the current again at a frequency of 1 kHz. Does the current (and hence resistance) vary much with frequency? 5. Replace your resistor with a capacitor. With the 100 Hz, 1 V output, measure the current and voltage. Calculate the "resistance" (actually reactance, i.e., V /I). Using your calculated reactance, calculate the capacitance (see Fig. 1.1 on page 20 if you're unsure of the definition of C). Measure again at a frequency of 1 kHz. Does the current (and hence reactance) vary much with frequency? Again calculate the capacitance--it should be nearly the same even though the current should be nearly ten times bigger. 0. Basic scope measurements. Note: A "scope trace sketch" should include: horizontal and vertical scale settings (and the size of a div on your sketch--this is one place where quad ruled notebook paper helps!), the location of ground (zero volts), and the signal frequency. All of these numbers are on the scope display. In addition report the type of Coupling being used on the displayed channel. Feel free to report "same settings as previous" if that is the case. 1. Set up the scope and a DMM to monitor the function generator as shown. Set the function generator to produce a 1 V sine wave at a frequency of 1 kHz. Turn on the scope and hit default setup. You will probably be using a 1 probe; make sure the scope knows that. Hit autoset. Fiddle (or not) with the controls until you have a nice stable display of the signal on the scope. Sketch the scope trace. scope V 2. Record the size (in units of divisions on the display) for: the peak-to-peak voltage (Vpp ), the amplitude or peak voltage (Vp ), and the period (T ) of the wave. Convert these to physical units (volts, seconds) using the scale factors. Calculate the frequency from the period and compare to the set value. Calculate the rms voltage from the peak voltage and compare to the DMM value. Hit the measure menu and arrange the simultaneous display of Freq, Period, Pk-Pk voltage, Cyc RMS voltage, and Max voltage. Compare these values to those you calculated from divisions. 3. Run the function generator frequency up to and beyond the limiting value determined in previous section, #3. Notice that the scope (correctly) shows a constant amplitude even as the voltmeter (incorrectly) shows a varying voltage. 4. Return the frequency to 1 kHz. Vary the function produced by the function generator: try triangle and square waves. For each wave, record the scope-reported peak-to-peak and rms voltages. Also record the rms voltage reported by the DMM. For a square wave: Vrms = Vpp /2; for a triangle wave: Vrms = Vpp /2 3. Compare these calculated rms voltages to those directly reported by the scope and DMM. Which device appears to be a "true rms" meter? Explain. Electrical Measurements Lab 39 5. In the measure menu change the measurement of Period to Mean (voltage). Find the dc offset knob on the function generator. Monitor the scope display and produce a 1 V amplitude sine wave with a 1 V dc offset. Sketch the resulting scope trace. Record the DMM ac voltmeter reading; record the scope Cyc RMS voltage. Switch the DMM to dc volts and record the dc voltage the reading; record the scope Mean voltage. Switch the scope's input to CouplingAC, and repeat the above sketches and measurements. Comment on the results. 0. Filters In many circumstances in electronics we are concerned with the ratio of voltages (or powers, currents, etc). For example the aim of an amplifier is to produce a Vout Vin and one is usually primarily concerned with the amplification (or gain), which is the ratio: A Vout /Vin . Such ratios are most often reported in decibels (dB): 20 log10 (Vout /Vin ) (2.1) Thus if an amplifier increases the voltage by a factor of 10 (i.e., A = 10), we would say the amplifier has a gain of 20 dB. Some common examples: amplification A dB 100 40 10 20 2 6 2 3 1 0 attenuation A dB 1 -40 100 1 -20 10 1 -6 2 1 -3 2 1 0 For the following measurements you will use a sinusoidal signal with no dc offset and a very large range of frequencies and voltages: Remember to adjust your scope scales appropriately. Your voltage measurements are most easily done as Pk-Pk or Cyc RMS (it doesn't matter which, just be consistent) with CouplingAC. 1. Solder together a lead from your resistor and a lead from your capacitor. 1 "BoDee" plots were popularized by Hendrik Wade Bode's book Network Analysis and Feedback Amplifier Design 1945. Bode worked at Bell Labs and became head of the lab's Mathematics Department in 1944. 2 Note: The general definition of f-3dB is not that the gain is .707, rather that the gain has changed from some standard value by a factor of .707, i.e., the gain is 3 dB less than `usual'. In #2 the `usual' A is 1, so f-3dB is where A = .707 In #3 the `usual' A is .5, so f-3dB is where A = .354 An electronic filter aims to pass certain frequencies and attenuate others. For example, a radio antenna naturally picks up every frequency; an electronic filter must attenuate all but the desired frequency. We start here with a low pass filter : a filter that lets low frequencies pass nearly untouched, but attenuates high frequencies. It is impossible to have step changes in attenuation, so as the frequency is increased, A goes smoothly to zero. Traditionally the curve of A vs. f is presented as a log-log plot and is called a Bode1 plot. Filters are usually characterized by their "-3 dB" frequency (f-3dB ), i.e., the frequency2 at which A = 1/ 2 .707. 40 Electrical Measurements Lab 2. RC low pass filter. Use your RC combination to construct the low pass filter shown. You will monitor the input to the filter (Vin ) on ch 2 of the scope and the output (Vout ) on ch 1. Describe how this filter's `gain' (but here A < 1, so `attenuation' might be a better word) changes as the frequency is varied from 50 Hz to MHz. Find 1 the filter's f-3dB (at which Vout = Vin / 2 ). Make a table of attenuation (Vout /Vin ), dB, and f measured at frequencies of about 0.1, 0.2, 1, 2, 10, 20, and 100 times f-3dB . (Note that if, by adjusting the amplitude on the Wavetek you make Vin = 1, you won't need a calculator to find Vout /Vin .) By hand, make a log-log graph of attenuation vs. f using the supplied paper. Check that the filter's Bode-plot slope for f > f-3dB is 6 dB per octave (or, equivalently, 20 dB per decade). Sketch (together) the scope traces at f-3dB . Does Vout lead or lag Vin ? (Note: a RC low pass filter is behind your scope's HF Reject option; a RC high pass filter is behind your scope's CouplingAC option.) 3. LC Filter I've constructed the beastie shown. (For the curious, it's a 5-pole, low-pass Butterworth filter in the -configuration: see Appendix H of Horowitz and Hill if you 1 want more details.) At low frequencies the gain never exceeds about 2 ; therefore f-3dB is defined as the frequency at which A is .707 its low frequency value, i.e., A = .707 .5 = .354. Using a sine wave as input, take data of attenuation vs. frequency and plot them on log-log paper. Use frequencies of about 0.01, 0.1, 0.2, 0.5, 1, 2, 4, and 8 times f-3dB . (At high frequencies the attenuation is so large that Vout may be lost in the noise unless you increase Vin to well above 1 V.) How does the steepness of this filter's cut-off compare with that of the simple RC filter? (I.e., compare the dB-per-octave falloff of this filter to the simple RC filter.) Vin 500 10 mH 10 mH Vout Vin Vout .01 F .033 F .01 F 560 4. Tape your resistor+capacitor combination in your notebook. 3: Thermionic Emission Purpose While we think of quantum mechanics being best demonstrated in processes that show discontinuous change, historically quantum mechanics was first revealed in systems where a large number of particles washed out the jumps: blackbody radiation and thermionic emission. In this lab you will investigate these two phenomena in addition to classical space-charge limited electron emission: Child's Law. Introduction Metals, as demonstrated by their ability to conduct an electric current, contain mobile electrons. (Most electrons in metals, particularly the "core" electrons closest to the nucleus, are tightly bound to individual atoms; it is only the outermost "valence" electrons that are somewhat "free".) These free electrons are generally confined to the bulk of the metal. As you learned in E&M, an electron attempting to leave a conductor experiences a strong force attracting it back towards the conductor due to an image charge: Fx = - e2 40 (2x)2 (3.1) where x is the distance the electron is from the interface and e is the absolute value of the charge on an electron. Of course, inside the metal the electric field is zero so an electron there experiences zero (average) force. You can think of these valence electrons as bouncing around inside a box whose "walls" are provided by the image-charge force. (Odd to think: the "walls" are non-material force fields; the "inside" of the box is filled with solid metal.) Since temperature is a measure of random kinetic energy, if we increase the temperature of the metal, the electrons will be moving faster and some will have enough energy to overcome the image-charge force (which after all becomes arbitrarily small at large distances from the interface) and escape. This is electron "evaporation". The higher the temperature the larger the current of escaping electrons. This temperature induced electron flow is called thermionic emission. Starting in 1901, Owen Richardson studied this phenomenon and in 1929 he received the Nobel prize in Physics for his work. A hot wire will be surrounded by evaporated electrons. An electric force can pull these electrons away from the wire -- the larger the electric force, the larger the resulting current of electrons. The precise relationship between the voltage and the resulting current flow 41 42 Thermionic Emission electric current density JA V=0 cathode metal vacuum VA anode accelerating electrons x x=0 x=b Figure 3.1: A planar cathode and a planar anode are separated by a distance b. A positive potential difference VA attracts electrons from the cathode to the anode, so the speed of the electrons v(x) increases as they approach the anode. The moving electrons constitute an electric current from anode to cathode. The resulting steady current density is called JA . is called Child's law1 (or the Child-Langmuir law, including Langmuir who independently discovered it while working at G.E.). In this experiment you will measure both Child's Law and the Richardson Effect. Child's Law Consider a planar interface between a metal (x < 0) and "vacuum" (x > 0). Vacuum is in quotes because this region will contain escaped electrons--a `space charge'--rather than being totally empty2 . The number of electrons per volume (i.e., the number density) is denoted by n. In this experiment, the metal will be heated (i.e., its a `hot cathode' or filament) which will result in a supply of electrons `evaporated' from the metal into the vacuum. An additional conducting sheet (the anode) is located at x = b. A positive potential difference, VA , between the cathode and the anode plane provides a force pulling these electrons from the vicinity of the cathode towards the anode. The result is a stream of moving electrons (a current); the number density n(x) and speed v(x) of these electrons will depend on location, x, between the plates. The negatively charged electrons moving to the right constitute a steady electric current density to the left, i.e., a steady conventional electric current from the anode to the cathode: J = -en(x)v(x) = -JA (3.2) Since the electrons leave the metal with (nearly) zero speed at zero potential, we can calculate their speed along the path to the anode using conservation of energy: 1 mv 2 - eV (x) = 0 (3.3) 2 v= 1 2 2e V (x) m (3.4) Clement Dexter Child (18681933) Born: Madison, Ohio, A.B. Rochester, Ph.D. Cornell In fact a perfect vacuum is not possible, so the word "vacuum" actually refers simply to a region with relatively few particles per volume Thermionic Emission 43 where V (x) is the potential difference ("voltage") at x and m is the mass of an electron. Because the accelerating electrons constitute a steady current (i.e., JA doesn't depend on position), n(x) must decrease as the electrons speed toward the anode. The varying space charge density affects the electric potential in the "vacuum" according to Poisson's equation3 : 2V (x) en(x) =- = (3.5) 2 x 0 0 Putting these pieces together with have the differential equation: d2 V JA = = 2 dx 0 v(x) 0 JA 2e m (3.6) V (x) Since the electric field will be zero at the interface, we have a pair of initial conditions: V = 0 x x=0 V |x=0 = 0 This differential equation looks a bit like Newton's second law: d2 x 1 F (x(t)) = 2 dt m as you can see if in Newton's second law you substitute: t - x (3.9) (3.7) (3.8) x(t) - V (x) JA 1 F (x(t)) - m 0 2e V (x) m Recall that force problems are often most simply solved using conservation of energy and that conservation of energy was proved using an integrating factor of dx/dt. If we try the analogous trick on our voltage problem, we'll multiply Poisson's equation by dV /dx: dV d2 V dx dx2 1 dV 2 dx 2 = 0 = 0 JA 2e m V -2 V 1 2 1 dV dx (3.10) JA 2e m 1 2 (3.11) 1 dV 2 dx 2 = 0 JA 2e m V 1 2 1 2 + constant (3.12) The initial conditions require the constant to be zero, so 1 dV 2 dx 3 2 = 0 JA 2e m V 1 2 1 2 (3.13) Poisson's equation is derived in the Appendix to this lab. 44 or Thermionic Emission dV = dx This differential equation is separable: dV V V 3 4 1 4 3 4 4JA 0 2e m V 1 4 (3.14) = 4JA 0 2e m dx (3.15) = 4JA 0 2e m x (3.16) where again the initial conditions require the constant of integration to be zero. Finally: 2 3 4 9JA x 3 V (x) = (3.17) 40 2e m Of course, V (b) is the anode voltage VA , so we can rearrange this equation to show Child's law: 3 2e 40 2 (3.18) VA JA = 9b2 m Much of Child's law is just the result of dimensional analysis, i.e., seeking any possible dimensionally correct formula for JA . Our differential equation just involves the following constants with dimensions (units) as shown: b : L E M L2 /T 2 VA : = Q Q 0 2e k : m JA : Q2 Q2 Q 2 1 = 3 EL M 2 M 2 L3 /T 2 Q/T L2 1 5 (3.19) (3.20) (3.21) (3.22) where the dimensions are: L=length, T =time, M =mass, E=energy, and Q=charge. To make a dimensionally correct formula for JA , we just eliminate the M dimension which we can only do with the combination: VA k 2 3 : Q3 T3 2 2 (3.23) We can then get the right units for JA with: VA k 3 b2 2 3 2 = k 3 V2 b2 A k 3 V2 b2 A : Q/T L2 (3.24) Thus the only possible dimensionally correct formula is JA (3.25) Thermionic Emission cathode: hot filament, radius a 2 45 anode anode l cathode 1 4 b Figure 3.2: Coaxial cylinders: an inner wire (radius a) and outer cylindrical anode (radius b), form a vacuum tube diode. The cathode is heated so electron evaporation is possible, and a potential difference VA attracts electrons from the cathode to the anode. The speed of the electrons v(r) increases as they approach the anode. The moving electrons constitute a steady electric current from anode to cathode. Since the same current is spread out over larger areas, the current density, J, between the cylinders must be proportional to 1/r. The exact proportionality constant, found from the differential equation, is (as usual) is not hugely different from 1. We have derived Child's law for the case of infinite parallel plates, but you will be testing it in (finite length) coaxial cylinders. The inner wire (radius a) is the cathode; the outer cylinder (radius b) is the anode. Your cylinder with have some length , but we will below consider infinite length coaxial cylinders. Note that dimensional considerations require that the anode current per length should be given by a formula like: I/ j 3 k 2 VA b (3.26) although we could have an arbitrary function of the radius ratio: b/a on the right-hand-side. From Poisson's equation4 we have: 2 V = J I j = = 0 v(r) 2r0 v(r) 2r0 V -2 1 2e m (3.27) Using the Laplacian in cylindrical coordinates we find: 2V j 1 V = + 2 r r r 2r0 V -2 1 2e m (3.28) There is no known formula for the solution to this differential equation, but we can make considerable progress by writing the differential equation in terms of dimensionless quanti4 Poisson's equation is derived in the Appendix to this lab. 46 ties: Thermionic Emission r/a = V = yielding: (3.29) ja 20 2e m 2 3 f () (3.30) 1 1 f 1 1 2f + = f () + f () = f - 2 2 (3.31) with initial conditions: f (1) = 0 f (1) = 0 We can numerically solve this differential equation using Mathematica: NDSolve[{f''[p]+f'[p]/p==1/(p Sqrt[f[p]])}, f[1]==0, f'[1]==0, {f},{p,1,200}] It's actually not quite that simple. The cathode, at = 1, is actually a singular point of the differential equation (i.e., f (1) = ). However the situation very near the cathode is well approximated by the planar case, where we've shown: 9I 9JA 4 V (x) = x3 = 40 2e 2a40 m 2 3 2 9 3 r-a ja = 4 a 2e 2 0 m (3.32) (3.33) 2 3 2e m 4 3 2 3 4 3 (r - a) = 9ja 240 2e m 2 3 r-a a 4 3 (3.34) So, near the cathode (i.e., slightly larger than 1): 9 f () 4 2 3 ( - 1) 3 4 (3.35) We can use this approximation to start our numerical differential equation solution at a non-singular point (like = 1.00001). Real devices are designed with b/a 1. The behavior of f for large can be determined by finding A and for which f = A is a solution to the differential equation. One finds: f= 9 4 2 3 (3.36) A useful approximation for the range: 100 < b/a < 1000 is: f= 9 4 2 3 +2 (3.37) Thermionic Emission f() for 1 f() for 47 10 60 f 8 50 40 30 20 6 4 f() exact 2 10 50 100 150 200 2 4 6 8 10 12 14 Figure 3.3: The plot on the left displays the dimensionless voltage f obtained by numerical solution to the differential equation. The plot on the right compares various approximations for f to this numerical solution. (For example, the device used in lab has b/a = 121.5. For this value, the differential equation gives f = 44.136; the above approximation gives: f = 44.130.) We recover Child's law by rearranging (3.30): 20 a 2e m VA f (b/a) 3 2 = j = I/ (3.38) So: Note: Langmuir's original work (Phys. Rev. 22, 347 (1923)) on this subject is expressed in terms of where: - 3 1 ( - 1)2 2 4 f 2 () = (3.39) - 9 1 80 9b 2 2e m 3 2 VA = I (3.40) 2 = 1.072 for the device used in lab. Richardson's Law Most any thermal process is governed by the Boltzmann factor: exp - E kT = e-E/kT (3.41) where k is the Boltzmann constant. Approximately speaking the Boltzmann factor expresses the relative probability for an event requiring energy E in a system at (absolute) temperature T . Clearly if E kT , the probability of the event happening is low. If an electron requires an energy W (called the work function) to escape from the metal, The 48 Thermionic Emission Boltzmann factor suggests that this would happen with relative probability e-W/kT . Thus you should expect that the current emitted by a heated metal would follow: I e-W/kT (3.42) Clearly you should expect different elements to have different work functions, just as different atoms have different ionization potentials. What is surprising is that the proportionality factor in the above equation includes a universal constant -- that is, a constant that just depends on the properties of electrons (and, most importantly, Planck's constant, h) and does not depend on the type of material. (This situation is similar to that of blackbody radiation, in which photons rather than electrons are leaving a heated body, which was Planck's topic in discovering his constant. We will take up this topic on page 53.) Thermionic emission probes the quantum state of the electrons statistically, whereas the photoelectric effect probes much the same physics electron by electron. (The next level problem is to explain why this universal constant (the Richardson constant, A) in fact does depend a bit on the material.) To show: J = AT 2 e-W/kT (3.43) where A= 4emk2 = 1.2 106 A/m2 K2 h3 (3.44) Quantum Theory: Free Electron Gas Instead of thinking about electron particles bouncing around inside a box, de Broglie invites us to consider standing waves of electron probability amplitude: = N exp(ikx x) exp(iky y) exp(ikz z) = N eikr (3.45) Recall 5 that vector k is the momentum, p = mv, of the electron and = h/2. Periodic boundary conditions on the box (which we take to be a cube with one corner at the origin and the diagonally opposite corner at the point r = (L, L, L)) require each component ki to satisfy: 2 ni ki = (3.46) L where each ni is an integer. Thus each wave function is specified by a triplet of integers: n = (nx , ny , nz ), the n-vector. Applying Schrdinger's equation, we find that this wavefunction o has energy: 2 k2 (2 )2 (n2 + n2 + n2 ) (2 )2 n2 x y z = = (3.47) E(n) = 2m 2mL2 2mL2 Notice that there is a definite relationship between the velocity vector of the electron and the n-vector. 2 n (3.48) v= mL Another way of saying the same thing is that allowed quantum-state velocities form a cubic lattice with cube-side 2 /mL. The number of states with electron velocities in some specified region (for example a velocity-space parallelepiped with sides: vx vy vz ) can 5 For a review see: Thermionic Emission 49 vacuum metal vacuum image-charge potential unoccupied levels U W occupied levels 0 EF zero-force (flat) potential inside metal Figure 3.4: Electrons in the metal experience a constant confining potential of depth U . Possible quantum mechanical states for these electrons are displayed as horizontal lines. Electrons fill all the available states up to the Fermi energy, EF . The work function, W , is defined at the minimum energy needed to remove an electron from the metal. As shown above: W = U - EF . be found from the number of 2 /mL sided cubes that fit into the region, which is the volume of that velocity-space region divided by (2 /mL)3 . Hence: number of states with velocity between v and v + v = vx vy vz (3.49) (2 /mL)3 vx vy vz number of states per volume with velocity between v and v + v = (2 /m)3 m 3 vx vy vz = N vx vy vz (3.50) = 2 where N is the (constant) density of states in velocity space. Quantum Theory: Fermi Energy Fermions (half-integer spin particles), in contrast to bosons (integer spin particles), cannot 1 group together. Since the electron is "spin 2 ", each of the above states can hold at most 2 electrons: one spin up and one spin down. The probability that a particular fermion state with energy E will be occupied is given by a generalization of the Boltzmann factor called Fermi-Dirac statistics: 1 (3.51) f (E) = 1 + exp E-EF kT where EF is called the Fermi energy. The Fermi energy is basically a disguise for the number of electrons, as, approximately speaking, it is the dividing line between occupied states and unoccupied states. (If the Fermi energy is high, there must be lots of occupied states and hence lots of electrons.) Note that if E EF , the exponential factor is huge and we can 50 Thermionic Emission v x t A Figure 3.5: Consider just those electrons with some particular x-velocity, vx . In order to hit the wall during the coming interval t, an electron must be sufficiently close to the wall: within vx t. The number of such electrons that will hit an area A will be equal to the number of such electrons in the shaded volume (which is the perpendicular extension of A a distance vx t into the volume). Note that many electrons in that volume will not hit A because of large perpendicular velocities, but there will be matching electrons in neighboring volumes which will hit A. To find the total number of hits, integrate over all possible vx . neglect the "+1" in the denominator so f (E) exp - E - EF kT (3.52) that is, if E EF Fermi-Dirac statistics approximate the Boltzmann factor. Classical Theory: Electron Escape The density of states combined with the Boltzmann factor gives us the number of free electrons per unit volume with a particular velocity. In order for an electron to escape during some time t, it must have vx sufficient to overcome the image-charge barrier and it must be sufficiently close to the wall. All the electrons with vx > 2U/m within a distance vx t, will escape, where U is the depth of the potential well for the electrons. Thus the number of electrons escaping through area A during t is: 2U/m dvx - dvy - dvz 2N f (E) Avx t e-mvx /2kT vx dvx (U - EF ) kT 2 = 2N eEF /kT At = 4m(kT )2 (2 )3 - e-mvy /2kT dvy 2 - e-mvz /2kT dvz (3.53) 2 2U/m At exp - where we have used the Gaussian integral: - e-z dz = 2 (3.54) Thermionic Emission 51 Table 3.1: G.E. FP-400 Specifications Filament (W) length Filament diameter Anode (Zr coated Ni) I.D. Maximum filament voltage Maximum filament current Maximum anode voltage Maximum anode current Maximum anode dissipation 3.17 cm 0.013 cm 1.58 cm 4.75 V 2.5 A 125 V 55 mA 15 W (1.25") (0.005") (0.620") The electric current density is the electric charge escaping per time per area: J= e number escaping 4em(kT )2 W = exp - 3 At h kT (3.55) which is Richardson's equation, with work function W given by U - EF . Experiment: Richardson's "Constant" is Material Dependent! Experimentally it is found that Richardson's constant depends on the material6 Why? 1. The work function depends on temperature (due to, for example, thermal expansion of the lattice of atoms). If the data analysis assumes it's constant, the resulting A will be grossly in error. 2. Classically reflection requires a turning point (where vx = 0), whereas quantum mechanical reflections are possible just due to sharp changes in potential. Quantum mechanical reflection at the metal boundary was not included in our calculations; we assumed every energetic electron headed toward the wall went through the wall. 3. Surface contamination can affect emission probability. In fact, it was originally thought that thermionic emission was 100% due to surface contamination. (You can think of surface contamination as a locally varying work function.) 4. Even in the absence of surface contamination, in typical experiments, the metal is polycrystalline and different crystal surfaces have different work functions. Experiment This experiment involves thermionic emission from the hot tungsten filament of a G.E. FP-400 vacuum tube. 6 This should remind you a bit of the material dependent emissivity, T , for blackbody radiation to be discussed on page 53. 52 Thermionic Emission 2 anode Keithley 2400 source-meter FP-400 GND 4 1 cathode Keithley 2420 source-meter Keithley 192 voltmeter Figure 3.6: Circuit diagram (note tube pin labels) for thermionic emission experiment. Temperature Determination Often temperature measurement in physics experiments is easy. In the "normal" range of temperatures there are many types of transducers which convert temperature to an electrical quantity (e.g., Pt resistance thermometers, thermocouples, thermistors, ICs). However at the extremes of high and low temperature, measurement becomes tricky. Questions like "What exactly defines temperature?" must then be answered. This experiment requires "high" temperatures in a vacuum, so we do not have direct contact with the material whose temperature we seek. In addition the FP-400 tube was not built by us, so we have limited direct information about the device. One common way to measure temperature is by using the temperature dependence of resistance. The resistance of metals is approximately proportional to temperature. Jones and Langmuir7 have published a table of resistance vs. temperature for tungsten, from which Kirkman has found an approximating formula: Tr = 112 + 202x - 1.81x2 where x is the ratio of the hot resistance to that at 293 K. A problem with this approach is that the measured resistance, Rmeasured , will include both the resistance of the tungsten8 filament, RW and the wires supporting it in the vacuum tube, Rsupport . Thus the quantity we seek (tungsten filament resistance, RW ) must be calculated as the small difference between two numbers: RW = Rmeasured - Rsupport (3.57) (3.56) a situation that produces big relative errors. Additionally, we have no independent way of measuring Rsupport (we can't take the tube apart); In the end you will measure Rsupport at room temperature and then assume it is constant9 . There is a further problem with any measurement of voltage when parts of the system are at different temperatures: thermally induced emfs (thermocouples). If the ends of the tungsten filament are at different temperatures, there will be a voltage induced approximately 7 GE Rev 30, 310 (1927) The chemical symbol for tungsten is W from the German Wolfram 9 See "Assuming Away Variation" page 18 in Chapter 0 8 Thermionic Emission 53 Temperature Dependence of Resistance: W 3000 2500 Temperature (K) 2000 1500 1000 500 0 5 10 Resistance Ratio: R/R293 15 Figure 3.7: The temperature dependence of tungsten-filament resistance from the data of Jones & Langmuir with an approximating curve. The x-axis is the ratio of the hot resistance to that at 293 K. proportional to the temperature difference between the two ends. This additional voltage source confuses the resistance determination. The phenomena can be detected and corrected by reversing the direction of current flow (which reverses the Ohm's law voltage, but does not affect the sign of the thermal voltage.) Thermal voltages are generally less than a mV, and so are negligible once our measured voltages approach one volt. Another approach is suggested by Jones & Langmuir. In a vacuum the temperature of a current-carrying wire is the result of an equilibrium between electrical power dissipated in the wire and energy lost in the form of radiation. (We assume that energy lost through conduction -- either through the wire-supports or residual air in the "vacuum" -- is negligible.) According to the Stefan-Boltzmann law, the power radiated from the surface of a hot black-body is given by: P = T 4 A (3.58) where is the Stefan-Boltzmann constant, T is the temperature of the body, and A is the surface area of the body. (In fact tungsten is not a black-body, so when applied to tungsten the above should be multiplied by a "fudge factor", the total emissivity T , about 0.3.) Using just crude approximations, we equate the electrical power dissipated to the power radiated: (3.59) I 2 T 2 I 2 T 2 I 2 R = T T 4 d T 4 d d 4d where d is the diameter of the wire and is the length of the wire. On the right hand side we've assumed that the resistivity of tungsten is proportional to temperature, and on the left hand side we've assumed T doesn't depend on temperature. We conclude: I 2 T 3 d3 I d2 3 2 3 (3.60) (3.61) T 54 Thermionic Emission Temperature vs. Current for W Filaments 3000 2500 Temperature (K) 2000 1500 1000 500 0 500 1000 1500 a=I/d^1.5 (A/cm^1.5) 2000 Figure 3.8: The temperature of an in-vacuum tungsten filament as a function of current from the data of Jones & Langmuir with an approximating curve. The x-axis is the current 3 divided by the diameter of the wire (in cm) raised to the 2 power. Absent the above approximations we can hope that temperature is a function of a I/d 2 . Once again Jones & Langmuir provide us with calibrating data for this expected relationship. For temperatures 400 K < T < 3000 K, Kirkman finds: Ti = 117 + 56a0.5 + 0.00036a1.8 (3.62) 3 Finally, attaining thermal equilibrium10 is a problem that affects most any temperature measurement. The balance between electrical power in and heat lost is not immediately achieved. The parts of the system with large heat capacities (particularly the filament supports and other large structures in the tube), will only gradually approach equilibrium. Thus "soak" time is required following each jump in heating power. The effect of this "thermal inertia" is seen in "hysteresis": temperatures obtained while increasing the temperature disagree with those found while decreasing the temperature. This will be an important source of uncertainty. Hands-on Electrical Measurements Support Resistance: Rsupport As shown in Figure 3.6, the tungsten filament (cathode) is powered by a Keithley 2420 source-meter. The filament+support voltage is measured directly at the socket with a Keithley 192 voltmeter. By combining the current through the filament (measured from the 2420) with the voltage across the socket (from the 192), a resistance, Rmeasured , can 10 See "Special Problem: Temperature" page 17 in Chapter 0 Thermionic Emission 55 be determined. At room temperature, the filament resistance, RW , can be calculated from the filament geometry (see Table 3.1) and the resistivity of W-filament material at room temperature: 293 = 5.49 cm. (You should calculate: RW .1 .) Then: Rsupport = Rmeasured - RW (3.63) Because room temperature Rmeasured is `small' (and hence error prone), you will make three distinct measurements of it. Begin by sourcing 1 mA and then 10 mA into the room temperature filament (using the 2420), reading the resulting voltages on the 192, and calculating the corresponding Rmeasured . Follow up those two measurements with a fourterminal resistance measurement just using the 192. (If Rmeasured is substantially larger than .2 , confirm that you have firm, low-resistance contacts between the socket and the tube. Working the socket-tube contact may be required to find the lowest resistance spot.) Maximum Filament Current, 2420 Voltage Compliance Limit Following your determination of Rsupport in a room temperature tube, check out the conditions required to stay just below the tube's maximum ratings (4.75 V, 2.5 A). Using the 2420, successively source filament currents of 2.0 A, 2.1 A, 2.2 A, . . . to directly determine the maximum current you can use without exceeding the 4.75 V limit across the tube's filament. Note that at just below maximum conditions, the 2420 will probably need to produce voltages above 4.75 V (because of the resistance of the external wires: the voltage drop across the connecting wires is not zero). Record the maximum 2420 voltage and tube current allowed by the tube's ratings; you will need these numbers in step #2 of your computer program. Data Collection Plan You will be collecting two types of data at the same time: thermal characteristics of the filament and the thermionic properties of the tube (anode current/voltage relationship). Starting at a filament current of 0.9 A, increase the current flowing through the filament in steps of 0.1 A to a maximum current (found as described above, about 2.4 A) and then reverse those steps down to a filament current 1.0 A. The up-sweep in filament current followed by the down-sweep will allow you to test for hysteresis. At each step in current, allow the filament to approach thermal equilibrium (wait, say, 15 seconds) and then measure the voltage across and current through the cathode/anode. Calculate filament temperature two ways (Equations (3.56) and (3.62)). Average the two to estimate the temperature, and use half the absolute value of the difference to estimate the uncertainty. You see above a classic example of systematic error. The temperature is measured two different ways. Direct application of error propagation formulas to these temperatures calculated from 6-digit meter values would suggest small uncertainties. However the two temperatures in fact disagree. If only one method had been used to measure temperature, we would have badly underestimated the error. 56 Thermionic Emission T 4 vs. Power: Testing Stefan-Boltzmann By conservation of energy we know that the power dumped into the filament (mostly from electrical heating, but also from other sources like radiation from the room temperature environment to the filament) should equal the power out of the filament (from black-body radiation and other factors like conduction down the supports). Thus: T AT 4 = I 2 RW + constant 1 I 2 RW + constant T4 = T A y = bx + a (3.64) (3.65) (3.66) A graph of T 4 vs. power should be a straight line from which you will determine T . (Note that the error in power is quite small, so we have properly used it as an x variable.) In order to test this relationship you will want to make a file containing the filament power, 4 T 4 (use the average of the two temperatures: (Ti4 + Tr )/2), and the error in T 4 (use half 4 - T 4 |/2). the difference from the two temperatures: |Ti r IA vs. VA : Testing Child and Richardson You will collect anode current vs. voltage curves for each different filament (cathode) temperature. Use the Keithley 2400 to sweep the anode voltage logarithmically from 2 V to 120 V. (Note the maximum limits for the anode: 0.055 A or 125 V. Do not exceed either!) 3 2 According to Child's law, the anode current, IA , should increase in proportion to VA . Of course, at sufficiently high voltage the current will be limited by the maximum electron evaporation rate, and a current plateau forms at a level given by Richardson's law. At the maximum filament current (corresponding to the maximum filament temperature and evaporation rate), plateau formation occurs at very high voltage and you have the longest run of data following Child's law. Make a file containing VA , IA , and IA which you can use to fit to the Child's law functional form: IA = k1 (VA - k2 ) 2 3 (3.67) In addition, you will want to make a big continuous file containing: VA , IA at every temperature tested. The current plateaus can be extracted from this file and fit to the Richardson relationship: IA = k1 AT 2 e-k2 /T (3.68) Computer Data Collection As part of this experiment you will write a computer program to control the experiment. Plagiarism Warning: like all lab work, this program is to be your own work! Programs strikingly similar to previous programs will alarm the grader. I understand that this will often be a difficult and new experience. Please consult with me as you write the program, and test the program (with tube disconnected!) before attempting a final data-collecting run. Your program will control all aspects of data collection. In particular it will: Thermionic Emission 0. Declare and define all variables. 1. Open (i.e., create integer nicknames--i.e., iunit--for) the enets gpib0 and gpb1. 57 2. Initialize meters--requires knowing the gpib primary address--i.e., iadd--of the meter and the iunit it is attached to. Get the status of each meter after you have initialized it. (a) Each source-meter must be told the maximum voltage and current it must produce during the experiment. Initialize the 2400 (anode voltage/current) for the tube maximum ratings. (b) Initialize the 2420 (filament voltage/current) for the near tube-maximum ratings found above11 . In the following I assume the current maximum is 2.4 A, but it may be different for your tube. (c) Initialize the 192 for autorange DC voltage measurements. 3. Open the files: (a) filament.dat (intended for: If , Vf , Tr , Ti of filament). (b) stefanB.dat (intended for: power, T 4 , T 4 of filament). (c) VI.dat (intended for: all VA , IA of anode, with comments (`!') for filament If , Tr , Ti ). (d) child.dat (intended for: VA , IA , IA of anode at maximum filament current). (e) child-.dat (like above but intended for a downsweep of anode voltage). (f) rich.dat (intended for: Ti , Tr , IA -- i.e., the estimated temperatures and the corresponding maximum anode current for Richardson's Law). 4. Tell the 2420 source-meter to source a filament current of 0.9 A. 5. Let the system sleep for 60 seconds to approach thermal equilibrium. 6. Do a sequence of measurements where the filament temperature is sequentially increased (i.e., a temperature up-sweep) with filament currents ranging from 0.9 A to some maximum (e.g., 2.4 A) current in steps of 0.1 A. (a) Tell the 2420 source-meter to source a filament current (If ). (b) Let the system sleep for 15 seconds to approach thermal equilibrium. (c) Request a logarithmic sweep of the anode voltage (controlled by the 2400 sourcemeter) from 2 V to 120 V. Receive the resulting arrays: VA and IA . (d) Turn off the anode voltage. (e) Repeat (a) thus receiving an updated version of the filament current (it will be very close to the requested current). (f) Read the 192 to get the filament voltage (Vf ). (g) Using Eqs. (3.56) and (3.57), calculate Tr based on the calculated tube resistance Rmeasured , the calculated room temperature filament resistance and Rsupport . (h) Calculate Ti from Eq. (3.62) 11 See Hands-on Electrical Measurements, p. 54. Recall: the 2420 maximum voltage will need to be a bit above 4.75 V. If you have not yet completed those measurements, temporarily initialize with 4.75 V. 58 Thermionic Emission (i) Write a line to the file filament.dat reporting: If , Vf , Tr , Ti . 2 (j) Write a line to the file stefanB.dat reporting: filament power (RW If ), T 4 , and T 4 (see p. 56). (k) Write a comment line (i.e., starts with `!') to the file VI.dat reporting filament data: If , Tr , Ti . (l) Write the anode sweep data (one data-pair per line) in the file VI.dat. (m) Write a line to the file rich.dat reporting Ti , Tr , IA . Use IA at VA =120 V as the estimated plateau current. (When the experiment is over, you will need estimate IA based on hysteresis, and may need to delete IA values if, for example, the current did not plateau or if cold emission substantially added to the plateau current.) (n) Increment the filament current by 0.1 A and return to (a). 7. Collect data for Child's Law. Begin by repeating a normal anode sweep at the maximum filament current; follow all the steps (a)(m) outlined in 6 above. In addition, write the anode sweep data (VA , IA , IA ) in the file child.dat (one data-triplet per line). In this case, IA will be calculated from the manufacturer's specs: percent+digits; the fortran function eAk2400 can do this error calculation automatically. Now check for hysteresis by doing a reverse anode sweep: from 120 V down to 2 V. Write this reverse anode sweep data (VA , IA , IA ) in the file child-.dat. 8. Do a sequence of measurements where the filament temperature is decreased (i.e., a temperature down-sweep) by sequentially sourcing filament currents from one step down from maximum (e.g., 2.3 A) back to 1.0 A. Follow steps (a)(m) used in the temperature up-sweep (part 6 above) and then: (n) decrement the filament current by 0.1 A and return to (a). 9. Turn off the output of the 2420. 10. Close all files. Note that the 0.9 A filament current data is just trash collected so the 1.0 A filament current data is taken on a pre-warmed filament. Observations 1 While the computer collects the data observe the light from the filament. (There is a 16 " diameter hole in the anode allowing light from the mid-point of the filament to escape.) Initially the filament will not appear to be incandescent (i.e., not a source of light at all: dark) so it may help to turn off the lab lights to better observe the beginning of visible incandescence. According to the Stefan-Boltzmann law the light energy radiated depends on T 4 , so small changes in T produce big changes in light intensity. In addition to changes in light intensity, you should look for the more subtle change in light color: from a dull red to a brilliant yellow. Record your observations! At what filament temperature did you first see the filament producing light? Thermionic Emission 59 FP400 Vacuum Tube .01 .001 Ia (A) 1.E04 1.E05 1.E06 1.E07 1 10 Va (V) 100 Figure 3.9: The temperature dependence of thermionic emission in a FP-400 vacuum tube. Each curve shows the anode current-voltage relationship at a particular filament temperature. At high filament temperatures and low anode voltages the anode current follows Child's law: an upward sloping straight line on this log-log plot. At sufficiently high anode voltage, the filament current plateaus at a value given by Richardson's law. At low filament temperatures and high anode voltages, "cold emission" may augment (distort) the plateau. Data Analysis Beginnings The main result of this experiment is a plot similar to Figure 3.9 showing the anode currentvoltage relationship at various filament temperatures. Production of such a multi-plot is a bit complex, and you will almost certainly need further detailed instructions from your instructor on using the program Nplot. Note that since this is a log-log plot, negative anode currents must also be edited out of VI.dat to make this plot. It may be difficult to determine the plateau level for the lowest and highest filament temperatures, in which case you must edit out those data points in rich.dat. Note: If the high-temperature V I sweep reaches a plateau, then Child's Law will not apply at high VA so child.dat will require editing. If it does not reach a plateau, then Richardson's Law does not apply so rich.dat will require editing. In Figure 3.9, I see no sign of a plateau at the maximum filament current, but in the range If = 1.1 to 2.3 A, plateau currents can be determined. (I.e., I removed from rich.dat the non-plateau data point for If = 2.4 A, and also the aberrant initial data for If = 1.0 A.) Additionally a glance at the file rich.dat should demonstrate systematic temperature error: Ti = Tr . Clearly fits assuming T = Tr will produce different results from fits assuming T = Ti . Absent further information, our best estimate for T must be something like the 60 Thermionic Emission Richardson's Law .01 Anode Current (A) .001 1.E04 1.E05 2600 2400 2200 2000 1800 Temperature (K) 1600 Figure 3.10: A Richardson Plot of thermionic emission in a FP-400 vacuum tube. Each data point displays the plateau anode current at a particular filament temperature. The curve through the data fits for the work function; the slightly steeper curve uses the book value for the work function. Child's Law .01 Ia (A) .001 .0001 1 10 Va (V) 100 Figure 3.11: A plot of the space-charge limited thermionic emission in a FP-400 vacuum tube (Child's law). The data was taken at a filament current of 2.4 A. Every other data point has been eliminated so the fit line is not obscured. Note that the fit line systematically misses the data, sometimes a bit high others a bit low. The measurement errors are tiny, so these small misses do result in a "too-large" 2 . Nevertheless, the law provides an excellent summary of the data over a huge range of variation. Thermionic Emission 61 StefanBoltzmann Law 4.E+13 3.E+13 T^4 (K^4) 2.E+13 1.E+13 0 2 4 6 Power (W) 8 10 Figure 3.12: A test of the Stefan-Boltzmann law: that power radiated is proportional to T 4 . Note that the fit line hits well within each error bar. The 2 for this fit will be "small". Evidently the average temperature is a better measure of temperature than you might expect from the deviations between Ti and Tr . average of Ti and Tr . The difference between fit parameters produced assuming T = Tr and those produced assuming T = Ti will allow use to estimate the systematic error in those parameters. Now according to Richardson's law, these plateau currents should satisfy: I = AAT 2 e-W/kT = k1 AT 2 e-k2 /T (3.69) where A is the tungsten filament surface area. However, since we are currently only seeking error estimates, simplified analysis is justified. Clearly the largest error is in T , so the standard approach would be to put T on the y-axis and I on the x-axis: the opposite of what is implied by the above equation. However, we can't simply solve the above equation for T without some seemingly huge approximations. It turns out that e-k2 /T is the significant factor in the above equation, so we start by ignoring the T 2 and assume: I = a e-b/T Now if we take ln of both sides: ln(I) = ln(a) - k2 or 1 T (3.71) (3.70) 1 1 = ln(a)/k2 - ln(I) T k2 (3.72) This equation is now in a form12 known to WAPP+ . Thus you can quickly WAPP+ (I, Ti ) and (I, Tr ) data from rich.dat (for simplicity assume no x-error and no y-error), and 12 Inverse-Natural Log: 1/y = A + B log(x) 62 Thermionic Emission Work Function Systematic Error 1000 Temperature (K) 1500 2000 2500 1.E08 1.E06 1.E04 Anode Current (A) .01 Figure 3.13: Simplified analysis suggests Richardson's Law data can be fit to the InverseNatural Log relationship of Eq. 3.72. The filled-square points use y = Ti , the unfilled-square points use y = Tr . Systematically different temperature measurements yield systematically different B (slopes in the above plot) and hence systematically different work functions W = -k/B. Note the inexact pairing of the data due to temperature hysteresis. determine the range of k2 = -1/B consistent with the differing temperature data. Request WAPP+ make a linearized plot of your data with x-scale: log and y-scale: inverse as in Fig. 3.13. This should allow you to check for aberrant data points. (Usually the high temperature curve has not plateaued, and so the high temperature data point must be removed. Occasionally the low temperature curve is also aberrant.) The range for W = k2 k has been found by simplified fits; the best value for W will be found (see following) with a fit to the proper functional form with due account for the uncertainty in current caused by hysteresis and the best estimate for the actual temperature. The Stefan-Boltzmann Law is the easiest law to check: you can quickly WAPP+ the data in the file stefanB.dat to produce a plot similar to Fig. 3.12. I expect you'll find a small reduced 2 due to the large systematic error in temperature, so a bit of additional work will be required to estimate T . Child's Law At sufficiently high filament temperature (and low anode voltage), Child's law governs the anode current-voltage relationship: IA = 80 9b 2 2e m 3 2 VA (3.73) (see vicinity of Eq. 3.40, page 47, for a definition of symbols) At the highest filament temperature (e.g., filament current of 2.4 A) you have saved the (VA , IA , IA ) data in the Thermionic Emission file child.dat. Now fit and plot this data to the functional form: IA = k1 (VA - k2 ) 2 3 63 (3.74) (where k2 represents an offset between ground and the actual average voltage of the filament), producing a result similar to Fig. 3.11. Do not be surprised if you get a huge reduced 2 . Find an estimate for k1 error either by a `fudged fit' or a bootstrap. Follow exactly the same process for the downsweep data in the file child-.dat. Generally you will find that the k1 values for the two sweeps differ by more than computer-based value of k1 . Systematic error (here a type of hysteresis) is responsible. Note that the usual reduced 2 alerts us to a problem, but measuring twice (in different ways) provides an estimate (perhaps still an under-estimate) for k1 : half the difference between the two values of k1 . We expect that k1 = 80 9b 2 2e m (3.75) so with known geometry, the electron charge-mass ratio e/m can be calculated. Since the FP-400 is a finite length cylinder (rather than the infinite cylinder discussed in our theory) use the effective length13 = 0.7 as the length of the cylinder. But what can we use as errors for the `book' values for b and ? Thermal expansion of the filament should, by itself, change by about 1% (notice the spring-tensioned support for the filament in the FP-400). Based on the sigfigs in the reported values, I estimate: (b/) 3% (b/) Calculate e/m and its error. Compare to the `known' value (citation required!). (3.76) Stefan-Boltzmann Law You should have already checked for an approximately linear relationship between electrical power in and T 4 : Power = T AT 4 (3.77) and found a reduced 2 indicative of large systematic uncertainty in temperature. We now seek an estimate for T (with error) at the highest temperature. The best value for T can be found by plugging in our best estimates for T 4 and the measured power (found in the file stefanB.dat), and A (calculated based on the dimensions recorded in Table 3.1). Alternatively T could be calculated based on slope as suggested by Eq. 3.66. But how should we incorporate the large systematic errors in T 4 and the unknown systematic error in A? For the surface area A, all we know is the `book' values for the dimensions of the filament. Based on the sigfigs in the reported values, I estimate: A 10% A (3.78) This effective length corrects for current loss through the ends of the cylinder under space-charge situations. A smaller correction should be applied when large anode voltages are designed to sweep up all evaporated electrons, i.e., for Richardson's Law, where 90% electron collection seems appropriate. The details of handling such corrections to theory may be found in reference 1. 13 64 Thermionic Emission (mostly due to the filament diameter, where a `small' uncertainty leads to a large fractional uncertainty). We can then use the `high-low' method of 191 to estimate the range of possible values for T , given the range of possible values for T and A (assume the error in power is negligible). Richardson-Dushman Law You should have already checked for an approximately exponential-inverse relationship between T and IA , edited out aberrant data (e.g., not yet plateaued), and have a range of possible values (due to systematic error in temperature) for k2 in the expected relationship: IA = k1 AT 2 e-k2 /T (3.79) We now seek a treatment incorporating the hysteresis error in IA and the proper functional form, to obtain the best possible value for k2 . We will need to manipulate14 the data in rich.dat to bring together equivalent data from the temperature upsweep and downsweep. An easy way to get the data into the gnumeric spreadsheet, is to type it to the screen using the UNIX command15 cat: cat rich.dat You can then copy and paste16 this data into gnumeric. Our aim to put together data with the same filament current, i.e., to combine the first line of data with the last; to combine the second line of data with the next-to-last; etc. This is easily accomplished by typing the data bottom-to-top to the screen using the UNIX command17 tac: tac rich.dat The results can be copy and pasted next to the previous data so that the data we aim to combine is on the same line. The best estimate for T is the average of the four T s (the two Ti should be nearly identical); the best estimate for IA is the average of the two IA s; for IA use half the difference between the two values of IA (|IA1 - IA2 |/2); for T use half the difference between Tr and Ti (either pair or average the two differences). The result of this `data reduction' is half as many data points, but with a value for IA based on hysteresis. (The alternative for IA would be the meter manufacturer's specifications, and we know they are way to small from analysis of Child's Law.) Report the relative importance of hysteresis and calibration in temperature uncertainty determination. For If = 1.2 A record the difference in Tr due to hysteresis and the difference between Tr and Ti which is a temperature calibration uncertainty. Copy and paste this reduced data into a new file, and fit it to Eq. 3.79. Produce a plot similar to Fig. 3.10; include lines both for your best fit and the `book' values of k1 and k2 . A glace at Eq. 3.69 on page 56 shows that k1 = A (Richardson's Constant) and kk2 = W (Work Function). This processing could very easily have been done within the program itself. I've instead opted to make the program as simple as possible at a cost of additional `by-hand' processing in a spreadsheet. 15 from concatenate--commonly this command is used to combine several files 16 Note use of "See two two separators as one": Alt-e 17 clever or what? 14 Thermionic Emission 65 The work function is an atomic quantity, and it is usually expressed in the atomic scale unit eV18 . Calculate the work function from your value of k2 in joules and eV and compare to the book19 value of 4.5 eV. The book value for A is 0.72 106 A/m2 K2 ; This will serve as a fair initial guess for k1 (required for fitting), but we need a better way to calculate A (and particularly the error in A) based on your data20 . Since our temperatures are so uncertain, particularly at the low end, the best estimate for the Richardson constant A comes from substituting the book value of the work function and the plateau (T, I) measurement for the highest valid filament temperature into Eq. 3.69. We can then estimate the systematic uncertainty in A by using the `high-low' method of 191 (A+ I + , T - , A- ). Report Checklist 1. Write an introductory paragraph describing the basic physics behind this experiment. For example, why did higher cathode currents produce ever higher plateaus of anode current? (This manual has many pages on these topics; your job is condense this into a few sentences and no equations.) 2. Calculations (no errors) of room temperature RW . Measurements (4-wire ohmmeter and direct voltage/current) of Rmeasured at room temperature. Calculation of Rsupport . 3. Observations of the light intensity and color as a function of filament temperature. 4. Data files and computer program: Leave them in your UNIX account. Print out a copy of your program, the file filament.dat, and the data you used to fit Richardson's Law; Tape them into your lab notebook. 5. Plots similar to Figures 3.9, 3.10 (with fit curve Eq. 3.68 and also the Richardson function using "book" values for k1 and k2 ), 3.11 (with fit curve Eq. 3.67), 3.12 (with line Eq. 3.66) and 3.13 (two separate WAPP+ plots one with Tr data and the other with Ti data). Note that Figure 3.9 is complex to produce. Use a file of Nplot commands and feel free to talk to me about how to go about doing this. Carefully note the use of log and inverse scales (which requires positive data--edit to achieve this!). Also include a fit report for each fit curve. 6. Experimental values (with error range) for: W (in eV), A, e/m and T . 7. Show the missing steps leading to Equations (3.31) and (3.36). Substitute the 1 approximation (Eq. 3.35) into the differential equation Equation (3.31). Show that while we do not have an exact solution to the differential equation, the singular parts (i.e., those that approach infinity as 1) cancel. 8. Make a final results table, recording your final results (with proper units and sigfigs) adjacent to the corresponding `book' values. Don't forget to record your tube's identifying letter! Recall: 1 eV = 1.6022 10-19 J is the energy an electron gains in going through a potential difference of 1 V. 19 Blakemore, Solid State Physics 20 Recall that when systematic errors dominate computer-based errors are usually some sort of non-sense. 18 66 Thermionic Emission Comment: Classical vs. Quantum I said above that the presence of an in Richardson's constant shows that the process is governed by quantum mechanics. It is not quite so simple. The evaporation of water is not fundamentally different from that of electrons, while the former (due to the larger mass of a water molecule) is well-approximated by a classical calculation. Classical evaporation can be calculated much as quantum evaporation using the Maxwell-Boltzmann speed distribution and the number density (of water molecules) rather than the disguised version of this: Fermi 3 2 energy (EF number density). We can write the classical rate in terms of the quantum with no visible: 3 EF 2 4 quantum flux (3.80) classical flux = 3 kT The different temperature dependence21 for the classical flux (T 2 e-W/kT vs. T 2 e-W/kT ) cannot be detected experimentally: the Boltzmann factor overwhelms all other temperature dependencies. On the other hand, since EF kT , the expected classical rate is much larger than the quantum rate. This played a role in the mistaken idea that thermionic emission was due to surface contamination: the experimental rate seemed too small to be thermal evaporation. On the other hand a more fruitful interpretation of the "low" rate was that only a fraction ( kT /EF ) of the electrons were thermally active. This connected with other observations (like the "small" specific heat of the electron gas) and provided a link to the idea of a degenerate Fermi gas. 1 Comment: Uncertainty Inspection of Figures 3.9-3.12 shows that something "funny" is going on. (I can't say "abnormal" or "unusual" as it is neither.) Figure 3.12 shows the unmistakable signs of "small reduced 2 ": The fitted line goes nearly dead-center through all 30 error bars, never coming even close to an edge. For error bars sized at one standard deviation (), you should expect total misses of the error bar about 1/3 of the time. In addition recall that each data point is really a double: the same filament current was sourced as part of a temperature up-sweep and as part of a temperature down-sweep. These repeated measurement should also frequently miss by a standard deviation, but here they are so close that the two points often look like just one. The answer to this puzzle is that the error bars are not displaying statistical (`random') error. Instead the temperature was measured two different ways (Ti and Tr ), and the error bar represented the deviation between these two measurement methods. When different methods of measurement produce different answers for the same quantity, we have a textbook example of systematic error (in contrast to statistical error). Notice that if we had used the statistical deviation of just one measure of temperature, we would seriously underestimated the error. Furthermore since quite accurately measured electrical quantities were used to calculate the temperature (via Equation 3.62 or Equation 3.56), application of the usual error propagation methods would also have produced very small errors in T . The full range of our uncertainty in temperature is made evident only by measuring T two different ways. (This is typically how systematic errors are detected.) 21 Saul Dushman ( Phys. Rev. 21, 623636 (1923)), while working at G.E., provided a general thermodynamic argument for the T 2 dependence and the universal nature of A. The resulting relationship is therefore sometimes known as the Richardson-Dushman relationship. Thermionic Emission 67 Having detected systematic errors, we should seek an explanation. . . In addition to the problems previously cited (particularly use of book values for filament dimensions and the problems associated with Rsupport ), nonuniform filament temperature due to filament supports may be the problem. Koller (p. 89) reports the filament temperature 0.5 cm from a support is reduced by 15% (and of course the effect is greater closer to the support). Thermionic emission is reduced a similar amount 1.3 cm from a support. Thus the quantity we are seeking (a filament temperature) does not even exist, so it is hardly surprising that different measurements give different results. (It turns out the Tr is something like the average temperature; whereas Ti is something like the central temperature.) These effects have been investigated, and Koller gives the theory to correct such measurements, but such considerations are beyond the scope of this experiment. Figure 3.9 shows the opposite problem, "large reduced 2 ": The fitted line systematically misses almost every error bar. In this case, the miss might be called "small", but the error bar is smaller still. Once again, errors were not calculated statistically (manufacturer's specifications were used), so "reduced 2 = 1" should not really be expected. In this case, my guess is that the problem is with our simplified theory (we assumed: infinite cylinders, no random component to the electron velocity [zero electron temperature], uniform filament [temperature, voltage, emissivity, . . . ], perfect vacuum, no incipient current plateau). We could of course test these hypotheses by further experiments with different tubes, but such work is beyond the scope of this experiment. (Indeed I have detected VA -sweep hysteresis; correction for this dramatically reduces reduced chi2 , but not all the way to 1.) Summary: Very large or very small reduced 2 suggests significant non-statistical errors, a very common --perhaps even the usual-- situation. Computer generated errors are some sort of none sense in this circumstance. Presumably your theory and/or measurement methods are less than perfect. That is, of course, No Surprise. Science next requires you to guess and perhaps further investigate which imperfections are the leading source of the problems, i.e., what changes to the experiment would ameliorate the problem. References 1. Jones & Langmuir GE Review, The Characteristics of Tungsten Filaments as Functions of Temperature 30 (1927) Part I pp. 31019, Part II pp. 35461, Part III pp. 40812 2. Melissinos Experiments in Modern Physics, 1966, pp. 6580 3. Preston & Dietz The Art of Experimental Physics, 1991, pp. 14147, 15261 4. Blakemore Solid State Physics, 1974, pp. 18895 5. Koller The Physics of Electron Tubes, 1937, Ch. I, VI, VIII 6. G.E. FP-400 Description and Rating FP400.pdf found at 68 thickness: x Thermionic Emission E(x) x Area: A E(x+x) x Figure 3.14: Gauss' Law is used to calculate the charge between two plates of area A separated by a distance x. Since (by assumption) the potential just depends on x, the electric field is in the x direction and is given by E = -dV /dx. Appendix--Poisson's Equation Equation 3.5 and Equation 3.28 made reference to "Poisson's Equation", which is really a topic for Physics 341, rather than part of Physics 200. In this appendix, Poisson's Equation is derived starting from two Physics 200 results: Gauss' Law: that the electric flux leaving a region depends just on the charge enclosed by that region: ^ E n dA = Qenclosed /0 and the relationship between electric field and electric potential (voltage): Ex = - dV dx (3.82) (3.81) Poisson's Equation is a differential equation equivalent to Gauss' Law. It is usually written in terms of the Laplacian (2 ), which in turn can most easily be expressed in terms of second derivatives w.r.t. x, y, and z: 2 V = 2V 2V 2V + + = -/0 x2 y 2 z 2 (3.83) where is the electric charge density. We need Poisson's Equation only in cases where the electric potential depends on just one variable (x or cylindrical r), which simplifies the required proofs considerably. As shown in Figure 3.14, if V is a function of x alone: V (x), we can find the charge between two plates of area A using Gauss' Law: Qenclosed = 0 A (E(x + x) - E(x)) 0 A Thus the charge density between the plates is given by: = 0 A dE x Qenclosed dE d2 V dx = = 0 = -0 volume Ax dx dx2 (3.85) dE x dx (3.84) Thermionic Emission 69 thickness: r E(r+r) l E(r) r+r Figure 3.15: Gauss' Law is used to calculate the charge between two coaxial cylinders of length l separated by a distance r. Since (by assumption) the potential just depends on r, the electric field is in the r direction and is given by E = -dV /dr. which provides what is needed for Equation 3.5. As shown in Figure 3.15, if V is a function of r alone: V (r), we can find the charge between two coaxial cylinders using Gauss' Law: Qenclosed = 0 l {2(r + r)E(r + r) - 2rE(r)} = 0 l {2r(E(r + r) - E(r)) + 2rE(r + r)} dE 0 l 2r + 2E(r) r dr (3.86) Thus the charge density between the cylinders is given by: = 0 l {2r dE/dr + 2E(r)} r Qenclosed = = 0 volume 2rlr d2 V 1 dV = - 0 + 2 dr r dr 1 dE + E dr r (3.87) (3.88) which provides what is needed for Equation 3.28. 70 Thermionic Emission 4: Photometry The mathematical thermology created by Fourier may tempt us to hope that. . . we may in time ascertain the mean temperature of the heavenly bodies: but I regard this order of facts as forever excluded from our recognition. Auguste Comte Cours de la Philosophie Positive1 (1835) . . . within a comparatively few years, a new branch of astronomy has arisen which studies the sun, moon, and stars for what they are in themselves and in relation to ourselves. Samuel Pierpont Langley The New Astronomy (1888) Purpose In 1835 the French philosopher Auguste Comte noted that since we know stars only by their light (and cannot take bits of stars into the laboratory), our knowledge of stars would be forever limited essentially to location. Fifty years later astrophysics--the study of the physics of stars--was beginning, and the first measurements of the Sun's temperature and composition were being made. Evidently, careful measurement of starlight (photometry) allows the intrinsic properties of stars to be determined. In this lab you will use broadband photometry to measure the temperature of stars. Introduction We begin our study of stars with blackbody radiation, which you measured in the Thermionic Emission lab (p. 53). There we claimed (and you confirmed) that the surface area of an object emits light energy at the rate (power in watts): P = T T 4 A (4.1) where is the Stefan-Boltzmann constant, T is the temperature of the body, A is the surface area of the body, and T is the total emissivity. Our `spherical cow' model2 of a provides this as an etext translated and edited by Harriet Martineau. 2 This model of a star works best with stars similar in temperature to the Sun, but every star's light is considerably modified by absorption in the star's atmosphere. In hot stars, UV wavelengths that can photoionize hydrogen (H(n = 2) + H+ + e- ) are highly attenuated producing the Balmer jump. The 1 71 72 Photometry star is an incandescent ball (radius R) of gas shining as a blackbody. Thus the total power output of a star (called the luminosity 3 ) is L = T 4 4R2 (4.2) The light produced by a hot object is akin to audio noise: in both cases a random process produces a simultaneous superposition of a wide distribution of wavelengths. In his Opticks, Newton noted certain systematic dependencies in the light emitted by incandescent objects and suggested something fundamental was behind the process4 . Experimentally it was found that the distribution of photon wavelengths was broadly similar to the Maxwell-Boltzmann speed distribution: a bell-shaped curve in which the location of the peak depended on temperature. In 1893 Wilhelm Wien5 concluded that the wavelength of this peak must be inversely proportional to temperature. Experiment confirmed that the wavelength of this peak was given by: 2898 m K (4.3) max = T Together the Stefan-Boltzmann and Wien displacement laws--both the results of classical physics--explain much of what was observed in the Thermionic Emission lab. At room temperature the filament did not appear to be a source of light: it appeared dark. This is a result both of the small rate of emission given by Stefan-Boltzmann and (from Wien) for T = 300 K, we find max 10 m, that is the wavelengths typically emitted are well outside the range of light that can be detected by the eye6 . On the other hand for an object like the Sun: T = 6000 K, so max 0.5 m -- right in the middle of of the visible range. The equation for the exact distribution of photon wavelengths produced by a blackbody was actively researched with erroneous equations produced by Wien and Rayleigh & Jeans. In 1900 Max Planck7 published his derivation of the wavelength distribution of light from a blackbody. However before we can discuss his result we must explain what precisely is meant by `wavelength distribution'. In Physics 211 you were required to make plots (histograms) of speed distribution based on a list of 1000 molecular speeds. You counted the number of molecules (`hits') that had speed in a range (bin) (v - v/2, v + v/2) for various speeds v. (Here the bin is centered on the speed v and has bin-size v.) Clearly the number of hits in a bin depended on the bin-size (or range), but you should expect that the hit density (number of hits divided by the bin-size) should be approximately independent on the binsize. Your histogram was then a plot of hit density vs. bin center. In a similar way we can sort the light into various wavelength bins, and calculate the total light energy per bin-size. For sorting visible light by wavelength, an appropriate bin-size would be a fraction of a m, and the resulting measurement of light intensity (hit density) would have units of J/m. In the case of the thermal radiation continuously emitted from the surface of a body, we are spectra of cool stars shows broad absorption bands due to molecules. is an applet allowing a comparison of blackbody light to actual star light. 3 The MKS unit for luminosity is watt, but the numbers are so large that it is usually expressed as a multiple (or fraction) of the Sun's luminosity (a.k.a., solar luminosity) L = 3.846 1026 W. 4 Query 8: Do not all fix'd Bodies, when heated beyond a certain degree, emit Light and shine. . . Query 11:. . . And are not the Sun and fix'd Stars great Earths vehemently hot. . . 5 Wilhelm Wien (18641928) German physicist, 1911 Nobel Prize in Physics 6 The human eye can detect light with a wavelength in the range 0.4 m (violet) to 0.7 m (red). 7 Max Planck (18581947) German physicist, 1918 Nobel Prize in Physics Photometry 73 Blackbody Thermal Radiation 10.E+07 8.E+07 F (W/m2 m) 6.E+07 4.E+07 2.E+07 0 6000 K 4000 K 0 .5 1.0 1.5 wavelength (m) 2.0 Figure 4.1: The distribution of the light emitted by a blackbody plotted as a function of wavelength. Hotter objects emit much more light than cool ones, particularly at the shorter wavelengths. interested in the rate of energy emission per area of the body, with units (W/m2 )/m. This quantity is called the monochromatic flux density and denoted F 8 . Planck's result was: F = 1 2hc2 5 exp(hc/kT ) - 1 (4.4) where h is Planck's constant, whose presence is a sign of the importance of quantum mechanical effects. The primary assumption in Planck's derivation of this distribution was the photon: an indivisible packet of light carrying total energy E = hc/. With this identification notice the Boltzmann factor in Planck's equation: exp(E/kT ). As before, this factor means that high energy (i.e., short wavelength) photons are rarely in the mix of emitted photons. Thus only "high" temperature stars will emit an abundance of UV radiation. Of course, the light emitted from the surface of a star gets spread over an ever larger area as it moves away from the star. Thus the flux coming into a telescope a distance r from the center of the star would be: F = 2hc2 R2 1 5 r 2 exp(hc/kT ) - 1 (4.5) Theoretically its now easy to measure the temperature of stars: simply measure the wavelength distribution (`spectra') of the starlight and fit the measurements to the above theory with two adjustable parameters: T and R/r. (Do notice that increasing R or decreasing r produces exactly the same effect--an overall increase in the flux density--so we cannot separately determine R and r from the spectra. A large, distant T = 6000 K star could 8 One could just as well measure bin-size in Hz. The result is denoted F with units (W/m2 )/Hz. A helpfully sized unit for F is the jansky: 1 Jy = 10-26 W m-2 Hz-1 . F is commonly used in radio astronomy whereas F is commonly used in optical astronomy. 74 Photometry have exactly the same spectra as a small, close T = 6000 K star.) Let me enumerate a few (of the many) reasons this does not work perfectly. 1. We assumed above that space was transparent: that the only effect of distance was diluting the light over a greater area. However, our Galaxy is filled with patchy clouds of dust. Unfortunately the dust in these clouds has the property of being more likely to absorb blue light than red. Thus a dust-obscured star will lose more blue than red light and the observed spectra will be distorted from that emitted. A dust-obscured star would measure cooler and appear more distant (i.e., dimmer) than it actually was. These two effects of dust are called `reddening' and `extinction'. Of course substantial dust absorption is more likely for more distant stars. It is usually `small' for stars within 300 Ly of Earth. However, 300 Ly covers a very tiny fraction of our Galaxy. In this lab we will side step this problem by selecting stars with little dust absorption. However, with a bit of additional work, dust absorption could be measured and corrected for using photometry. 2. Starlight measured from the surface of the Earth must traverse the Earth's atmosphere. The non-transparency of the Earth's atmosphere will also modify the observed starlight. In addition, the telescope and detector typically introduce additional `nontransparency', that is blue and red photons at the telescope aperture are not equally likely to be counted. The `efficiency' of the system depends on wavelength. Unless extraordinary care is taken, this efficiency can change on a daily basis. (And, of course, the atmosphere's transparency can change in a matter of minutes.) As a result frequent calibration of the system is required. Calibration at the multitude of wavelengths required to make a full spectra is obviously more difficult than calibration at just a few wavelengths. 3. There is a competition between the accurate measurement of the value of F and the bin-size used. As you should have noticed in Physics 211, if you select a small bin-size, each bin captures relatively few counts which results in relatively large N errors. So if you have only a few thousand photons, you may do better to use a big bin-size (so you capture enough counts to get an accurate measurement of F ), but then only have a few bins (widely spaced in wavelength so, unfortunately, the spectrum has been reduced to just a few points). Of course, you could always collect starlight for a longer period of time (longer `integration time') or use a telescope with a larger aperture. However, these solutions miss the point: The boundary between the known and unknown in astronomy is almost always at the edge of what you can just barely detect. Thus you must always come up with the maximally efficient use of the available photons. Telescope time is allocated for a detailed spectra only when it is expected that the results cannot be obtained in a `cheaper' way. 4. Stars are not exact blackbodies, and of course they are not at a temperature. Certainly if we move from the surface to core of a star we would experience quite different temperatures. But even on the `surface' of our Sun we see regions (`sunspots') with unusually low temperature. In the Langmuir Probe lab, you will find that the electrons is a small region may have a different temperature from the co-mingled atoms. Similarly, in the atmospheres of stars it is not unusual for the various components of the gas to be at different temperatures (that is to say the gas is not in local thermodynamic equilibrium [LTE]). Hence it may make no sense to try to obtain high-precision measurements of `the' temperature. Photometry 75 Oddly enough the fact that stars are not blackbodies is actually helpful as it allows a variety of information to be decoded from the starlight. (The only information in a blackbody spectra is the temperature.) Varying absorption in the star's atmosphere then means that some wavelengths of light come from deeper in the star and hence represent a higher temperature. Ultimately the absorption lines in a star's spectra provide the bulk of the information we have about stars. Chemical composition of the star's atmosphere and temperature--accurate to a few percent--are best obtained this way. However, consideration of these absorption lines is beyond the aims of this lab. Temperature I Our aim in this lab is to measure the temperature of stars without resorting to a detailed measurement of the star's spectra. We begin by considering measurement of F at just two wavelengths: F1 at 1 and F2 at 2 where 1 < 2 . (Think of 1 as blue light and 2 as red light.) There is a huge cancellation of factors if we look at the ratio: F2 /F1 : 5 exp(hc/1 kT ) - 1 F2 1 = 5 F1 2 exp(hc/2 kT ) - 1 (4.6) In general, ratios are great things to measure because (A) since they are dimensionless they are more likely to be connected to intrinsic properties and (B) it is often the case that systematic measurement errors will (at least in part) cancel out in a ratio. While the ratio has considerably reduced the complexity of the formula, it would help if we had an even simpler formula. Towards that goal we make the approximation that the exponential terms are much larger than 1: F2 F1 = log(F2 /F1 ) Example Consider the case where 1 = .436 m (blue) and 2 = .545 m (yellow or `visible'). For historical reasons, astronomers prefer to consider 2.5 log 10 (F2 /F1 ), and if we evaluate all of the constants we find: 7166 K - 1.21 (4.10) 2.5 log(F2 /F1 ) T Do notice that for cool stars this quantity is large (more `visible' than blue light) and that extremely hot stars (T ) will all have much the same value for this quantity: -1.21. This last result is a consequence of the `Rayleigh-Jeans' approximation for the Planck distribution, valid for max . In this large wavelength (or high temperature) limit, the Boltzmann factors are nearly one and: exp(hc/kT ) - 1 hc/kT (4.11) 5 1 5 2 5 1 5 2 hc kT exp(hc/1 kT ) - 1 exp(hc/2 kT ) - 1 (4.7) (4.8) (4.9) 5 1 1 hc exp(hc/1 kT ) 1 = 5 exp - exp(hc/2 kT ) kT 1 2 2 1 1 - log e + 5 log(1 /2 ) 1 2 76 Photometry Color Index vs. Temperature 2 Color Index 1 0 1 2000 4000 8000 Temperature (K) 20,000 40,000 Figure 4.2: In the case where 1 = .436 m (blue) and 2 = .545 m (yellow or `visible'), the `color index' 2.5 log(F2 /F1 ) is plotted as a function of temperature. The dotted line shows the approximate result Eq. 4.10; the solid line shows the actual ratio of Planck functions. Do remember that high temperature stars have a smaller, even negative, color index. so9 F 2c 2hc2 kT = kT 4 5 hc (4.12) Since temperature is now an over-all factor, it will cancel out in any flux ratio. Color Index At increasing distance from the star, both F2 and F1 will be diminished, but by exactly the same fraction (assuming space is transparent, or at least not colored). Thus this ratio is an intrinsic property of the star, and you saw above that this ratio is related to the temperature of the star. The ratio is, of course, a measure of the relative amounts of two colors, and as such is called a color index. (Our experience of color is independent of both the brightness of the object and the distance to the object. It instead involves the relative amounts of the primary colors; hence the name for this ratio.) Any two wavelengths 1 , 2 can be used to form such a color index, but clearly one should select wavelengths at which the stars are bright and max (to avoid the Rayleigh-Jeans region where the ratio is independent of temperature). 9 Do notice the absence of h in this classical result. Photometry 77 Magnitude For historical reasons, astronomers typically measure light in magnitudes. Magnitudes are related to fluxes by: m = -2.5 log 10 (F/F0 ) (4.13) where F0 is the standardizing10 flux that corresponds to m = 0. In astronomy log commonly11 refers to log10 , although we could equivalently write this equation in terms of other bases: m = -2.5 log10 (F/F0 ) = -1.085736 ln(F/F0 ) = - log2.511886 (F/F0 ) (4.14) The minus sign in this definition means that magnitudes work backwards: the brighter the star, the smaller the magnitude. Very bright objects would have negative magnitudes. Since the measured flux of a star depends on distance, magnitude is not an intrinsic property of a star. Assuming transparent space, magnitude can be related to luminosity and distance: m = -2.5 log L 4r 2 F0 (4.15) Thus if a star's luminosity is known and its magnitude measured, its distance can be calculated. Astronomers define the absolute magnitude of a star to be the magnitude the star would have if observed from a distance of 10 pc12 . Thus: m = -2.5 log = -2.5 log = -2.5 log L 4r 2 F0 L 4(10 pc)2 F0 10 pc r 2 (4.16) (4.17) (4.18) (4.19) L r + 5 log 2F 4(10 pc) 0 10 pc r = M + 5 log 10 pc where M is the absolute magnitude and the remaining term: 5 log(r/10 pc) is known as the distance modulus. The main point here is that distance affects magnitude as a constant offset. This result is a lemma of the general result: Theorem: If the measured flux Fm is a fraction () of the actual flux F (as would occur due to a less-than-perfectly efficient detector, a dusty mirror, atmospheric absorption, absorption due to interstellar dust,. . . ) the resulting magnitude as measured (mm ) is just a constant (2.5 log ) off from the actual magnitude (m). Proof: mm = -2.5 log(Fm /F0 ) = -2.5 log(F/F0 ) = -2.5 log(F/F0 ) - 2.5 log() (4.20) (4.21) (4.22) (4.23) = m - 2.5 log() 10 11 Historically F0 was defined as the flux of the star Vega, so Vega by definition had magnitude zero. Pun intended. Use of log 10 is a bit old fashion. If you see `log' (without an explicit base) in a modern physics document, you should probably assume ln = loge is intended, unless something historical is being discussed (like stellar magnitudes, dB or pH). 12 pc = parsec = 3.0857 1016 m = 3.2616 Ly 78 Photometry JohnsonCousins UBVRI 1.0 .8 Transmission .6 .4 .2 0 U B V R I .4 .6 .8 wavelength (m) 1.0 1.2 Filter Name U B V R I (m) 0.36 0.44 0.55 0.71 0.97 (m) 0.07 0.10 0.09 0.22 0.24 F0 (W/m2 m) 4.22 10-8 6.40 10-8 3.75 10-8 1.75 10-8 0.84 10-8 Figure 4.3: The characteristics of the standard filters: U (ultraviolet), B (blue), V (visible), R (red), I (infrared) from Allen's Astrophysical Quantities p. 387 and The General Catalogue of Photometric Data: Much of this lab will involve finding the constant offset that relates our measured (instrumental) magnitudes to the actual (standardized) magnitude. Filters In the above equation for magnitude, the flux might be the total (all wavelengths or `bolometric') light flux, or a monochromatic flux density F , or the flux in a particular range of wavelengths. Much of astronomy is concerned with the flux through the five standard13 U , B, V , R, and I filters. The characteristics of these filters are detailed in Figure 4.3, but essentially each filter allows transmission of a range of wavelengths , centered on a particular wavelength. The name of the filter reports the type of light allowed to pass: U (ultraviolet), B (blue), V (`visible', actually green), R (red), and I (infrared). The filters are broad enough that the resulting bin really doesn't well represent a value for F . (Narrowband filters called u, v, b, y, are also commonly used.) The magnitude of a star as measured through a B filter is called the B magnitude and 13 Johnson, H. L. and Morgan, W. W. (1951) ApJ 114 522 & (1953) ApJ 117 313 Cousins, A.W.J. (1976) memRAS 81 25 Photometry 79 Temperature vs BV Temperature vs RI 14000 12000 10000 8000 6000 4000 0 .5 1.0 BV 1.5 2.0 12000 10000 8000 6000 4000 Temperature (K) Temperature (K) 0 .5 1.0 RI 1.5 2.0 Figure 4.4: Using data from Allen's Astrophysics Quantities calibration curves relating stellar temperature to color index can be obtained. The fit equations for these curves is given in the text. is simply denoted: B. Notice that we can now form color indices just by subtracting two magnitudes. For example, the most common color index is B - V : B-V = 2.5 (log(FV /FV 0 ) - log(FB /FB0 )) = 2.5 log(FV /FB ) + constant (4.24) (4.25) (4.26) = 2.5 (log(FV /FB ) + log(FB0 /FV 0 )) So B - V is related to the flux ratio FV /FB and so is an intrinsic property of the star related to temperature. Furthermore, while Eq. 4.10 was derived assuming monochromatic flux densities (in contrast to the broad band fluxes that make up B and V ), we can still hope equations similar to Eq. 4.10 can be derived for B - V . Allen's Astrophysical Quantities provides calibration data to which Kirkman has fit a curve: B-V = and the inverse relationship: T = 1700 + 5060 - 1600(B - V ) (B - V ) + .36 (4.28) 35800 - 3.67 + 1.08 10-4 T T + 3960 (4.27) The data with fitted curve is plotted in Figure 4.4. Notice that for B - V < 0 small uncertainties in B - V result in large uncertainties in T : it would probably we wise to switch to a different color index like U - B to measure the temperature of such a hot star. Similar work can be done for R - I, with results: T = 2750 + 1400 + 160(R - I) (R - I) + .314 (4.29) For the Sun, Allen's Astrophysics Quantities reports: T =5777 K, B - V =0.65, R - I=0.34 whereas Eq. 4.28 gives 5703 K and Eq. 4.29 gives 5659 K. In general errors of a few percent should be expected. 80 Photometry Temperature II The starting point for physics is usually hard intrinsic quantities like temperature, density and pressure. However, in astronomy the transducers14 used to measure these quantities are parts of the star itself (say the absorption lines of a particular element in the atmosphere of the star). We will always be a bit uncertain about the exact situation of these transducers, and hence the calibration of these transducers is correspondingly uncertain. For example stars with unusual chemical composition (like population II stars) or unusual surface gravity g (like giant stars) really deserve separate calibration curves. Since the usual physical quantities are in astronomy provisional and subject to recalibration, astronomers attach primary importance to the quantities that are not subject to the whims of revised theory: the actual measurements of light intensity. Thus astronomers use as much as possible hard quantities15 like B - V directly without converting them to temperature using some provisional formula. For example just using B - V , stars can be arranged in temperatureincreasing order, since there is a monotone relationship between B - V and T . Of course, in this lab the aim is to measure star temperatures in normal units. Summary In this lab you will measure the temperature of stars by measuring their B, V, R, I magnitudes, calculating the color indices B -V and R-I, and then, using the supplied calibration curves, find T . The calibration curves are based on detailed spectra of bright, normal stars; In using these calibration curves we are automatically assuming our (distant and hence dimmer) target stars are also `normal'. Detector System In this lab photons are counted using a charge coupled device or CCD. Our Kodak KAF0401E CCD consists of a typical 24-pin DIP integrated circuit with a window that exposes to light a 7 5 mm field of 765 510 light sensitive regions or pixels. Photons incident on a pixel free electrons via the photoelectric effect16 . Each pixel stores its charge during an exposure, and then each bucket of charge is transferred (in a manner similar to a bucket brigade) to a capacitor where it produces a voltage proportional to the number of electrons. An analog-to-digital converter then converts that voltage to a 16-bit binary number. A In normal usage, a transducer is a device which converts a physical quantity (e.g., pressure, temperature, force,. . . ) to an electrical quantity (volts, amps, Hz . . . ) for convenient measurement: for example, a microphone. Here I've broadened the usual meaning to include natural `devices' whose emitted or absorbed light allows a physical quantity to be determined. Obviously we are connected to the stars by light, not copper wires. 15 In this lab you will see that even as simple a quantity as B is actually only distantly related to direct meter readings. The direct meter readings must be `reduced' to eliminate confounding factors like detector efficiency or atmospheric absorption. Nevertheless, since these problems are `at hand' astronomers believe that they can be properly treated. Astronomers willing archive B data, whereas T calculations are always considered ephemeral. 16 Albert Einstein (18791955) received the 1921 Nobel prize for his 1905 theory explaining this effect. Robert Millikan's [(18681953), B.A. Oberlin College, Ph.D. Columbia] experimental studies (19121915) of the photoelectric effect were cited, in part, for his Nobel in 1923. 14 Photometry Telescope aperture focal ratio focal length Focal Reducer effective f Filter Wheel Camera pixels image scale Meade LX200 12" f /10 120" Meade f/3.3 FR/FF 50" SBIG CFW-8A SBIG ST-7E ABG 765 510 18 12 1.45"/pixel CCD pixel full-well read noise gain .65 m .55 m .45 m .40 m Kodak KAF-0401E 9 m 9 m 50, 000 e- /pixel 15 e- rms 2.3 e- /ADU Quantum Efficiency 60% 50% 40% 30% 81 Table 4.1: CCD photometry equipment used at the SJU observatory. Note: A Celestron f/6.3 FR/FF (which yields 0.97"/pixel) may be employed when unusually good seeing is anticipated. CCD image is thus a 765 510 matrix of integers. The units of these integers is simply ADU for analog to digital units, but ideally the result should be proportional to the number of incident photons. The combined effects of photon e- efficiency (`quantum efficiency') and e- ADU (`gain') means that there are 5 times as many photons as ADU counts. Electron storage during an exposure is limited to a full-well capacity of 50, 000 e- which corresponds to a count of 20, 000 ADU. The Kodak CCD is contained within (and controlled by) a ST-7E camera made by the Santa Barbara Instrument Group17 . In front of the camera, an SBIG CFW-8A filter wheel allows computer controlled placement of one of the U, B, V, R, I filters. The SBIG detection system is mounted on a Meade 12" LX200 Schmidt-Cassegrain Catadioptric telescope. Once the telescope has been initialized, Meade's telescope control software allows the telescope to be directed to a long list of cataloged objects or to any location specified by RA/dec18 . Linearity In an ideal CCD, the count in a particular pixel should be proportional to the number of photons absorbed, which in turn should be proportional to the exposure time. This proportionality fails for large exposures and the CCD saturates as the count approaches 20,000 ADU. (This is due to the limited capacity of a pixel to store electrons: the full-well capacity.) Of course, if a pixel is saturated, additional photons are missed, and the ADU count falls below actual photon flux. CCD linearity was tested by varying the exposure time. The results are displayed in Figure 4.5. The calculated magnitude of a star--a measurement of the photon flux (counts/time)-- should not depend on the exposure time. Of course, for very short exposures the handful of counts above background is subject to a relatively large random fluctuation, and so the magnitude will have a relatively large uncertainty. As we've just learned, sufficiently long 17 18 See page 88 for a list of online tutorials covering basic astronomy vocabulary. 82 Photometry Linearity Test 4000 3.E+04 3000 2.E+04 ADU ADU 2000 Linearity Test 1.E+04 1000 0 0 5 10 Exposure Time (s) 15 0 0 .5 1.0 Exposure Time (s) 1.5 Figure 4.5: Below 30, 000 ADU (12 second exposure) the response of our CCD seems to be linear, with saturation quite evident in the 16 second exposure. In fact, the CCD should not be trusted to be linear above 20, 000 ADU. The linear fit may look good, but it is in fact not great: reduced 2 =8. Systematic error in the shutter speed control is probably a problem for the short exposures. (Elimination of the two shortest exposures results in a reduced 2 =1.6.) Also note that the y intercept is not exactly zero: with this camera zero-time exposures are designed to produce an output of about 100 ADU. exposures will result in pixel saturation and hence a systematically low count and a too-large (dim) magnitude. As shown in Figure 4.6, there is a large range of exposures (including those with slight saturation) which can produce an accurate magnitude. Flat Frame Figure 4.7 shows that individual pixels have different sensitivities to light. While the effect is not large ( 10%), it would be a leading source of uncertainty if not corrected. The solution is (in theory) simple: take a picture of a uniformly illuminated surface, and record the count in each pixel. Evidently this `flat frame' count records how each pixel responds to one particular flux level. If in another frame a pixel records some fraction of its flatframe count, linearity guarantees that the measured flux level is that same fraction of the flat-frame flux level. Thus to correct for varying sensitivities we just divide (pixel by pixel) the raw CCD frame by the flat frame. In practice it is difficult to obtain a uniformly illuminated surface. In a sky flat, one assumes that the small section of a blue sky covered in a CCD frame is uniform. (Typically this is done after sunset so the sky is dim enough to avoid saturation in the short exposure, but not so dim that stars show through.) In a dome flat we attempt to uniformly illuminate a white surface with electric lights. Whatever the source of the flat, it is best to average several of them to average out non-uniformities in the source. It should be clear that a flat frame records the net efficiency of the system: both the varying quantum efficiency of individual pixels and the ability of the telescope to gather light and concentrate it on a pixel. Generally the optical efficiency of a telescope decreases for sources far from the optical axis. The off-axis limits (where the optical efficiency approaches zero) Photometry 83 Saturation 39 magnitude 40 41 42 10 Exposure Time (s) 100 Figure 4.6: The magnitude of four stars was calculated from frames with various exposures. The brightest two stars have saturated pixels in 20 sec exposures and the dimming magnitude is evident in exposures longer than 30 sec. The third brightest star begins to be saturated at 240 sec. The fourth star is not saturated. Notice that a handful of saturated pixels will generally not result in a large magnitude error. Linear Response: U 8000 10000 8000 6000 4000 2000 2000 0 Linear Response: I 6000 ADU ADU 0 100 200 300 Exposure Time (Sec) 400 500 4000 0 0 2 4 6 Exposure Time (Sec) 8 Figure 4.7: Below saturation, the response of our CCD seems to be linear. But different pixels have slightly different sensitivities to light (fairly extreme cases are plotted). Since common light sources produce few UV photons per second, much longer exposures were required to obtain 10,000 counts through the U filter (left) compared to the I filter (right). 84 Photometry define a telescope's maximum field-of-view. The limited optical efficiency near the edge of the field of view leads to `vignetting': the gradual fade of optical efficiency to zero at the extreme edge. In addition optical flaws, like dust on optical surfaces, is recorded in a flat frame. As a result a flat frame depends on the entire detection system (telescope, filter, chip) and will gradually change, as for example, dust accumulates on the mirror. Bias Frame A careful examination of the linearity plots in Figure 4.5, shows that they do not go through the origin. That is, a zero-time19 exposure produces some counts, simply by the process of reading the CCD. The uncertainty in this bias frame (the `read noise') represents the minimum background in any frame. Flaws in the CCD are also most evident in the bias frame20 . Measurements on our CCD shows that the difference between two bias frames shows an approximately normal distribution (aside from the flaws) with a standard deviation 6 ADU. Also note that the system artificially introduces an offset of 100 into each pixel, so 2 of the pixels in a bias frame are expected in the range 100 6 ADU. 3 Dark Frame Even in the total absence of light, pixels will slowly accumulate charge. The electric current leaking into the pixels can be controlled by temperature, hence the name `thermal current' and it produces accumulated charge in the absence of starlight, hence the name `dark current'. The professional solution is to cool the CCD using liquid nitrogen, which reduces the dark currents to near zero. The ST-7 uses a less cumbersome (but less effective) solution: a thermoelectric cooler which can cool the CCD 30 C below ambient temperature, reducing the dark currents by about a factor of 10. The resulting dark currents are not negligible for exposures longer than a few minutes. A `dark frame' is simply an `exposure' with the shutter closed; it is a bias frame plus accumulated charge due to dark currents. A dark frame can be subtracted from an (equally timed) exposed frame to compensate for this non-photon-induced current, however the frame-to-frame variation in the dark current (`noise') results in a corresponding uncertainty in the adjusted pixel counts. There is tremendous pixel-to-pixel variation in the dark current. A few percent of the pixels will have dark currents 10 times the typical value. Thus a dark frame looks black except for a handful of bright (`hot') pixels; it could be mistaken for the night sky, except the `stars' are not bright disks rather isolated, single hot pixels. Since taking a dark frame takes as much time as a real exposure, use dark frames only when required: when you are taking real data. Simply learn to disregard the sprinkling of isolated hot pixels. Data Frames The term `object frame' here refers to a raw CCD frame, with subtracted dark frame. A `reduced frame' is an object frame that has been flat-field corrected. By adjusting for non-photon-induced charge and varying pixel sensitivity, we hope our reduced frame shows 19 20 With the ST-7 the minimum exposure is actually 0.12 s In fact our CCD has a flaw at pixel (592, 194) which in fact affects the entire 592 column Photometry 85 Distribution of Pixel Dark Current .999 .99 .9 Percentile ADU 10 C +10 C 2000 4000 Along a Row: Dark Current .5 800 400 200 .1 .01 .001 200 400 800 ADU 2000 4000 8000 0 10 20 pixel 30 40 (a) The distribution of pixel dark current at +10 C and -10 C. All but a few of the pixels have a (log)normal distribution of dark current. A few percent of the pixels show much larger dark current: these are the `hot' pixels. Note that both classes of pixels respond to temperature: lowering the temperature reduces every pixel's dark current. (b) Part of a row of pixels is displayed at three temperatures: = +10 C, = 0 C, = -10 C. Isolated hot pixels are randomly sprinkled throughout the image (here at columns 12 and 20). The data here and in (a) are from 15 minute dark frames. Temperature `cools' Hot Pixels 4000 150 +10 C 140 130 120 0 C 1000 10 C 0 5 Time (min) 10 15 100 0 110 Normal Pixels at 10 C 3000 ADU 2000 ADU 5 Time (min) 10 15 (c) Dark current produces ever larger stored charge in a pixel; of course reducing the temperature reduces the dark current. (d) While it is more evident in this low count data, there is always deviation in counts. While we can subtract the average dark charge, deviations from the average produce `noise' in our images. Reducing the temperature reduces this noise. Figure 4.8: Dark frames are `exposures' with the shutter closed. The source of these counts is not starlight, rather a thermally induced current allows charge to accumulate in each pixel. This `dark current' can be exceptionally large in `hot pixels'. 86 Photometry the actual distribution of photons. It should be noted that all of these corrections will be applied by making the proper requests of the CCD control software. In the end, you should retain both the fully processed reduced frame and the object frame. Basic Plan Given a reduced frame, determining the number of counts/second from a star is a relatively straightforward21 task which will be performed (with your help) by the software. This resulting rate must be converted to flux by multiplying (or dividing) by a variety of factors: collecting area of the telescope, reflectivity of the mirror, quantum efficiency of the CCD, CCD gain, . . . . However the main point is that there is a proportionality between this rate and the actual flux. By our theorem, this means that a magnitude calculated just using this rate (an `instrumental magnitude') is just a constant off from the actual standardized magnitude. If we have a star in the reduced frame of known magnitude, that constant difference can be calculated just by taking the difference between the instrumental magnitude and the known magnitude. And then that calibrating constant can be applied to every star in the reduced frame, converting each instrumental magnitude to a standardized magnitude. In practice it is wise to use several calibrating stars and use the average value of magnitude difference as the calibration constant. (The standard deviation of the magnitude differences then gives us an estimate for the error.) The calculation of instrumental magnitudes (i.e., calculating -2.5 log(rate)) is also a straightforward task best left to software. All you will need to do is calculate (probably using a spreadsheet) the magnitude differences of the calibrating stars, and then find their mean and standard deviation. Complications Air Mass Corrections Magnitudes are intended to be archivable numbers in which the particulars of the observation (e.g., telescope aperture, quantum efficiency, transparency of the atmosphere,. . . ) have been removed. As describe above, this is relatively easy if the reduced frame contains calibrated sources. On the other hand, if we have separate reduced frames for the object and calibrated stars, we must arrange that the entire detection system (atmosphere included) acts the same (i.e., the same proportionality constant between rate and flux) for both frames. However, it is difficult to find calibrated stars that have the same angular altitude as the object; Thus one frame must be shot through a thicker slice of atmospheric than the other. Air mass corrections adjust for this path length difference and assume that the atmospheres (and optics) were otherwise identical. The absorption of light is a multiplicative phenomenon; that is, if the first meter of the atmosphere allows fractional transmission of light T1 and the next T2 , etc, then the intensity Many important details are hereby hidden from view. A short list: (1) determining the background level to be subtracted from the star pixels (2) determining the proper aperture that encloses all the stars light or accounting for the star's light that has blended into the background (3) accounting for the star's light that is between pixels (4) dealing with pixels that are only fractionally within the aperture. In summary: background and point spread function (PSF) are the key issues. 21 Photometry 87 zenth # atmosphere L z L sec z Figure 4.9: Stars that are near zenith go through the minimum amount of the Earth's atmosphere. Stars at a zenith angle z go through sec z times that minimum. This figure displays the atmosphere as a finite thickness slab, whereas in fact the atmosphere gradually thins out to nothing. Nevertheless, meter-by-meter a star at angle z must go through sec z times as much atmosphere as a star at zenith. `Air mass' is defined as this multiplicative factor: sec z. of light through a sequence of meters is: N F = TN TN -1 T2 T1 F0 = Ti F0 i=1 (4.30) Since it is easier to deal with additive quantities rather than multiplicative quantities we define the optical depth of a filter by: T e- so F = TN TN -1 T2 T1 F0 = exp - N (4.31) i i=1 F0 (4.32) According the Beer-Lambert law, the absorption of light depends on atomic absorptivity (the `cross section' , which generally depends on the wavelength of light), the number of atoms per volume (`number density' n), and the thickness of the filter : = n (4.33) We can think of the atmosphere as a sequence of filters, so the light intensity through the atmosphere is: N F = exp - n i=1 F0 = exp - n d F0 (4.34) Of course the density of the atmosphere (and to some extent the composition of the atmosphere) depends on altitude. If we consider light that has taken the shortest path through the atmosphere (i.e., from a star directly overhead or at zenith), we call the above integral 0 (). (The dependence of 0 on the wavelength of light is of course due to the dependence of on the wavelength of light.) As shown in Figure 4.9, if a star is at zenith angle z, the path length (and hence ) in increased by the air mass factor sec z. Thus: F = exp (-0 sec z) F0 (4.35) 88 Photometry 31.00 31.05 vV 31.10 31.15 31.20 1.8 2.0 2.2 2.4 Air Mass 2.6 2.8 Figure 4.10: The instrumental magnitude minus the standardized magnitude (v - V ) for a red star ( ) and a blue star ( ) as a function of the air mass. Notice that the slopes are quite similar (i.e., .177 sec z), but the intercepts are slightly different ( = 0.016(B - V )), due to the difference in color. This color correction is explained in the following section. 1 The rms deviation in these fits is 0.005 magnitudes which is about 3 the intercept shift. The magnitude that corresponds to this flux is: m = 2.50 log e sec z + m0 (4.36) where m is the magnitude measured at zenith angle z and m0 is the magnitude as measured without any atmospheric attenuation (i.e., the standardized magnitude). For any single star one can (assuming the atmosphere doesn't change over the few hours required) plot (sec z, m) data as the star rises, culminates and sets, and determine the y-intercept which is m0 . Alternatively for a set of stars each with known m0 and measured m at varying sec z, one can determine the correction factors A and B required to convert a star's (sec z, m) to m0 : m - m0 = B sec z + A (4.37) Thus lacking in-frame calibrated stars, several calibration frames (at different sec z) are required for each filter, and the reduction process becomes much more complicated. Note that the zenith angle z can be calculated22 from: cos z = sin sin + cos cos h cos (4.38) where and h are the star's declination and hour angle and is the observatory's latitude. The above terms from astrometry (declination, hour angle, altitude, zenith, right ascension, . . . ) are part of the everyday vocabulary of astronomers. If you are not familiar with these terms you should read the following online tutorials: 22 xephem can do this calculation for you Photometry 89 60 Air Mass 1.1 1.2 1.5 2.0 2.5 3.0 Z 40 Declination 20 0 20 8 6 4 2 0 2 Hour Angle 4 6 8 Figure 4.11: Air mass (sec z) depends on where a star is in the sky. Air mass is plotted here as a function of the star's declination and hour angle for the SJU observatory location. Z marks zenith, where the minimum (1) air mass is found. For example at SJU ( = 45 34.5 ) for stars on the celestial equator ( = 0), the minimum air mass (at h = 0h ) is sec z = 1.43. The sequence of air masses: sec z=1.5, 2.0, 2.5, 3.0 occurs at hour angles h = 1.18h , 2.96h , 3.68h , 4.10h . Figure 4.11 plots these air mass values for other declinations. Color Corrections We learned in the above section that the Earth's atmosphere acts as a non-negligible filter, attenuating starlight before it reaches our telescope. The fact that the atmosphere looks blue tells us that the atmosphere in fact acts as a colored filter, attenuating some wavelengths more than others. (That is, depends on , with short- (blue) light more attenuated than long- (red).) As discussed in Figure 4.12, this means that two stars which, outside our atmosphere have the same magnitude but different temperatures, will be measured as having different magnitudes from the ground. The upshot of this is the atmospheric optical depth parameter 0 , depends (slightly) on the temperature, T , of the star: B = 2.5 0 (T ) log e sec z + B0 (4.39) 90 Photometry hot star F F cool star B B Figure 4.12: The atmosphere (in fact any optical component) acts as a colored filter on incident starlight. Typically short wavelength light is more strongly attenuated than long wavelength light. Consider two stars that have the same B magnitude outside the atmosphere. Of course, the hot star has a larger fraction of its light in the short wavelength side of the B band. As a result more of its light will be affected and its measured magnitude will deviate from that of the cool star. In a tremendous exaggeration I've assumed above 50% absorption at the short- end of the B band, and 0% absorption at the long- end. As a result the total measured B-band flux (the shaded area) will be larger for the cool star even though both stars had the same flux outside the atmosphere. If narrow band filters are used there will be much less change in absorption over the band and so this effect is much reduced. where B is the magnitude through the B filter at ground level and B0 is the atmosphere-free B magnitude. The temperature of the star is related to the actual B0 -V0 color index, which in turn is related to the instrumental color index b - v. If we make a first order Taylor expansion of 0 (T ) in terms of b - v, we have (after lumping together all the unknown constants) b B + B (b - v) + B0 (4.40) That is the offset between the instrumental magnitude and the atmosphere-free magnitude is not exactly a constant; instead we expect a slight dependence on the color of the star. A linear fit (using the data from the calibrated sources) can then be used to determine and , viz: b - B0 = B + B (b - v) (4.41) and similarly for the V magnitudes: v - V0 = V + V (b - v) (4.42) Subtracting these two equations gives us the relationship between the instrumental color index and the atmosphere-free color index: (b - v) - (B0 - V0 ) = B - V + (B - V )(b - v) (4.43) (4.44) [1 + (V - B )] (b - v) + (V - B ) = B0 - V0 Thus we expect a linear relationship (with a nearly unit slope) between the instrumental color index and the atmosphere-free color index. The parameters of this linear relationship can be determined by a fit using the data from the calibrated stars. Photometry 91 dim star bright star ADU FWHM FWHM pixel Figure 4.13: Every photon detected in a CCD frame has traversed (essentially) the same patch of atmosphere, so each star's image suffers the same distortion. Every stellar `image' (really point spread function or PSF) has the same shape, only the total flux in each image varies (with the star's magnitude). However, a brief inspection of an image will show clearly larger spots for brighter stars. As shown above this is simply a matter of the level chosen as bright. Aperture Corrections When you look at your CCD frames, you will not believe my statement that all the star images (really PSF) have the same size. However, both bright stars and dim stars are far enough away that their angular diameters are effectively zero (less than a thousandth of a single pixel). The photons from each star shoot through the Earth's atmosphere and are deflected this way and that producing a multi-pixel disk in the frame. (This is called atmospheric seeing.) But since every photon is shot (basically) through the same patch of atmosphere, these random deflections must have the same distribution. The difference between a dim star and a bright star is simply the total number of photons in the images. (That is a long time exposure of a dim star should look exactly like a short exposure of a bright star if the two cases produce the same number of photons.) To see that the distributions are identical you should look at the full width at half maximum (FWHM) of the PSF: First determine the peak count in an image, and then find the circle at which the count is half of this maximum. The diameter of the circle (typically a few arcsec) is the FWHM. Because of the obvious difference in the image size, when totaling the counts in a stellar image one is tempted to expand or reduce the range of pixels included as part of a star to match the apparent size of that star. Doing this means that an accurate count will be obtained for bright stars, whereas counts (really fractions of counts) will be missed in dim stars. While this doesn't sound like a particularly horrendous idea, recall that our theorem states that if we capture a consistent fraction of the counts in every star, our instrumental magnitudes will just be a constant off from the standardized magnitudes. So a consistent 92 Photometry fraction of the counts may be combined with all the other proportionality constants in the final difference between instrumental magnitude and standardized magnitude. Now there may be occasions where we must adjust the aperture used for magnitude calculation (for example to exclude a neighboring star), however then inconsistencies (and hence errors) are then being generated. The disadvantage of using a constant aperture is relatively more noise-prone `background' pixels will be included in dim stars. There are a variety of solutions to this problems (generally under the heading of PSF fitting), however they are beyond the aims of this lab. The blurring of star images (i.e., PSF) depends both on atmospheric seeing and the adjustment of the telescope. Clearly an out-of-focus telescope (see below) will produce larger star images than a properly focused telescope. It should also be noted that the telescope's `optical aberrations' result in additional blurring near the edge of the telescope's field of view. Focus The aim of focus is to achieve that smallest possible FWHM, so each star's light is concentrated in the fewest possible number of pixels. Since every pixel comes with a certain background `noise', fewer pixels means less noise in the stellar magnitude. You should plan on spending a good bit of time (maybe an hour) trying to achieve the best possible focus. There are two approaches to monitoring improved focus: (1) monitor the peak count in a bright star ( 4 mag), (2) watch for the appearance of numerous dim stars as their peak gets above the `bright' level. (You must learn to ignore the apparent size of bright stars: it is a poor measure of good focus.) A complicating factor is that both (1) and (2) vary independently of focus adjustments since atmospheric seeing varies from second to second, and since photon counts should be expected to vary N . Software xephem xephem is planetarium software designed to display various types of night-sky maps in advance of observing. You will also use it to tabulate astrometry data relevant to your observations. The file using.xephem.txt describes how to use this program to produce the tables and maps required for the pre-observation phase of the lab. In addition you can use it for astrometry and photometry on .fits images (although I recommend gaia below for these reductions). Aladin Aladin is a front-end to various astronomy databases on the internet. You will use it to find images to use as finder maps and to identify (and obtain data on ) the stars in those images. The Aladin server and SkyView+ are often a good sources of images; Simbad is the usual source for stellar information. Sometimes just information from the SAO catalog Photometry 93 is required; then VizieR can be used to access just the SAO catalog. You will use Aladin mostly in the pre-observation phase. The file using.Aladin.txt briefly describes how to use this program to produce the maps and data required for the pre-observation phase of the lab. Note: NASA has a very handy web page23 that also allows you to do many of these things. gaia (see below) is also and excellent way to make finder maps. CCDops CCDOps is software used to control the camera and create/save CCD frames. You will be using a small fraction of its capabilities. Typical commands will be: to set the temperature (cooling) of the CCD, focus, grab images (set the exposure time and perhaps take a dark frame), examine images (histogram, crosshair), save images (in both .ST7 and .fits formats), and reduce images by flat-fielding or averaging. ccdview ccdview will be used as a first-look program. Immediately following your observing night you will want to examine the image files (stored on /media/disk). (At the same time you should put copies of the important frames in your UNIX directory.) While you can use this program to determine magnitudes, gaia (see following) will do a better job. The file using.ccdview.txt briefly describes how to use this program. gaia gaia is the recommended reduced-frame analysis program. In order to use gaia you will need to convert your .ST7 frame into a ndf frame (.sdf) using the program sbig2ndf. The file using.gaia.txt briefly describes how to use this program. IRAF IRAF is the Image Reduction and Analysis Facility, a general purpose software system for the reduction and analysis of astronomical data. It is the standard for professional astronomers: designed for the expert user--with no compromises for the beginner. While its use is encouraged for those thinking of becoming astronomers, I cannot recommend it for this lab. 23 start with 10001000 pixel Digitized Sky Survey images with 0.5 Image Size, B-W Linear Inverse Color Table, and a Grid 94 Photometry Planning to Observe Minnesota weather is is unpredictable24 and we need an extraordinary, no-visible-clouds night. We must be prepared to use any/every of these rare nights on short notice. Of course, you have other commitments (your concert, the night before your math/physics exam, the St. Thomas game. . . ), the aim here is to communicate those constraints to me and to your lab partners in well in advance. Thus you will need to carefully record your schedule a lunar month ahead. Note that Murphy's Law guarantees that if you forget to X-out some important night (your girlfriend's birthday, the day before your term paper is due, a `free' day), that night will be clear, and somebody (I hope not me!) is going to be disappointed. My family commitments and intro astro labs are going X-out all Sunday nights and many Monday, Tuesday and Thursday nights. Before each full moon, mark on the class calendar the dates you and your lab partner(s) are committed to observe. Since clear nights are precious, `go' days need a well-thought-out plan. If you're following the standard project, your plan involves a target whose frame will include several stars with known25 B, V, R, I magnitudes. Thus you need to: (A) Find a target that has the required stars. (Stars with recorded R magnitudes are relatively rare.) (B) Know where those standard stars are in your target (so you can aim so your CCD frames to include the stars you need). (C) Know where your target is among the neighboring stars (so if your first try at finding the target fails, you have a plan [not `hunt-and-peck'] for getting to the right spot). (D) And of course, know that your target is well above the horizon when you are planning to observe. While the standard project does not require additional standard stars, you may want you to record alternative (out-of-frame) standards. Thus, you must produce (and turn in) the pre-observation documents described below. Using the following new moon as a date, use xephem, to select your targets. The problem is to find a star cluster that has the required calibrated stars. For example, in early September xephem makes M39 or NGC 6871 obvious choices. However, I'm unable to find any calibrating R magnitudes in those clusters. (Of course, correcting for air mass and using calibrated sources outside of the cluster is an option.) Starting with the options listed in Table 4.2, clusters between NGC 6940 (air mass 1.2) and M5 (air mass 2.0) are OK; NGC 6940 or NGC 6738 seems to have the smallest air mass. shows lots of R and I magnitude stars well within the capabilities of our system (V < 13), so either looks like a fine choice. Pre-observation Checklist The following should be prepared before each full moon (15-Sep-08, 14-Oct-08): 1. Sign up for `go' days during the following lunar month. 24 has a link designed to report (guess) sky conditions at the SJU observatory. It cannot be relied on, but I believe it's the best available information. Note: black is good for this clear sky clock. 25 Professionally, only very carefully measured `standard' stars--for example those measured by Landolt or Stetson--would be used for calibration. For the purposes of this lab you may use any star that has a `book' magnitude. It is not at all unusual for such literature magnitudes to have 10% errors Photometry 2. Using xephem set for the date of the following new moon print out the following: 95 (a) A Sky View showing the entire sky with the labeled location of the target. It is helpful to label the brightest stars with the Meade * number from the Meade250.edb file. (b) A .xephem/datatbl.txt file recording basic data (RA, dec, air mass, . . . ) for the target for several times through the night. (c) A copy&pasted list of basic data on the SAO stars used to locate the target and selected standard stars (`LX200 stars')--perhaps copy&pasted into the above file. Find and record nearby mag 4 stars to help achieve sharp focus. The file using.xephem.txt describes what data you need to have recorded. 3. Using Aladin (or web-based print out the following finding charts for your target cluster: (a) a low magnification (degree scale) image showing the relationship between the LX200 star and the desired CCD frame. (b) an image scaled to about 2 the CCD frame showing exactly the desired CCD frame. (Since the target CCD frame must include calibrated stars, clearly you must know where the calibrated stars are in the target.) Record the RA and dec for your target frame(s). Learning to effectively use these pre-observation programs will be a bit of a challenge and take several hours, so do not procrastinate and feel free to seek my help. In addition to the usual problems of learning new software, you will need to learn how to make the web-based astronomy databases find the information you need. Example 1: Plan for Lunar Month starting 21-July-2005 Immediately on starting xephem, Set the location: SJU Observatory, click on the Calendar for the following new moon ("NM", in this case 4-August-2005) and enter a Local Time of 21:00 (9 p.m.). Hit Update so xephem is ready to display the sky at that time. ViewSky View then displays the night sky for that location/date/time. (Print a copy of this sky view as described in using.xephem.txt and record: UTC Time: 2:00, Sidereal: 16:37.) IC 4665 is the obvious target; its RA (listed in Table 4.2) is most nearly equal to the sidereal time and hence it is near the meridian, However, with a diameter of 70 , IC 4665 is much larger than the CCD frame, so only a small fraction of it will be imaged--a fraction that must include several calibrated stars. A visit to the IC 4665 cluster web page26 finds a couple of dozen of VRIc observations. I copy and paste that data into a spreadsheet, and then seek corresponding UBV CCD data. The result is 12 stars with the full set of BV RI data. Next we need to find a CCD frame that will include as many of these stars as possible. (It is of course possible to use a different set of stars to calibrate different filters, but it will be easier if one set of stars will serve for all calibrations.) Surprisingly, most of the 12 stars have SAO27 identifications; only three stars must be located by RA and dec. Aladin can find a low magnification (1.5 1.5 ) image of IC 4665. (Print this image as a low magnification finder chart.) A request to VizieR for the SAO catalog objects near IC 4665 allows me 26 27 Smithsonian Astrophysical Observatory--a standard catalog of `bright' stars. 96 Photometry to locate the calibrated sources found above. Since IC 4665 is much larger than our CCD frame, we aim for the largest possible subset of these stars. A CCD frame centered near SAO 122742 will pick up four fully calibrated stars (73, 82, 83, 89). Simbad gives B -V data for several other bright stars in this frame (e.g., TYC 424-75-1, Cl* IC4665 P44), but I don't find additional R - I data in Simbad. Returning to, I find (depending on the exact position of the CCD frame) it may be possible to include R - I calibrated stars 67, 76, 84, & 90. You should decide exactly how to place the CCD frame and locate the calibrated stars on a finder chart that is about 2 the size of the CCD frame. For focus stars consider SAO 122671 (mag=3) and SAO 123005 (mag=5). Note that SAO 122723 would make a good (bright: mag=6.8) star to aim the LX200 near this object. (I call such aiming/finding stars `LX200 stars'. The LX200 object library includes all SAO stars brighter than magnitude 7, but any bright star can help assure the telescope is aimed at what you intend (and that the telescope's reported RA/dec are accurate). You can load the SAO LX200.edb database into xephem to display the LX200 stars; SAO mag75.edb includes about twice as many stars (down to magnitude 7.5); SAO full.edb includes about a quarter of a million stars (down below magnitude 9). In such a large cluster, it would be wise to select additional fields, for example one centered near SAO 122709. A CCD frame there might include R - I calibrated stars: 39, 40, 43, 44, 49, 50, 58, 62. To select alternative standards. . . WIYN has a web page28 with recommended standard star regions. (WIYN's stars are designed for a large telescope, and hence would require long exposures on our telescope). Peter Stetson's extensive list of standards is also online29 . The xephem databases LONEOS.edb, Landolt83.edb and Buil91.edb contain shorter lists of brighter stars: < 13mag, 9.5mag and < 8mag respectively.) Generally these good standard stars will be well-separated from the target, so air mass air mass correction would be required. (Since I have internal `calibrated' sources, I'm not required to take CCD frames of these standards, however I've decided to `be prepared' and hence have recorded the basic data I would need to observe them.) I select: #109 (PG1633+099, at RA=16:35, LX200 star: SAO 121671), #121 (110 506, at RA=18:43, LX200 star: SAO 142582), and #125 (111 1969, at RA=19:37, LX200 star: SAO 143678, also see: SAO 124878). These standard stars can be marked as xephem Favorites from the UBVRI.edb database. These sources should span a good range of air mass in the general direction of the target. Following the instructions in using.xephem.txt, I create a file of basic location data for my target and standard fields. By default, this file is: .xephem/datatbl.txt and can be printed (% lp filename) or edited (% kwrite filename). Example 2: Plan for Lunar Month starting 19-August-2005 After Setting xephem for location: SJU Observatory, the new moon date 3-September2005, and Local Time 21:00 (9 p.m.), I find and record: UTC Time: 2:00, Sidereal: 18:35. NGC 6633, whose RA is most nearly equal to the sidereal time and hence is near the meridian, is the obvious target. After adding NGC 6633 to xephem Favorites and loading the SAO database (DataFiles; FilesSAO mag75.edb), it's easy to zoom in (lhs scroll bar) on 28 29 Photometry 97 NGC 6633 and find the nearby bright (5.7 mag) star SAO 12351630 , which would be a good star to steer the LX200. For focus stars consider SAO 122671 (mag=3) and SAO 123377 (mag=5). A visit to the NGC 6633 cluster web page31 finds six VRIe observations for R. Cross-reference shows that all six are 8 mag SAO stars32 , however only two (50 & 70) are in the central region of the cluster. (They could also be found using SAO full.edb.) Stetson reports BV I data for NGC 6633, however the brightest ten of his stars are 13 mag, which is 100 dimmer than the 8 mag R-mag stars. Thus different frames are required to properly expose the Stetson standards and the R `standards'. Using the Buil91.edb database, I can find bright standard stars: SAO 085402, SAO 141956, SAO 145050. These bright SAO stars are themselves LX200 stars. Example 3: Plan for Lunar Month starting 21-June-2005 After Setting xephem for location: SJU Observatory, the new moon date 6-July-2005, and Local Time 22:00 (10 p.m. -- it's not dark at 9 p.m.), I find and record: UTC Time: 3:00, Sidereal: 15:43. Because of its larger declination, Upgren 1 has a bit less air mass than NGC 6633, so it becomes the target. Upgren 1 is not in xephem's databases, so it must be added to Favorites following the procedure recorded in using.xephem.txt. A visit to the Upgren 1 cluster web page33 finds seven stars with VRIc and UBV observations. However the color index disagreements are of order 0.05, so we can use this as an opportunity to find the correct values. Using the Landolt83.edb database, I find three neighboring standard stars with a range of colors: HD 102056 (LX200 star: SAO 81968, bluer than the Upgren 1 stars), HD 106542 (LX200 star: SAO 100009, redder than the Upgren 1 stars), and HD 107146 (LX200 star: SAO 100038, similar to the Upgren 1 stars). Using the Oja96.edb database, I find two very close stars with similar colors: BD+35 2356 and BD+34 2338. (For these dimmer, 10 magnitude, stars finder charts are required.) #133 (SAO 63257, with mag 5 partner 63256) is a neighboring mag 3 star, SAO 44230 is a mag 4 star -- both can help focus and alignment. SAO 63118 is the nearest LX200 star to Upgren 1 (since Upgren 1 is not in the LX200 catalog of clusters, the final jump must be made based on RA/dec). All but one of the Upgren 1 seven stars is an SAO star; they are all easy to identify with Aladin. (FYI: spiral galaxies M94 and M63 might be worth a look.) Target and standards were fairly close together and near the meridian during the measurements, with airmass varying from 1.01 to 1.14. This was not a sufficient range to detect airmass correction terms. Increasing airmass dims and reddens stars. As a result calibrated stars viewed through a larger air mass will have a smaller constant in the color calibration equation Eq. 4.44. In this case, stars at airmass 1.14 were used to determine the color calibration equation which was then applied to unknown stars at airmass 1. The result is a bias in the estimated color of the unknown stars: their actual B - V is likely to be bigger than that calculated using the calibration equation (i.e., color calibration constant should be larger at airmass 1). The effect can probably be ignored for airmass0.1; Adjusting for this effect (e.g., determining34 the term B in Eq. 4.37) typically requires measuring stars 30 VizieR reports R = 5.703, for this star, which may serve as an additional standard. FYI: this star is also known as: HD 170200 and HR 6928. 31 Again: 32 These dim SAO stars are not in SAO mag75.edb so I located them using Aladin 33 Again: 34 On a following night, stars from IC 4665 were used to provide high airmass data. The results were: 98 Names M34, NGC 1039 IC 348 Collinder 69 M35, NGC 2168 NGC 2264, cone NGC 2301 NGC 2420 M67, NGC 2682 Upgren 1 (asterism) M5, NGC 5904 IC 4665 NGC 6633 NGC 6738 (asterism) NGC 6823 NGC 6940 Constellation Per Per Ori Gem Mon Mon Gem Cnc CVn Ser Oph Oph Aql Vul Vul Photometry Right Ascension (2000) 02h 42m 05s 03 44 30 05 35 06 06 09 00 06 40 58 06 51 45 07 38 24 08 51 18 12 35 00 15 18 33 17 46 18 18 27 15 19 01 21 19 43 12 20 34 32 Declination (2000) +42 45 42" +32 17 00 +09 56 00 +24 21 00 +09 53 42 +00 27 36 +21 34 49 +11 48 00 +36 18 00 +02 04 58 +05 43 00 +06 30 30 +11 36 54 +23 18 03 +28 15 08 Diameter (arcmin) 25 10 70 25 40 15 10 25 15 23 70 20 30 12 31 Reddening (mag) 0.07 0.93 0.10 0.26 0.05 0.03 0.05 0.06 0.03 0.17 0.18 0.86 0.25 Table 4.2: Candidate targets. and Allen's Astrophysical Quantities p. 548 have lists of clusters. An ideal candidate would be within telescope range (Dec: -5 < < 50 ), a size that matches our camera ( 15 ), have little reddening, and contain calibrated sources. with airmass1. Sample Data The large open cluster IC 4665 was the target for two nights during August 2005. The results are reported in Figures 4.14 and 4.15. During the first night an overlapping mosaic of five frames was used to image the central region of the cluster. During a 2 hour period BV RI frames were taken at each pointing. (During this time the air mass to IC 4665 changed from 1.30 to 1.46, but the data show no sign of a change in the transparency of the atmosphere, so no air mass corrections were applied.) Dome flat-field frames (taken 12 hours earlier) were used to produce reduced frames. gaia was used to measure the instrumental magnitude of 18 standard stars and three `unknown' stars. The magnitudes were pasted into a gnumeric spreadsheet which was used to calculate the color indices: b - v and r - i. (Since overlapping fields were used typically a standard star appeared in a couple of frames. When this occurred average instrumental color indices were used.) A calibrating line relating the standard stars' known B - V and the measured b - v was fit using regression in gnumeric. An identical process was used to find the relationship between R - I and r - i. The results are displayed in Figures 4.14(c) and 4.14(d). Using these relationships and the measured instrumental color indices, we can determine the standardized color indices of any star in the imaged region. As shown in Table 4.3, Eq. 4.28 or Eq. 4.29 can then be used to find the temperature of B - V = 1.30(b - v) - .59 - .09 sec z, R - I = .85(r - i) + .84 - .06 sec z Photometry Label A B C Name BD+05 3486 GSC 00428-00981 TYC 424-517-1 SpType A2 G5 K5 b-v 0.404 0.833 1.047 r-i -0.708 -0.387 -0.151 B-V 0.2822 0.8399 1.1181 R-I 0.1826 0.4625 0.6683 TB (K) 7255 5133 4421 99 TR (K) 6966 5015 4306 Table 4.3: The instrumental color indices: b - v and r - i of the `unknown' stars in IC 4665 were measured using gaia and converted to standardized color indices: B - V and R - I using the calibration lines displayed in in Figures 4.14(c) and 4.14(d). Eq. 4.28 (TB ) and Eq. 4.29 (TR ) were used to calculate the temperature of the stars from each color index. The two temperatures do not exactly agree; as discussed in the text systematic errors of a few percent are to be expected. each `unknown' star. During the second night six (generally not overlapping) frames were selected with the aim of capturing a diverse (hot and cool) set of standard stars. The data collected during this night did show a change in the transparency of the atmosphere (there was a larger variation in air mass: 1.32 to 1.63; see Figure 4.15(a)), so air mass corrections were applied. Report Checklist 1. Write an introductory paragraph describing the basic physics behind this experiment. For example, why does smaller (even negative) B - V correspond to higher temperature? Why do we expect v - V to be approximately constant? (This manual has many pages on these topics; your job is condense this into a few sentences and no equations.) 2. Files containing the reduced frames and object frames for your cluster in ST7 format. (At least one for each filter.) Leave these in your UNIX account, but identify the filenames in your notebook. While basic frame characteristics (filter, exposure time, date/time of exposure, . . . ) are contained in each file, you should also record this information in your notebook. 3. Files containing the flat-field frames you used to flat-field your images. (One for each filter, but perhaps derived as an average of several frames.) Leave these in your UNIX account, but record the filenames and characteristics in your notebook. 4. Hardcopy finder charts, with your target, calibrated stars and nearby SAO LX200 stars labeled. (Similar to Figure 4.14(a).) Record an RA/dec grid on your finder charts. 5. xephem datafile containing basic data (RA, Dec, altitude, air mass for each observation night). Print this out and tape in your lab notebook. 6. Table containing the basic data on your calibrated stars, including standard names and magnitudes. (Perhaps included in the below file.) 7. Spreadsheet file, reasonably well documented, containing known and instrumental magnitudes. For each filter, calculate instrumental-standard magnitudes (e.g., v -V ) 100 Photometry vV Calibration 28.26 28.24 28.22 vV 28.20 28.18 28.16 28.14 .2 .4 .6 .8 1.0 bv 1.2 1.4 1.6 (a) This Digitized Sky Survey image shows the central 50 25 region of the open cluster IC 4665. The standard stars of J.W. Menzies and F. Marang (MNRAS, 282, 313316 (1996)) are labeled along with three `unknown' stars: A, B, C. (b) There is little evidence for a significant color correction in this data set. See Eq. 4.42. BV Calibration 1.0 1.5 .8 .6 RI .4 .5 .2 0 .2 .4 .6 .8 1.0 bv 1.2 1.4 1.6 1.0 .8 .6 RI Calibration BV 1.0 0 .4 ri .2 0 .2 (c) Calibration of the standardize color index B - V in terms of the instrumental color index b - v. See Eq. 4.44. (d) Calibration of the standardize color index R - I in terms of the instrumental color index r - i. Figure 4.14: An overlapping mosaic of five frames was used to image the central region of the open cluster IC 4665. This region includes 18 standard stars which were used to calibrate the relationships between standardized color indices and instrumental color indices. Using these relationships and the measured instrumental color indices, we can determine the standardized color indices of any star in the imaged region. Photometry 101 Air Mass Corrections 31.46 31.44 <bB> or <iI> 31.42 31.40 31.38 31.36 1.35 1.40 1.45 1.50 Air Mass 1.55 1.60 vV 30.96 30.94 30.92 30.90 30.88 30.86 .6 .8 Color Correction 1.0 1.2 bv 1.4 1.6 1.8 (a) Short wavelengths are more affected by atmospheric absorption than long; the above shows the effect on B magnitudes (filled squares) and I magnitudes (open circles). The y axis displays the difference between instrumental and standardized magnitudes averaged over all the blue stars in the frame. (b) Color correction is evident in this data set, but both air mass and color correction should be applied to this data. This full result was: v - V = 30.611 + 0.179 sec z + 0.048(b - v). The rms variation from this line was less than 0.02 mag. BV Color Calibration RI Color Calibration 1.5 .8 BV RI .4 .2 0 .6 .8 1.0 1.2 bv 1.4 1.6 1.8 .8 .6 .4 ri .2 0 .2 1.0 .6 .5 0 (c) Calibration of the standardize color index B - V in terms of the instrumental color index b - v. B - V = -0.67 + 1.29(b - v) (d) Calibration of the standardize color index R - I in terms of the instrumental color index r - i. R - I = 0.80 + 0.86(r - i) Figure 4.15: Six frames were used to image 22 standard stars in the open cluster IC 4665. The selected stars have a wide variation in color in order to to achieve robust color calibrations. During the 2+ hours of data collection the air mass increased from 1.32 to 1.63 and the resulting increase in atmospheric absorption increased (dimmed) B, V, R, and I magnitudes in proportion to the air mass (see Figure 4.15(a)). 102 Photometry and record the basic statistics (mean, standard deviation). (Note that aside from a small color correction, we expect v - V to be constant.) You should use this spreadsheet to calculate instrumental color indices and prepare the data for the following plots. 8. Hardcopy plot of B - V vs. instrumental b - v with a fitted line. (Similar to Figure 4.14(c).) 9. Hardcopy plot of R - I vs. instrumental r - i with a fitted line. (Similar to Figure 4.14(d).) 10. Hardcopy plot of v -V vs. known B -V with a fitted line. (Similar to Figure 4.14(b).) 11. Calculation of the temperature of several unknown stars based on b - v. (See Table 4.3.) Note: the error in B - V for the unknown stars can be estimated from the deviation-from-fit observed for the standard stars. The `high-low' method will allow you to propagate that B - V error into a TB error. Calculation of the temperature of those same stars based on r - i. 12. Comments on the uncertainty in temperature. What do you believe are the major sources of uncertainty? Other Projects The lab is aimed at one fairly simple project: measuring stellar magnitudes using in-frame calibrated stars. However, the techniques discussed here can be applied to a variety of other projects, some of which are listed below. 1. Imaging of extended objects (galaxies, nebulae, planets,. . . ) 2. Orbit determination of Solar System objects (moons, comets, asteroids,. . . ) 3. Magnitude as a function of time (a) Pulsing stars: Delta Scuti (, RR Lyrae, Cepheids (b) Cataclysmic variables ( (c) Eclipsing binaries ( (d) Novae and supernovae (e) Rotation period of an asteroid (f) American Association of Variable Star Observers ( projects 4. Careful determination of any significant parameter of the telescope system (e.g., S/N issues, atmospheric extinction, color corrections, tracking, . . . ) Photometry 103 References 1. Santa Barbara Instrument Group (, Operating Manual for ST-7E, st-7e.pdf 2. Santa Barbara Instrument Group (SBIG), Users Guide for CCDOps Version 5 CCDops v5.pdf 3. Meade (, Instruction Manual for 12" LX200 LX200 Classic Manual.pdf 4. Howell, Steve B., Handbook of CCD Astronomy, Cambridge, 2000 5. Howell, Steve B., Astronomical CCD Observing and Reduction Techniques, ASP Vol. 23 1992 6. Buil, Christian, CCD Astronomy, Willmann-Bell, 1991 7. Romanishin, W. An Introduction to Astronomical Photometry Using CCDs, 2006 8. J. Palmer & A.C. Davenhall [starlink], The CCD Photometric Calibration Cookbook, starlink ccd.pdf 9. ed. Cox, Arthur N., Allen's Astrophysical Quantities, AIP/Springer, 2000 10. American Association of Variable Star Observers (, CCD Manual aavso ccdmanual.pdf 104 Photometry 5: Bubble Chamber Purpose The purpose of this experiment is to determine the rest mass of the pion (m ) and the rest mass of the muon (m ). Introduction Particle physics (a.k.a. high energy physics) is the division of physics which investigates the behavior of particles involved in "high" energy collisions. ("High" here means energies greater than those found in nuclear reactions, i.e., more than 100 MeV=0.1 GeV. The highest energy particle accelerators available today produce collisions with energies of a few million MeV = TeV.) The first "new" particles discovered (circa 1940) by particle physicists were the pion () and the muon (). In spite of roughly similar masses (near 100 MeV, compare: electron mass = .511 MeV and proton mass = 938 MeV), these two particles have quite different properties. The muon is a relative of the electron (and hence is called a lepton). It comes in particle 1 (- ) and anti-particle (+ ) versions and has spin 2 . Unlike the electron, the muon is unstable. It decays into two neutrinos () and an electron (or positron) after a mean life of 2 10-6 s: + - + + e+ - (5.1) (5.2) - + + e - The pion belongs to the class of particles called mesons. Unlike leptons, mesons interact with protons and neutrons through an additional force called the strong nuclear force (a.k.a., color force). (Particles that can feel this force are called hadrons.) Unlike leptons, mesons are known to be composite particles: each is made of a quark and an antiquark. The pion comes in three versions: + , 0 , and - and has spin 0. All the pions are unstable; the + decays after a mean life of 3 10-8 s: + - + + . (5.3) (The 0 has a slightly smaller mass and decays much faster than the . It is not seen in this experiment.) 105 106 Bubble Chamber Particle Detection Since the particles studied by particle physics are sub microscopic and decay "quickly", particle detection is a problem. Most existing particle detectors rely on the fact that as a charged particle moves by an electron (e.g., an electron in an atom of the material through which the charged particle is moving), the electron feels a net impulse. If the charged particle comes close enough to the electron and/or is moving slowly enough (so the interaction is long enough), the impulse on the electron will be sufficient to eject the electron from its atom, producing a free electron and an ion. Thus a charged particle moving through material leaves a trail of ions. This trail can be detected in many ways (e.g., by direct electronic means as in a modern wire chamber or chemically as when the material is a photographic plate or emulsion). In this experiment the ion trail is made visible by vapor bubbles which are seeded by individual ions in boiling material (here liquid hydrogen). The bubbles are large enough to be photographed whereas the ion trail itself is much too narrow. Relativistic Kinematics Recall the following from Modern Physics: E = mc2 T pc = mvc = mc = E E 2 - (pc)2 = where: = v/c = 1 1 - 2 (5.8) (5.9) mc 2 2 (5.4) 2 2 E - mc (5.5) (5.6) 2 2 1 - 2 = mc (5.7) and v is the velocity of the particle with rest mass m, momentum p, total energy E and kinetic energy T . Note that E, T , pc, and mc2 all have the dimensions of energy; it is customary to express each in MeV and even say "the momentum of the particle is 5 MeV" or "the mass of the particle is 938 MeV." (Of course, technically the momentum of the particle would be 5 MeV/c and the mass 938 MeV/c2 . Basically what we are doing is redefining "momentum" to be pc and "mass" to be mc2 . Since the "c" has disappeared, this re-naming is sometimes called "setting c = 1".) For future reference, note from Equation 5.6 that if 1, E pc and from Equation 5.7 that if m = 0, E = pc. Of course, massless particles (like light) must travel at the speed of light (i.e., = 1). Momentum Measurements Classically a charged particle (with mass m and charge q) moving through a magnetic field B has an acceleration, a, given by ma = qv B (5.10) Bubble Chamber 107 Because of the cross product, the acceleration is perpendicular to both v and B. Thus there is zero acceleration in the direction of B, so v , the component of velocity parallel to B, is constant. On the other hand in the plane perpendicular to B, the acceleration and the velocity are perpendicular resulting in centripetal (circular) motion. Thus the particle moves in a circle of radius R even as it travels at constant speed in the direction of B. The resulting motion is a helix (corkscrew). Using to denote components in the plane perpendicular to B, we have: ma = 2 mv = qv B R p = mv = qBR (5.11) (5.12) This last relationship continues to hold for relativistic particles. SHOW: For a positron, Equation 5.12 means the momentum p c (in MeV) can be calculated as a simple product 3BR: p c (in MeV) = 3BR (5.13) where B is in Tesla and R is in cm.1 In this experiment, positrons (and electrons) from muon decay circle in an applied magnetic field. You will measure the radii of the positron orbits to determine positron p . Since the rest mass of the muon has been converted to kinetic energy of its decay products, positron p depends on muon mass and measurement of p allows calculation of m . Kinetic Energy Measurement As stated above, a charged particle moving through a material leaves a trail of ions. The energy needed to form these ions must come from the kinetic energy of the charged particle. Thus, every cm of travel results in a kinetic energy loss. It can be shown (Bethe-Block) that the decrease in kinetic energy depends on the inverse of the velocity squared: 2.1 dT =- 2 dx where is the density of the material ( = .07 example of a "calculator equation"2 . MeV/cm (5.14) g/cm3 for liquid H2 ). This is the second A particle with some initial kinetic energy T0 will travel some definite distance, L, before all of its kinetic energy is lost and it comes to rest. The relationship between T0 and L can be determined from the energy loss per cm: 0 L= T0 dx dT = dT T0 0 2 dT 2.1 (5.15) 1 This is an example of a "calculator equation" where we seemingly ignore units. That is if B = 2 T and R = 5 cm, this equation says p c = 3 2 5 = 30 MeV, units seemingly just tacked onto the answer. To derive such an equation, you must demonstrate how the units work out. In particular, p c -- which in MKS units in going to naturally come out in Joules -- must be converted to the energy unit MeV. The conversion factor is 1 MeV=1.6022 10-13 J. Of course, you already know 100 cm=1 m. 2 Thus if = .1 g/cm3 and = .5 we would conclude that dT /dx was .84 MeV/cm. 108 Actual Experiment camera Bubble Chamber Simplified Example in just 2 dimensions z actual path: length L y x apparent path (from photo): length L z actual path: length L x apparent path (from photo): length L Figure 5.1: This experiment uses photographs of particle paths in a bubble chamber. Two angles ( [0, 180 ], [0, 360 ]) are required to describe the orientation of the path in three dimensional space. The photographic (apparent) path length, L , is shorter than the actual path length, L, (of course, if = 90 , L = L). In general: L = L sin . The angle just describes the orientation of the apparent path in the photograph. We can make an easier-to-understand model of perspective effects by just dropping and considering a two dimensional experiment. In this case to generate all possible orientations [0, 360 ]. For particles moving much slower than the speed of light, Newton's mechanics is a good approximation: T = 1 mv 2 = 1 mc2 2 2 2 L= so T0 L1/2 . In this experiment muons produced by pion decay travel a distance L before coming to rest. You will measure the muon path length to determine muon T0 . Since the kinetic energy of the muon comes from the rest mass of the decaying pion, the mass of the pion can be calculated from muon T0 . 2 2.1mc2 T0 T dT = 0 2 T0 2.1mc2 (5.16) Perspective Effects In real particle physics experiments, decay events are reconstructed in three dimensions. However in this experiment you will measure apparent muon path lengths from photographs. Because of perspective effects, typically the true path length (L) is longer that the apparent (photographic) path length (L ), as the particle will generally be moving towards or away from the camera in addition to sideways. In this experiment we need to "undo" the perspective effect and determine L from the measurements of L . There are several ways this could be done. Perhaps the easiest would be to pick out the longest L , and argue that it is longest only because it is the most perpendicular, i.e., max ({L }) L (5.17) Bubble Chamber 109 .94 .83 .85 .49 .40 .18 .97 .94 .28 Figure 5.2: Nine randomly-oriented, fixed-length segments are placed on a plane and the corresponding horizontal lengths L (dotted lines) are measured (results displayed below the segment). The resulting data set {.94, .83, .85, .49, .40, .18, .97, .94, .28} of L can be analyzed to yield the full segment length L. (The angle [0, 360 ] describes the orientation, but it is not measured in this "experiment": only L is measured.) Essentially this is a bad idea because it makes use of only one collected data point (the maximum L ). For example, it is likely you will make at least one misidentification or mismeasurement in your 60+ measurements. If the longest L happens to be a bad point, the whole experiment is wrong. Additionally since L is the net result of interactions with randomly placed electrons, L is not actually exactly constant. (That is, Equation 5.14 is true only "on average".) Paths that happen to avoid electrons are a bit longer. The L-T0 relationship is based on average slowdown; it should not be applied to one special path length. One way of using all the data is to note that randomly oriented, fixed-length paths will produce a definite average L related to L. So by measuring the average L (which we will denote with angle brackets: L ), you can calculate the actual L. It will be easier to explain this method if we drop a dimension and start by considering randomly oriented, fixed-length segments in two dimensions. Figure 5.2 shows3 nine randomly oriented segments in a plane with the corresponding measured L . The different measured L are a result of differing orientations of a fixed-length segment: L = L| sin | (5.18) From a sample of N measurements of the horizontal distance L (i.e., a data set of measured L : {xi } for i = 1, 2, . . . , N , with corresponding orientations {i } with i [0, 2]), the average L could be calculated L 1 = N N i=1 L xi = N N i=1 | sin i | (5.19) The i should be approximately evenly distributed with an average separation of = 2/N (because there are N angles distributed throughout [0, 2]). Thus, using a Riemann sum approximation for an integral: L = L N L 2 N i=1 N | sin i | = L 2 i=1 | sin i | 2 0 2 0 2 (5.20) (5.21) 0 | sin | d = L | sin | d d 3 Note that if we applied Equation 5.17, we would conclude L = .97 with no estimate for the uncertainty in this result (i.e., L). 110 The above integral is easily evaluated: 2 0 Bubble Chamber | sin | d = 2 sin d = 2 0 - cos 0 =4 (5.22) Thus we have the desired relationship between L and L: L = L 2 (5.23) With the example data set we have: L = 0.653 with standard deviation L = 0.314. Using the standard deviation of the mean we have: 2 .314 0.65 = 0.65 .10 = L 10 1.03 .16 = L (5.24) (5.25) Note that our argument for finding averages is quite general, so if random values of x are uniformly selected from the interval [a, b], the average value of any function of x, f (x), can be calculated from: b f (x) dx (5.26) f (x) = a b dx a For the actual experiment, the path orientations have a uniform distribution in space. That is, if all the paths originated from the same point, the path ends would uniformly populate the surface of a sphere of radius L. The element of surface area of a sphere of radius L is: L2 d = L2 sin d d (5.27) where is called the solid angle and plays an analogous role to radian measure in a plane: plane angle in radians = solid angle in steradians = arc length R sphere surface area R2 (5.28) (5.29) Thus the relationship between L and L in three dimensions is: L = L sin = L sin d =L 4 d (5.30) SHOW this result! Note: d = sin d d and the range of the double integral is [0, ] and [0, 2] Comment: The above discussion has been phrased in terms of position vectors, but it applies as well to any vector. In particular, you will be measuring the perpendicular component of momentum, p , and need to deduce the actual momentum, p. Exactly as above, if the particles have the same speed with direction uniformly distributed in space: p = p 4 4 (5.31) If the particles actually have differing speeds we can still conclude: p = p (5.32) Bubble Chamber 111 1.0 .8 Cumulative Fraction .6 .4 .2 0 0 .2 .4 L .6 .8 1.0 Figure 5.3: The distribution of data set: {.94, .83, .85, .49, .40, .18, .97, .94, .28} of Figure 5.2 (nine randomly oriented segments) displayed as a cumulative fraction. Displaying Distributions As discussed above, when finding L it is best to use the entire data set. Although the L method uses all the data, it quickly reduces the whole data set to one number. Is there some way of graphically displaying and using all the data? In particular, is there some way of checking to see if the data have the expected distribution (i.e., the right proportion of long and short L s)? Perhaps the easiest way to understand the idea of a distribution is to consider the idea of the cumulative fraction function for some data set: {xi }, for i = 1, 2, . . . , N . The cumulative fraction function4 , c(x), reports the fraction of the data set {xi } that is less than or equal to x. Obviously if a < min ({xi }), c(a) = 0; if b max ({xi }), c(b) = 1; and if c(x) = .5 then x is the median (i.e., middle) data point. In the 2d example data set, c(.60) = 4/9, because four of the nine data points are smaller than .6. Similarly c(.84) = 5/9. Every time x increases past one of the xi , c(x) has a jump. See Figure 5.3 for a plot of this function. The function c(x) depends on the data set, so if the experiment is repeated generating a new data set {xi }, a new function c(x) is also generated. The new c(x) should be slightly different, but generally similar to the old c(x). If the data set is sufficiently large, the new and old c(x) will be quite similar and both c(x) would approximate the function c(x), the ^ cumulative fraction function that would be generated from an infinite-sized data set5 . How can c(x) be best approximated from one finite-sized data set {xi }? To answer this question ^ it will be convenient to consider the data set {xi } already sorted so x1 is the minimum and 4 The cumulative fraction function is also known as the empirical distribution function, and is closely related to percentiles and order statistics. 5 The usual name for c(x) is the distribution function. ^ 112 Bubble Chamber 1.0 .8 Cumulative Fraction .6 .4 .2 0 0 .2 .4 L .6 .8 1.0 Figure 5.4: The cumulative fraction function for the example data set is plotted along with the data points for the percentile estimate of the distribution function c(x). ^ xN is the maximum. Thus our example data set: {.94, .83, .85, .49, .40, .18, .97, .94, .28} becomes: {.18, .28, .40, .49, .83, .85, .94, .94, .97} As defined above, c(x) is given by: c(x) = i N where i is such that: xi x < xi+1 (5.35) (5.34) (5.33) That is to determine c(x) for some x, we see how far down the sorted list we must travel to find the spot where x fits between two adjacent data points: xi x < xi+1 . Clearly there are a total of i data points less than or equal to x (out of a total of N ), so c(x) = i/N . If x happens to equal one of the data points, things are a bit undefined because c(x) has a jump discontinuity at each xi . It turns out that the best estimate for c at these discontinuities is: ^ c(xi ) = ^ i i N +1 (5.36) Of course this estimate can be wrong; it has an uncertainty of = i (1 - i ) N +2 (5.37) See Figure 5.4 for a comparison of the estimated c(xi ) (called the percentile) and the ^ cumulative fraction function. (The mathematics of these results is covered under the topic "order statistics" or "nonparametric methods" in advanced statistics books.) Bubble Chamber 113 Figure 5.5: A particular line segment is displayed along with the measured L (dotted line). What fraction of randomly oriented segments would have a L smaller than this particular segment? The darkly shaded part of the circle shows possible locations for these small L segments. The fraction of such small L segments should be the same as the dark fraction of the circle: 4/2. Your experimental estimate of c should be compared to the theoretically expected c. The ^ ^ example data set was generated from randomly oriented line segments in a plane. As shown in Figure 5.5, it is expected that the fraction of a data set less than some particular value of L is: c(L ) = ^ = 4 where: = arcsin(L /L) 2 2 arcsin(L /L) (5.38) (5.39) Our formula for c involves the unknown parameter L; we adjust this parameter to achieve ^ the best possible fit to our experimental estimate for c. Using the program fit: ^ tkirkman@bardeen 7% fit * set f(x)=2*asin(x/k1)/pi k1=1. * read file cf.L.dat * fit Enter list of Ks to vary, e.g. K1-K3,K5 k1 FIT finished with change in chi-square= 5.4810762E-02 3 iterations used REDUCED chi-squared= 0.2289333 chi-squared= 1.831467 K1= 0.9922597 Using the covariance matrix to determine errors6 , we conclude k1 = 0.992 .025. This 1 reported random error is about 6 that obtained above using L . SHOW: Derive yourself the theoretical function c(L ) for line segments in space. Hint: Begin by noting that if the segments shared a common origin, the segment ends would uniformly populate the surface of a sphere of radius L. Segments with measured L less than some particular value would lie on a spherical cap, the three dimensional version of 6 Reference 2, Press et al., says usually error estimates should be based on the square root of the diagonal elements of the covariance matrix 114 Bubble Chamber 1.0 .8 Cumulative Fraction .6 .4 .2 0 0 .2 .4 L .6 .8 1.0 Figure 5.6: The theoretical distribution function (Equation 5.39) fit to the "experimental" data points derived (Equations 5.36 & 5.37) from the example data set. As a result of the fit we estimate: L = 0.992 .025. arc caps displayed in Figure 5.5. The ratio of the area of these caps to the total surface area of the sphere gives the expected value for c. You will need to calculate the area of a ^ spherical cap by integration. The above discussion has focused on path lengths as that is the quantity measured in pion decay. In muon decay, the radius of positron orbits in the applied magnetic field is measured. Weinberg-Salam theory provides a complete description of the decay process, including the distribution of positron momentum (which in turn determines the radius of positron orbits R). Kirkman has shown that the the Rs should be distributed according to c(R) = 3 2 x -1 2 1 - x2 + 1 - 1 4 x log 2 1+ 1 - x2 /x (5.40) where x = R/Rm , and Rm , the maximum value of R, is the value of R that corresponds to p c = m c2 /2. The adjustable parameter Rm (of course called k1 in fit), can be adjusted to give the best possible fit to the experimental distribution. From Rm , m can be determined. Biases and Robust Estimation The bane of every particle physics experiment is bias. Biases are data collection techniques that produce nonrepresentative data. For example, short L are harder to notice than long L , and thus long L tend to be over represented in the data sample, producing a high L . Use of the cumulative fraction function allows this biases to be detected. In addition Bubble Chamber 115 to biases, the cumulative fraction function allows you to detect likely mistakes: for example, particle path lengths that are extraordinary given the entire data set. The detection of a likely mistake suggests corrective actions like removing the "bad" point. You should almost never do this! (You will find a chapter in Taylor on this "awkward" and "controversial" problem.) A better option is to use analysis methods that are "robust", i.e., that are insensitive to individual "bad" points.7 Imagine we modify our example data set by adding a "bad" point: L = 2: {.18, .28, .40, .49, .83, .85, .94, .94, .97, 2.00} (5.41) Adding this outlying8 data point totally messes up the max({L }) method (the least robust method). Since it increases both the mean and the standard deviation, the estimated L based on the L method shifts from 1.03 .16 to 1.24 .26. Changing the number of data points requires recalculating the estimated distribution function for every point (because the value of c depends on the set size N ). If we carry through ^ the total analysis with our enlarged data set we find the fit L shifts from 0.992 .025 to 1.05 .05. We can conclude that the cumulative fraction method is less sensitive to bad data than the average method.9 Unconscious (uncontrolled) biases produce tainted data which can be rescued in part by robust estimation. Once bias is recognized the experiment can be rearranged to adjust for its effects. This requires that the bias be exactly reproducible. For example, short apparent muon paths (paths mostly towards or away from the camera) are inconspicuous and hence more likely to be missed on some occasions. One solution is formalize this bias and intentionally ignore all photographic paths shorter than say .3 cm long (about 5% of the data). This cut (formalized noncollection of data), can be included in the theoretical distribution function so it will not affect parameter estimation. Experimental Arrangement Our bubble chamber photographs were taken using the 385 MeV proton accelerator at Nevis Lab which is a part of Columbia University. A pion beam was produced by colliding accelerated protons with a copper target. The pion beam was directed through an absorber to slow the pions so that a sizeable fraction of the pions would come to rest in the adjacent bubble chamber. See Figure 5.7. The path of charged particles from the pion decay ( e) as recorded by a nearby camera, encodes the information needed to calculate m Removing a data point is a lie. A more subtle sort of lie comes from the existence of choice of methods. Clearly, you can analyze the data several different ways, and then present only the method that produces the answer you want. Darrell Huff's book How to Lie with Statistics (Norton, 1954) can help you if that is your goal. I probably don't need to remind you that schools with "Saint" in their name do not recommend this course of action. Choice and ethics are interlocking concepts. 8 While not exactly relevant, this data point is 2.34 above the mean and hence an outlier by Chauvenet's criterion (see Reference 4, Taylor). It is also an outlier by Tukey's criterion (see Reference 3, Hogg & Tanis). 9 Do note that both results remain consistent with the intended value of 1.00. Also note that the median could have provided a more robust alternative to the average. However, that would have required a discussion of the uncertainty in the median, which is beyond the intended aims of this lab. In this lab--and in most any experiment--there are many possible ways to analyze the data. Choice of method often involves art and ethics. 7 116 Bubble Chamber Pion Beam Absorber Magnet Coil Camera 12" 40" Bubble Chamber 6" Figure 5.7: A pion bean is slowed by an absorber so the pions are likely come to rest inside a liquid hydrogen bubble chamber. The path of charged particles from the pion decay ( e) is recorded by a nearby camera. (Of course, the uncharged neutrinos from the decay leave no ion trail, and hence no bubbles grow to mark their path.) A magnetic field (B = .88 T, directed toward the camera) produced by the current in the coil, bends the path of all charged particles into helixes, but the effect is most visible with the low mass electrons. The radius of the helix, as recorded by the camera, can be related to the particle's p . Since the muon decays into three particles, allowing varying distribution of energy, the electron's momentum can vary from 0 up to a maximum of m c/2. Since the pion decays into just two particles, there is only one way to distribute the released energy so the muon's initial kinetic energy is determined uniquely which produces a fixed stopping distance L. (Of course, the apparent muon path length, L , recorded by the camera will vary.) Bubble Chamber - - e- 117 key B magnetic field points out of page stationary muon decays muon path incoming pion electron path stationary pion decays Figure 5.8: A typical decay process as recorded in a bubble chamber photograph. A pion ( - ) slows and comes to rest inside the bubble chamber. A short time later it decays into a muon and an antineutrino ( - - + ). A kink in the path marks the decay location. The muon is in turn slowed and comes to rest after traveling a short distance. A short time later the muon decays into an electron and two neutrinos (- e- + + ). The high speed electron leaves a sparse track in accord with the Bethe-Block Equation (5.14) (reduced energy loss due to large means fewer ions produced and hence fewer bubbles). You will be measuring apparent muon path lengths, L , and electron helix radii R. and m . Apparent muon path length, L , will allow you to determine the actual muon path length L, from which in initial muon kinetic energy T0 can be determined. A magnetic field (B = .88 T, directed toward the camera) bends the path of all charged particles into helixes; but the effect is most visible with the low mass electrons. The radius of the electron helix determines (Equation 5.13) the electron's p c. From the distribution of electron p c, you can determine both pc and the maximum pc, from which m can be determined. Figure 5.8 shows an idealized decay sequence as might be recorded in a bubble chamber photo. The paths of interest start on the left (pions from the accelerator), have a short (1 cm) kink (muon), connecting to a sparse loop (electron). Procedure Measurement of Rs and L s is computerized. (Indeed almost all of your data collection this semester will be computerized.) As I'm sure you know, while computers can be useful devices, they seemingly have a knack for unintended/unexpected disasters, which are called `user error'. Thus the most important lesson of computer use is: GIGO (`Garbage In; Garbage Out'10 ) -- a computer's output must be considered unreliable until, at the very least, you know the limitations/uncertainties in the input data producing that output. In this lab you will be using a CalComp 2500 digitizing tablet to measure distances. The process seems simple (aligning a point between crosshairs and clicking to take the data10 Sometimes this acronym is reported as `Garbage In, Gospel Out' stressing many people's (mistaken) faith in computer output. 118 Bubble Chamber point) but involves problem of definition errors (including systematic biases, see page 13) in addition to more familiar device limitations (random and calibration errors). To have some justified confidence in this process, you must measure a known and see what the computer reports (`trust, but verify'). I have provided you with a simulated bubble chamber photo in which all the path lengths are 1 cm and all the curvatures are 20 cm. (If you don't trust this fiducial--and you might not since it depends on the dimensional stability of printers and paper--you can measure the `tracks' with an instrument you do trust.) Begin by logging into your UNIX account using the Visual 603 terminal with attached CalComp 2500 digitizer, and running the program bubbleCAL: tkirkman@linphys8 1% bubbleCAL The following directions are displayed: The general procedure will be to place the cursor crosshair at the needed place and press a cursor button to digitize. Press cursor button "0" when obtaining muon path lengths (digitize beginning and end of track); press cursor button "1" when obtaining electron radii of curvature (digitize three points on curve, from which the computer can figure R); press cursor button "2" to cancel an in-progress data point or clear error; press cursor button "3" to remove the last data point of the presently selected type. Additional data points may be removed by number when done. Digitizer will beep between data points. Hit ^D (control D) on keyboard when done. Files containing your data (unsorted) will be created: Lcal.DAT & Rcal.DAT. THIS PROGRAM IS FOR CALIBRATION/TESTING NOT DATA COLLECTION With the simulated bubble chamber photo taped in place on the digitizer, check some long distances (> 10 cm) on the scales. Then measure 16 path lengths and 16 curvatures. Record the data reported by the program (means, standard deviations, 95% confidence intervals, etc.). Does the probable range for each mean include the known value? (If not discuss the problem with Dr. Kirkman.) Note that if you took more data points, the probable range for the mean would become increasingly small and eventually you would detect a systematic error limiting the ultimate accuracy of the device. The program bubble works very much like bubbleCAL, except in the end it will produce files containing the cumulative fraction of L (cf.L.dat) and R (cf.R.dat). Please do not deface the bubble chamber photos! Lightly tape each bubble chamber photograph to the digitizer to keep the photo from moving as you collect data. Collect 64 L s and 64 Rs. (This will require scanning about 20 bubble chamber photos.) The program will make four files: cf.R.dat contains the usual three columns: R, the percentile estimate for c(R) calculated from your data, and the error in the estimate, cf.L.dat similarly contains the sorted L , estimated c(L ) and error, L.DAT and R.DAT contain the (unsorted) raw data. Use the web11 or spj (see page 173) or gnumeric (Linux spreadsheet) to calculate L and R and their standard deviations. Note that the raw data files (L.DAT and R.DAT) can be used to transfer the data to these applications (either by copy & paste or READ). 11 paste form.html Bubble Chamber 119 Use fit (see page 167) to fit each dataset (cf.R.dat & cf.L.dat) to the appropriate theoretical distribution function. Produce a hardcopy of your fit results to include in your notebook. Use plot (see page 161) to produce a hardcopy plot of your data with fitted curve to include in your notebook. The formula for c(R) (Eq. 5.40) is a bit tricky to type in, so I've provided you with a shortcut: c(R) will be automatically entered into fit (with K1 as the adjustable parameter Rm ) if, in fit, you: * @/usr/local/physics/help/ and c(R) will be automatically entered into plot (with K1 as Rm ) if, in plot, you * @/usr/local/physics/help/ Copy and paste is the easiest way to transfer K1 between fit and plot. Calculation of m Since energy is conserved, the rest energy of the muon must end up in its decay products: E = m c2 = Ee + E + E (5.42) where E , Ee , E , E are respectively the total energy of the muon, electron, neutrino, and antineutrino. You should expect that the rest energy of the muon would, on average, be evenly divided between the three particles. Thus Ee E E 1 m c2 3 (5.43) In fact, the Weinberg-Salam theory of weak decays predicts Ee = .35m c2 (5.44) Since electrons with that much energy have = .9996, to a good approximation Ee = pe c. Thus 4 pe c (5.45) .35m c2 = Ee = pe c = so you will find m c2 from pe c (which, in turn, may be found from R ). Additionally, you will fit the unknown parameter Rm in the theoretical expression for c(R) ^ (Equation 5.40) to the experimental data. From the above discussion you already know 3BRm = p c = m c2 /2 allowing you to determine m c2 from B and Rm . Recall: Standard deviation of the mean is used for errors in averages (see Taylor). The square root of the diagonal elements of the covariance matrix determine the errors in fit parameters (See Press, et al.). (5.46) 120 Bubble Chamber Calculation of m The pion mass can be determined from the muon path length L: From L you can find the initial kinetic energy of the muon (T0 ); adding the rest energy of the muon (use a high accuracy book-value for m c2 ) gives you the total muon energy, E . E = T0 + m c2 (5.47) From momentum conservation and the fact the neutrinos have nearly zero rest mass and hence travel at the speed of light, you can show 2 E = p c = p c = E - (m c2 )2 1/2 (5.48) Finally, energy conservation of the decaying pion requires 2 m c2 = E + E = E + E - (m c2 )2 1/2 (5.49) from which m c2 can be calculated. We have discussed two ways of determining L (using Equation 5.30 with a measured value for L and by fitting the theoretical c(L ) curve to your experimental data), so you will ^ produce two values for m . Note: The error in the book value of m used to calculate m should be small enough to ignore, so the error in m c2 (m c2 ) is due to the uncertainty in L (given by fit) or the uncertainty in L (given by the standard deviation of the mean). To calculate m c2 , m c2 and E = T0 and apply Eq. E.9 (or E.12) in the form: find E m c2 = m c2 E E (5.50) Some Words About Errors We noted above that different analysis methods yield different statistical error estimates. And robust methods that produce smaller uncertainties are preferred. But no amount of statistical gamesmanship can erase a systematic error in the original measurements so there is little point in reducing statistical errors once systematic errors dominate. Said differently: the aim is to understand the systematic errors and then reduce the statistical errors until they become irrelevant. In this experiment one finds potential systematic errors in constants given without error (B = .88 T, = .07 g/cm3 , the "2.1" in the Bethe-Block Equation 5.14) and theoretical simplifications (the bubble chamber's 6" thickness means paths closer to the camera are enlarged in photographs compared to those further from the camera, `straggling' where some muons would have stopping distances a bit more or less than that calculated from the BetheBlock Equation, varying magnetic field within the bubble chamber, expansion/contraction of the photographs due varying to humidity). Bubble Chamber 121 m c2 (MeV) 113 109 106 102 103 110 112 114 114 105 97 109 109 mean: standard deviation: discrepancy (mean-known): 108 5 2 1 m c2 (MeV) 140.36 140.10 139.61 139.98 140.17 140.15 140.26 139.55 139.92 140.09 140.12 140.24 140.49 140.08 0.27 0.51 .02 Table 5.1: Two years of student results for the muon mass (m c2 in MeV) and the pion mass (m c2 in MeV) as determined using a fit to the appropriate cumulative fraction function. Values using averages were similar (if slightly more discrepant) but with much larger errors. The error listed here is just typical: clearly each experiment will report both differing errors and values. Note particularly that results show much more variation than would be expected given the reported errors and the fact that identical equipment was used. 122 Bubble Chamber The error in L determined from the fit to the theoretical cumulative fraction function is often obviously too-small12 . The covariance matrix typically suggests an L error of about .001 cm. Whereas bubbleCAL typically suggests systematic errors in L measurements greater than .01 cm. Further the likely error in the and 2.1 that occur together with L in Eq. 5.16, would suggest L accuracy below 5% is irrelevant. It is helpful to review multiple results collected over a couple of years given in Table 5.1. The mean discrepancy is not too worrisome: a few percent change in the given constants would erase that difference. Of greater concern is large mismatch between the calculated statistical error and the standard deviation upon repeated measurement with identical equipment (but different data collectors). I'm sure part of this variation is due to `personal placement decision' detected by bubbleCAL, but additionally varying ability to notice short L will create varying results. (Most student cumulative fraction plots show a deficit of short L .) In any case we see here an example of calculated statistical errors reduced way below other uncontrolled uncertainties. Given this table of experiment repeats, would it be fair to report the standard deviation of the mean as the uncertainty in, say m c2 ? Numerically you can see that the result would 1 be disastrous (as the SDOM is about 7 the discrepancy). Since we are attributing the deviation to different observers (since the equipment and photos are the same) and we have no right to suggest that the average observer is the perfect observer (indeed one might say the perfect observer is exceptional), the answer must be "No". In short you can only use the SDOM as the error if you are convinced the deviation is caused by a balanced (exactly as much high as low) process. (And that implies you have a good idea of what is causing the deviations.) In our experiment, the actual experimental error is available to us (since m and m have since been measured to high accuracy), but how could the original investigators estimate these errors? A short answer to this question is calibration: "placing" known pathlengths and radii into the bubble chamber, and comparing the known values to the measured results. Indeed, experiments exactly like this lab are used to calibrate the detection systems in modern experiments. Histograms The cumulative fraction function described above is generally used when fewer then a thousand data points are available. Histograms are a more familiar way of displaying distributions. Histograms are made by dividing the range of the data set {xi } into several (usually equal-sized) sub intervals call bins. The data are sorted into these bins, and the number of data points in each bin is recorded as the y value for the average x value of the bin. (According to Poisson statistics, the uncertainty in y would be the square root of y.) The resulting plot is closely related to dc/dx (also known as the probability density function). Histograms are valued because they immediately show which bins are highly populated, i.e., what values of x occur frequently. 12 The reason too-small errors are found when fitting to cumulative fraction functions may be found in Reference 2, Press, et al. Bubble Chamber 123 Report Checklist 1. Write an introductory paragraph describing the basic physics behind this experiment. For example, why did different decays of pions result in different L ? Why did different decays of muons result in different R? What are the relationships between L , R and more usual variables of motion? (This manual has many pages on these topics; your job is condense this into a few sentences and no equations.) 2. Book values for muon ( ) and charged pion ( ) mass. Find (and cite) a source that reports at least five significant digits. 3. Derivations: (a) Equation 5.13 (b) Equation 5.30 (c) Area of a spherical cap (d) Use above to derive c(L ) ^ 4. Read: Let's call the last two digits of your CSB/SJU ID number: XY. (a) Go to the web site: and select the file: 4 X.dat. This file contains 4 random data points. By hand draw in your notebook the cumulative fraction step-curve for this dataset along with the percentile values with errors. The result should be similar to Figure 5.4. Show your work! (b) From the same web site, select the file: 1000 Y.dat (or 1000 Yw.dat with the same data in multiple columns). Select reasonable bin values and by hand draw the histogram for this dataset including error bars. Show your work! Copy & paste the 1000 data points into the web13 site and produce a hardcopy percentile plot. Comment on the relationship between features in your histogram and features in the percentile plot. (Note: I'm requesting individual work--rather than partnered work--to check that everyone understands these plots. Do feel free to help your partner succeed, but don't just do the work for him/her.) 5. Results and conclusions from use of the program bubbleCAL: Was systematic error detected? If so, how much. If not, how large could the systematic have been and still go undetected? (This is called an upper limit on the error.) 6. Two values (with errors) for m : one based on R and the other on a fit to Equation 5.40. Hardcopies of the fit results and a plot of the best-fit curve with your data points. (The plot is analogous to Figure 5.6.) 7. Two values (with errors) for m : one based on L and the other on a fit to the equation for c(L ) you derived in 3(d) above. Hardcopies of the fit results and a ^ plot of the best-fit curve with your data points. 13 paste form.html 124 Bubble Chamber 8. A comparison of the values you reported above (5 & 6) and the "known" values cited above in 1. (Make a nice table.) 9. Typically in this lab, the error estimates based on fits that are much smaller than the errors based on averages. Are these error estimates accurate? (If your masses (with errors) are inconsistent with book values, it is quite likely that something is wrong with your error estimates.) Explain how your error estimates could be improved. (For example, sometimes error estimates can be improved (and reduced) by simply taking more data. In other situations, improved errors come from proper consideration of systematic errors and result in larger estimates of error.) A few words about statistical and systematic errors are probably required; you might consult Chapter 0 if nothing occurs to you. References 1. Lederman, Leon & Teresi, Dick The God Particle (1993)-- In this general-audience book Nobel Laureate Lederman describes his work as an experimental particle physicist. Of particular interest: his work with pions in a cloud chamber using the Nevis accelerator (pp. 219224); building cloud and bubble chambers (pp. 244247); parity violation detected in the e decay sequence (pp. 256273). 2. Press, Teukolsky, Vetterling & Flannery Numerical Recipes (1992)-- Title says it all. Cumulative fraction is discussed in the section on the Kolmogorov-Smirnov Test; covariance matrix in chapter on modeling data. 3. Hogg & Tanis Probability and Statistical Inference (1993)-- Order statistics are discussed in Chapter 10. 4. Taylor, J.R. An Introduction to Error Analysis (1997)-- derivative rule error propagation, standard deviation of the mean, histograms, random and systematic errors 5. I know of no book that derives Equation 5.40. A starting point is the Kurie Plot for weak decays, for example, from Halzan & Martin Quarks and Leptons (1984) p. 263: G2 d = m2 E 2 dE 12 3 3- 4E m (5.51) 6. Enge, H.A. Introduction to Nuclear Physics (1966)-- Chapter 7 has a nice discussion on stopping power, range, the Bethe-Block equation and particle detectors including bubble chambers. 7. Coughlan, G. D. & Dodd, J. E. The Ideas of Particle Physics: An Introduction for Scientists (2003)-- An excellent introduction to particle physics at undergraduate physics major level. 8. This lab is based on "The Pi-Mu-e Experiment" 33-3908 (1966) by Ealing Corporation. 6: Poster Presentation Poster Topics: Presentation of a lab project as a poster is the final component of this course. I will be assigning you a topic; feel free to drop me email with a list of your preferences! (First come, first served. Choose a topic ASAP and avoid the crunch at finals!) Topics: 1. Photometry: TB , TR 2. Thermionic Emission: Stefan-Boltzmann Law 3. Thermionic Emission: Richardson-Dushman Law 4. Thermionic Emission: Child-Langmuir Law 5. Bubble Chamber: m 6. Bubble Chamber: m 7. Langmuir Probe: Te 8. Langmuir Probe: ne 9. Langmuir Probe: Vf , Vp Poster Basics: size: fonts: maximum width: 6 feet; maximum height: 4 feet title: 3045 mm (160 pt) authors and affiliations: 2530 mm (96 pt) main headings: 10 mm (48 pt) subheadings: 8 mm (32 pt) text: 5 mm (24 pt) Title, Author(s) with affiliation, Abstract, Introduction, Methods (not required for standard techniques; a block diagram of the apparatus is often helpful), Theory, Results and/or Discussion (if possible, use figures to display results; never display raw data), Conclusions, References, Acknowledgments 125 sections: 126 first steps: Poster Presentation Assemble your figures. Make sure that they convey your main points and can be understood from a distance. For each main section of your poster, try to boil the essential points down to about 350 words (which will, in the proper font, approximately fill an 8 1 " 11" sheet of paper). These section pages 2 can then be assembled along with the figures to form a rough version of your poster. See the references below for book-length instructions on creating an effective poster. a 12 minute `talk' that explains how you collected you data and why the data you are displaying behaves as it does. (If your data displays `unexpected' behavior, explain what was expected and how your data differs from the expected.) This talk is to serve as an introduction to your experiment for folks who are too lazy to read your poster or who have questions about your work. Focus on the main points as poster space is limited. Eliminate everything but the essentials. There should be a clear path through the poster: typically left to right and top to bottom. Avoid unusual fonts and unnecessary font changes. Size fonts for viewing at a distance. (This is particularly important for visual elements like plot axes labels.) The focus of this course has been systematic error, so discussion of the systematic error in your project should be part of your poster. Posters are visual, so your figures must convey your main messages. prepare: hints: warning: Two years ago Dr. Gearhart flustered many students with the question: "What is the basic physics behind your experiment?" Inability to explain simply, in a few sentences (and no equations), what is going on in your experiment puts your entire work in jeopardy as listeners figure if you don't understand the basics, you cannot be trusted on the details. You really must be fully prepared for such questions. Feel free to practice your explanation with me. Here, for example, are 200 words about Child's law: Child's law states that the current through the vacuum tube is proportional to the voltage drop across the tube to the three-halves power. The current here consists of electrons in the `vacuum' moving from cathode to anode because of the electric field caused by the voltage difference. It makes perfect sense that larger voltages (and hence larger forces on the electrons) result in faster electrons and hence more current. The fact that the power is exactly 1.5 is a result from dimensional analysis. The physics behind that dimensional analysis involves current conservation (that the same current must pass through any cylinder between cathode and anode) and `space charge': the charge density (due to the presence of electrons) in the `vacuum' modifies the electric field (which can be calculated via Poisson's equation). Current conservation means Poster Presentation that the electron density is lowest where the electrons are moving fastest, so the highest space charge is near the cathode (i.e., before the electrons have moved very far `downhill'). Child's Law is limited by the maximum possible rate of electron evaporation from the cathode: a temperature dependent effect known as Richardson's Law. 127 References 1. Briscoe, Mary Helen Preparing Scientific Illustrations 1996, Springer-Verlag, Q222.B75 2. Gosling, Peter J. Scientist's Guide to Poster Presentations 1999, Kluwer 3. 4. 5. 6. 7. 128 Poster Presentation 7: Langmuir's Probe Purpose The purpose of this lab is to measure some basic properties of plasmas: electron temperature, number density and plasma potential. Introduction When you think of electrical conductors, you probably think first of metals. In metals the valence electrons are not bound to particular nuclei, rather they are free to move through the solid and thus conduct an electrical current. However, by far the most common electrical conductors in the Universe are plasmas: a term first applied to hot ionized gases by Irving Langmuir (see below). In conditions rare on the surface of the Earth but common in the Universe as a whole, "high" temperatures1 provide enough energy to eject electrons from atoms. Thus a plasma consists of a gas of freely flying electrons, ions, and yet unionized atoms. It should come as no surprise that during the extraordinary conditions of the Big Bang, all the matter in the Universe was ionized. About 380,000 years after the Big Bang, the Universe cooled enough for the first atoms to form. Surprisingly about 400 million years after that the Universe was re-ionized, and the vast majority of the matter in the universe remains ionized today (13.7 billion years after the Big Bang). Some of this ionized matter is at high density (hydrogen gas more dense than lead) and extremely high temperature at the center of stars, but most of it is believed to be at extremely low density in the vast spaces between the galaxies. Perhaps the most obvious characteristic of conductors (metals and plasmas) is that they are shiny; that is, they reflect light. A moment's thought should convince you that this is What does "high temperature" mean? When you are attempting to make a Bose condensation at less than a millionth of a degree, liquid helium at 4 K would be called hot. When you are calculating conditions a few minutes after the Big Bang, a temperature of a billion degrees Kelvin would be called cool. An important insight: Nothing that has units can be said to be big or small! Things that have units need to be compared to a "normal state" before they can be declared big or small. Here the normal state refers to conditions allowing normal solids and liquids to exist. Tungsten, which is commonly used in the filaments of light bulbs, melts at about 3700 K; Carbon does a bit better: 3800 K. The "surface" of the Sun at 6000 K has no solid bits. At temperatures of about 5000 K most molecules have decomposed into elements which in turn have partially "ionized": ejecting one or more electrons to leave a positively charged core (an ion) and free electrons. I'll pick 5000 K as my comparison point for "hot", but of course some elements (e.g., sodium) begin to ionize a lower temperatures, and others (e.g., helium) ionize at higher temperatures. The key factor determining the ionized fraction in the Saha equation is the "first ionization energy". 1 129 130 Langmuir's Probe not an "all-or-nothing" property. While metals reflect radio waves (see satellite TV dishes), microwaves (see the inside of your microwave) and visible light, they do not reflect higher frequency light like X-rays (lead, not aluminum, for X-ray protection). The free electron number density (units: number of free electrons/m3 ), ne , determines the behavior of light in a plasma. (Almost always plasmas are electrically neutral; i.e., the net electric charge density is near zero. If the atoms are all singly ionized, we can conclude that the ion number density, ni , equals ne . In this lab we will be dealing with partially ionized argon with a neutral atom density nn ne ni .) The free electron number density determines the plasma frequency, p : p = 2fp = ne e2 0 m (7.1) where -e is the charge on an electron and m is the mass of an electron. If light with frequency f is aimed at a plasma: the light is reflected, if f < fp ; the light is transmitted, if f > fp . Thus conductors are shiny only to light at frequencies below the plasma frequency. In order to reflect visible light, the plasma must be quite dense. Thus metals (ne 1028 m-3 ) look shiny, whereas semiconductors (ne 1024 m-3 ) do not. The plasma at "surface" of the Sun (with ionized fraction less than 0.1% and ne 1020 m-3 ) would also not look shiny. You will find the plasma used in this lab has even lower ne ; it will look transparent. The defining characteristic of conductors (metals and plasmas) is that they can conduct an electric current. Since conductors conduct, they are usually at a nearly constant potential (voltage). (If potential differences exist, the resulting electric field will automatically direct current flow to erase the charge imbalance giving rise to the potential difference.) At first thought this combination (big current flow with nearly zero potential difference) may sound odd, but it is exactly what small resistance suggests. In fact the detection of big currents (through the magnetic field produced) first lead to the suggestion2 of a conductor surrounding the Earth--an ionosphere. Edward Appleton (1924) confirmed the existence and location of this plasma layer by bouncing radio waves (supplied by a B.B.C. transmitter) off of it. In effect the ionosphere was the first object detected by radar. Appleton's early work with radar is credited with allowing development of radar in England just in time for the 1941 Battle of Britain. Appleton received the Nobel prize for his discovery of the ionosphere in 1947. (Much the same work was completed in this country just a bit later by Breit & Tuve.) The plasma frequency in the ionosphere is around 310 MHz (corresponding to ne 1011 1012 m-3 ). Thus AM radio (at 1 MHz) can bounce to great distances, whereas CB radio (at 27 MHz) and FM (at 100 MHz) are limited to line-of-sight. (And of course when you look straight up you don't see yourself reflected in the ionospheric mirror, as flight 5 1014 Hz. On the other hand, extra terrestrials might listen to FM and TV, but we don't have to worry about them listening to AM radio.) The actual situation is a bit more complex. In the lowest layer of the ionosphere (D region), the fractional ionization is so low that AM radio is more absorbed than reflected. Sunlight powers the creation of new ions in the ionosphere, so when the Sun does down, ionization stops but recombination continues. In neutral-oxygen-rich plasmas like the D region, the plasma disappears without sunlight. Higher up in the ionosphere (the F region, where ne is higher and nn lower) total 2 Faraday (1832), Gauss (1839), Kelvin (1860) all had ideas along this line, but the hypothesis is usually identified with the Scot Balfour Stewart (1882). Langmuir's Probe 131 recombination takes much more than 12 hours, so the plasma persists through the night. Thus AM radio gets big bounces only at night. I have located a plasma (a hot ionized gas) high in the Earth's atmosphere, yet folks climbing Mt. Everest think it's cold high up in the Earth's atmosphere. First, the ionosphere starts roughly 10 times higher than Mt. Everest, and in the F Region (about 200 km up) "temperature" is approaching 1000 K, warm by human standards if not by plasma standards. But the electrons are hotter still. . . up to three times hotter (still not quite hot by plasma standards). This is an odd thought: in an intimate mixture of electrons, ions, and neutral atoms, each has a different temperature. As you know, in a gas at equilibrium the particles (of mass M ) have a particular distribution of speeds (derived by Maxwell and Boltzmann) in which the average translational kinetic energy, EK is determined by the absolute temperature T : 3 1 (7.2) EK = M v 2 = kT 2 2 where k is the Boltzmann constant and denotes the average value. Thus, in a mixture of electrons (mass m) and ions (mass Mi ) at different temperatures (say, Te > Ti ), you would typically find the electrons with more kinetic energy than the ions. (Note that even if Te = Ti , the electrons would typically be moving much faster than the ions, as: 1 2 m ve 2 ve rms = = = 2 ve 1 2 Mi vi 2 Mi vi m (7.3) rms (7.4) 40 1827 that is the root-mean-square (rms) speed of the electrons will be Mi /m 270 times the rms speed of the Argon ions, in the case of an 40 Ar plasma). How can it be that the collisions between electrons and ions fail to average out the kinetic energy and hence the temperature? Consider a hard (in-line) elastic (energy conserving) collision between a slow moving (we'll assume stopped, ui = 0) ion and a speeding electron (vi ). initial vi ui = 0 vf final uf We can find the final velocities (vf & uf ) by applying conservation of momentum and energy. The quadratic equation that is energy conservation is equivalent to the linear equation of reversal of relative velocity: vi = uf - vf (7.5) (7.6) mvi = mvf + M uf with solution: uf = 2m vi (7.7) m+M You can think of the ion velocity as being built up from a sequence of these random blows. Usually these collisions would be glancing, but as a maximum case, we can think of each 132 Langmuir's Probe blow as adding a velocity of u = 2mvi /(m + M ) in a random direction. Consider the ion velocity vector before and after one of these successive blows: uf u2 f = ui + u = u2 i + (u) + 2 ui u 2 (7.8) (7.9) Now on average the dot product term will be zero (i.e., no special direction for u), so on average the speed-squared increases by (u)2 at each collision. Starting from rest, after N collisions we have: u2 f 1 M u2 f 2 = N (u)2 1 = N M (u)2 2 2 2m 1 vi = N M 2 m+M M 1 4m 2 mvi = N m+M m+M 2 (7.10) (7.11) (7.12) (7.13) Thus for argon, N 18, 000 hard collisions are required for the ion kinetic energy to build up to the electron kinetic energy. Note that in nearly equal mass collisions (e.g., between an argon ion and an argon atom), nearly 100% of the kinetic energy may be transferred in one collision. Thus ions and neutral atoms are in close thermal contact; and electrons are in close contact with each other. But there is only a slow energy transfer between electrons and ions/atoms. In photoionization, electrons receive most of the photon's extra energy as kinetic energy. Slow energy transfer from the fast moving electrons heats the ions/atoms. When the Sun goes down, the electrons cool to nearly the ion temperature. Note that the hottest thing near you now is the glow-discharge plasma inside a fluorescent bulb: Te > 3 104 K. . . hotter than the surface of the Sun, much hotter than the tungsten filament of an incandescent light bulb. The cool surface of the bulb gives testimony to the low plasma density (ne 1016 1017 m-3 ) inside the bulb. Containing this hot but rarefied electron gas heats the tube hardly at all -- when the plasma's heat gets distributed over hugely more particles in the glass envelope, you have a hugely reduced average energy, and hence temperature. Plasma People Irving Langmuir (18811957) Born in Brooklyn, New York, Langmuir earned a B.S. (1903) in metallurgical engineering from Columbia University. As was common in those days, he went to Europe for advanced training and completed a Ph.D. (1906) in physical chemistry under Nobel laureate Walther Nernst at University of Gttingen in Germany. Langmuir returned to this country, taking o the job of a college chemistry teacher at Stevens Institute in Hoboken, New Jersey. Dissatisfied with teaching, he tried industrial research at the recently opened General Electric Research Laboratory3 in Schenectady, New York. Langmuir's work for G.E. involved the G.E. calls this lab, which opened in 1900, the "first U.S. industrial laboratory devoted to research, innovation and technology", but Edison's Menlo Park "invention factory" (1876) would often claim that honor. Bell Labs (founded 1925), with six Nobel prizes in physics, would probably claim to be the world's preeminent industrial research lab, but the break up of the "Ma Bell" monopoly has also reduced Bell Labs. 3 Langmuir's Probe 133 then fledgling4 electric power industry. He begin with improving the performance of incandescent electric light bulb. (Langmuir is in the inventors hall of fame for patent number 1,180,159: the modern gas-filled tungsten-filament incandescent electric light.) His work with hot filaments naturally led to thermionic emission and improvements in the vacuum triode tube that had been invented by Lee de Forest5 in 1906. Working with glow discharge tubes (think of a neon sign), he invented diagnostic tools like the Langmuir probe to investigate the resulting "plasma" (a word he coined). "Langmuir waves" were discovered in the plasma. Along the way he invented the mercury diffusion pump. In high vacuum, thin films can be adsorbed and studied. As he said in his 1932 Nobel prize lecture: When I first began to work in 1909 in an industrial research laboratory, I found that the high-vacuum technique which had been developed in incandescent lamp factories, especially after the introduction of the tungsten filament lamp, was far more advanced than that which had been used in university laboratories. This new technique seemed to open up wonderful opportunities for the study of chemical reactions on surfaces and of the physical properties of surfaces under well-defined conditions. In 1946, Langmuir developed the idea of cloud seeding, which brought him into contact with meteorologist Bernard Vonnegut, brother of my favorite author Kurt Vonnegut. That's how Langmuir became the model for Dr. Felix Hoenikker, creator of "ice-nine" in the novel Cat's Cradle. In fact Langmuir created the ice-nine idea (a super-stable form of solid water, with resulting high melting point, never seen in nature for want of a seed crystal) for H.G. Wells who was visiting the G.E. lab. Lyman Spitzer, Jr (19141997) Lyman Spitzer was born in Toledo, Ohio, and completed his B.A. in physics from Yale in 1935. For a year he studied with Sir Arthur Eddington at Cambridge, but that did not work out so he returned to this country and entered Princeton. He completed his Ph.D. in 1938 under Henry Norris Russell, the dean of American astrophysics. Following war work on sonar, he returned to astrophysics. His 1946 proposal for a large space telescope earned him the title "Father of the Hubble Space Telescope". Because of the bend in The Curve of Blinding Energy6 , the lowest energy nuclei are of middle mass (e.g., 56 Fe). Thus nuclear energy can be released by breaking apart big nuclei (like 235 Ur and 239 Pu): fission as in the mis-named atomic bomb or by combining small nuclei (like 2 H deuterium and 3 H tritium): fusion as in the hydrogen bomb. In 1943 Edward Teller and Stanislaw Ulam started theoretical work on bombs based on thermonuclear fusion then called "super" bombs. The end of WWII slowed all bomb work, but the explosion of the first Russian atomic bomb, "Joe 1", in 1949, rekindled U.S. bomb work. This history of renewed interest in H-bomb work is mixed up with Russian espionage--real and imagined, Although not germane to my topic, I can't resist mentioning the famous AC (with Tesla and Westinghouse) vs. DC (with Edison, J.P. Morgan, and G.E.) power wars just before the turn of the century. The battle had a variety of bizarre twists, e.g., each side inventing an electric chair based on the opposite power source aiming to prove that the opponent's source was more dangerous than theirs. Easy voltage transformation with AC guaranteed the victory we see today in every electrical outlet worldwide. Unfortunately the AC frequency did not get standardized so its 60 Hz here and 50 Hz in Europe. 5 18731961; "Father of Radio", born Council Bluffs, Iowa, B.S & Ph.D. in engineering from Yale 6 Title of an interesting book by John McPhee; ISBN: 0374515980; UF767 .M215 1974 4 134 Langmuir's Probe "McCarthyism", and the removal of Robert Oppenheimer's7 security clearance in 1954.8 Our piece of the fusion story starts with Spitzer's 1951 visit with John Wheeler9 in Los Alamos at just the right moment. The building of the Super, in response to Joe 1, had been failing: difficult calculations showed model designs would not ignite. Energy lost by thermal radiation and expansion cooled the "bomb" faster than nuclear energy was released.10 But just before Spitzer arrived, Ulam and Teller had come up with a new method (radiation coupling for fuel compression) that Oppenheimer called "technically sweet". Meanwhile, in a story Hollywood would love, Argentine president Juan Pern announced that his proteg, o e Ronald Richter an Austrian-German chemist, working in a secret island laboratory had achieved controlled fusion. The story (of course) fell apart in less than a year, but it got both Spitzer and funding agencies interested. Spitzer's idea (the "Stellarator") was a magnetically confined plasma that would slowly fuse hydrogen to helium, releasing nuclear energy to heat steam and turn electrical generators. Spitzer and Wheeler hatched a plan to return to Princeton with a bit of both projects (of course funded by the government11 ): Project Matterhorn B would work on bombs and Project Matterhorn S would work on stellarators. Matterhorn B made calculations for the thermonuclear stage of the test shot Mike (1 Nov 1952)--the first H-bomb. The device worked even better than they had calculated. From 1951 until 1958 stellarator research was classified. Optimistic projections12 for fusion reactors were believed by all--after all physicists had completed key projects (atomic bomb, radar, H-bomb) like clockwork. Why should controlled fusion be much different from the carefully calculated fusion of an H-bomb? Early hints of fusion success (neutron emission) turned out to be signs of failure: "instabilities" or disturbances that grew uncontrollably in the plasma. Turbulence--the intractable problem in hydrodynamics13 from the 19th century--came back to bite physics in the 1950s. Instead of the hoped-for "quiescent" plasma, experiment found large amplitude waves: a thrashing plasma that easily escaped the magnetic field configurations physicists had thought should confine it. In 1961 Spitzer turned directorship of the Princeton Plasma Physics Laboratory (PPPL) over to Melvin Gottlieb, and largely returned to astrophysical plasmas. 7 J. Robert Oppenheimer (19041967): born New York, NY, Ph.D. (1927) Gttingen. Directed atomic o bomb work at Los Alamos during WWII; `father of the atomic bomb'. 8 See: Dark Sun, by Richard Rhodes, UG1282.A8 R46 1995 9 John Archibald Wheeler (1911-2008): born Jacksonville, FL, B.S.+Ph.D. (1933) Johns Hopkins, Feynman's major professor at Princeton in 1942. Famous book: Gravitation. Coined the word "black hole". 10 Curve of Binding Energy p. 64: One day, at a meeting of people who were working on the problem of the fusion bomb, George Gamow placed a ball of cotton next to a piece of wood. He soaked the cotton with lighter fuel. He struck a match and ignited the cotton. It flashed and burned, a little fireball. The flame failed completely to ignite the wood which looked just as it had before--unscorched, unaffected. Gamow passed it around. It was petrified wood. He said, "That is where we are just now in the development of the hydrogen bomb." 11 The U.S. Atomic Energy Commission (AEC) initiated the program for magnetic fusion research under the name Project Sherwood. In 1974 the AEC was disbanded and replaced by the Energy Research and Development Administration (ERDA). In 1977, ERDA in turn was disbanded and its responsibilities transferred to the new Department of Energy (DOE). Since 1977, DOE has managed the magnetic fusion research program. 12 An August 1954 report on a theoretical Model D Stellarator (only Model B, with a 2" tube, had actually been built), using assumptions that proved false, projected a power output approximately four times that of Hoover Dam. The usual joke is that controlled fusion will always be just ten years away. 13 Turbulence in hydrodynamics is one of the Clay Millennium Prize Problems (essentially Nobel + Hilbert for mathematics in this century): 1 million dollars for a solution! Langmuir's Probe 135 The current focus for magnetically confined plasma research is the "tokamak": a particular donut-shape (torus) configuration that confines the plasma in part using a large current flowing through the plasma itself. Designed by Russians Igor Tamm and Andrei Sakharov, the T-3 Tokamak surprised the plasma physics world when results were made public in 1968. In 1969, PPPL quickly converted the C-Stellarator into the ST tokamak. Returning to astrophysics, Spitzer's influential books demonstrate his connection to plasma physics: Physics of Fully Ionized Gases (1956 & 1962), Diffuse Matter in Space (1968), Physical Processes in the Interstellar Medium (1978), and Dynamical Evolution of Globular Clusters (1988). Summary Almost everywhere in time and space, plasmas predominate. While present in some natural phenomena at the surface of the Earth (e.g., lightning), plasmas were "discovered" in glow discharges. In the first half of the 1900s, plasma physics was honored with two Nobels14 Langmuir worked to understand "industrial" plasmas in things that are now considered mundane like fluorescent lights. Appleton located the ionosphere: an astronomical plasma surrounding the Earth. Both Nobels were connected to larger historical events (the rise of radio and radar in time to stop Hitler at the English channel). In the second half of the 1900s, plasma physics was connected unpleasant problems: politics (McCarthyism), espionage, and turbulence. While H-bombs worked as calculated, controlled fusion proved difficult and only slow progress has been achieved. Astrophysical plasmas (for example, around the Sun) have also proved difficult to understand. In this century, "industrial" plasmas are again newsworthy with plasma etching for computer chip manufacture and plasma display screens for HDTV. Since the calculation of plasma properties has proved so difficult, measurements of plasma properties ("plasma diagnostics") are critical. In all sorts of plasmas (astrophysical, thermonuclear, industrial), the primary plasma diagnostic has been that first developed by Langmuir. The purpose of this lab is to measure basic plasma properties (Te , ne ) using a Langmuir probe. Glow Discharge Tube In a glow (or gas) discharge tube, a large voltage (100 V) accelerates free electrons to speeds sufficient to cause ionization on collision with neutral atoms. The gas in the tube is usually at low pressure (1 torr), so collisions are not so frequent that the electrons fail to reach the speed required for ionization. Making an electrical connection to the plasma is a more complicated process than it might seem: A. The creation of ions requires energetic collisions (say, energy transfer 10 eV). Kinetic 14 An additional Nobel for plasma physics: Hannes Alfvn (1970). Note that Tamm and Sakharov (named e in the context of the Tokamak) also received Nobels, but not for plasma physics work. 136 Langmuir's Probe cathode dark space cathode glow plasma sheath Cathode plasma Anode Faraday dark space glass tube x V -60 V Figure 7.1: When a current flows between the anode (+) and cathode (), the gas in the tube partially ionizes, providing electrons and ions to carry the current. The resulting plasma is at a nearly constant potential. Electric fields (from potential differences) exist mostly at the edge of the plasma, in the plasma sheath. The largest potential drop is near the cathode. The resulting cathode glow is the region of plasma creation. Increased discharge current Ic results in expanded coverage of the cathode by the cathode glow, but not much change in the cathode voltage Vc . Note that if the anode/cathode separation were larger, a positive column of excited gas would be created between the Faraday dark space and the anode. energy for the collision must in turn come from potential differences of > 10 V. However, we've said that conductors (like the plasma) are at approximately constant potential. Thus ion creation must occur at the edge of the plasma. B. It turns out that attempts to impose a potential difference on a plasma fail. Typically potential differences propagate only a short distance, called the Debye length D , into the plasma: 0 kTe (7.14) D = e2 ne Thus we expect the "edge" of the plasma to be just a few D thick. C. The small electron mass (compared to ions), guarantees high electron mobility in the plasma. Thus we expect electrons to carry the bulk of the current in the plasma. But this cannot be true in the immediate vicinity of the cathode. The electrons inside the cold cathode (unlike the heated cathode in thermionic emission) are strongly held-- they will not spontaneously leave the cathode. Thus near the cathode the current must be carried by ions which are attracted to the negatively charged cathode. Once in contact with the cathode an ion can pick up an electron and float away as a neutral atom. Note particularly that there is no such problem with conduction at the anode: the plasma electrons are attracted to the anode and may directly enter the metal to continue the current. Thus we expect the active part of the discharge to be directly adjacent to the cathode. D. If you stick a wire into a plasma, the surface of the wire will be bombarded with Langmuir's Probe 137 electrons, ions, and neutrals. Absent any electric forces, the impact rate per m2 is given by J= 1 1 n v = n 4 4 8kT M (7.15) where n is the number density of the particles and v is their average speed. If the particles follow the Maxwell-Boltzmann speed distribution, the average speed is determined by the temperature T and mass M of the particles. (Recall: vrms = 3/8 v 1.085 v for a Maxwell-Boltzmann distribution.) Since the electron mass is much less than the ion mass (m Mi ) and in this experiment the temperatures are also different (Te Ti ), the average electron speed is much greater than the average ion speed. Thus an item placed in a plasma will collect many more electrons than ions, and hence will end up with a negative charge. The over-collection of electrons will stop only when the growing negative charge (repulsive to electrons, attractive to ions) reduces the electron current and increases the ion current so that a balance is reached and zero net current flows to the wire. The resulting potential is called the floating potential, Vf . The upshot of these considerations is that objects immersed in a plasma do not actually contact the plasma. Instead the plasma produces a "sheath", a few Debye lengths thick, that prevents direct contact with the plasma. We begin by demonstrating the above equations. The starting point for both equations is the Boltzmann factor: probability = N e-E/kT (7.16) which reports the probability of finding a state of energy E in a system with temperature T , where N is a normalizing factor that is determined by the requirement that the total probability adds up to 1. The energy of an electron, the sum of kinetic and potential energy, is 1 (7.17) E = mv 2 - eV 2 where V is the voltage at the electron's position. (See that for an electron, a low energy region corresponds to a high voltage region. Thus Boltzmann's equation reports that electrons are most likely found in high voltage regions.) To find N add up the probability for all possible velocities and positions: + + + 1 = N = N = N dvx - + - 2 dvy - dvz + 2 dV exp 1 - 2 mv 2 + eV (r) kT + - ...