This book presents a thorough and selfcontained introduction to modern optics, covering in full the three componentsr
571 99 8MB
English Pages 640 [653] Year 2008
Table of contents :
Contents......Page 14
1.1 Aims and contents......Page 24
1.2 Electromagnetic waves......Page 29
1.3 The velocity of light......Page 31
1.4 A sketch of electromagnetic wave theory......Page 33
1.4.1 More general waveforms......Page 37
1.5 The electromagnetic spectrum......Page 39
1.5.1 Visible spectra......Page 41
1.6 Absorption and dispersion......Page 42
1.7 Radiation terminology......Page 45
1.8 Black body radiation......Page 49
1.9 Doppler shift......Page 50
2.1 Light rays and Huygens’ principle......Page 54
2.1.1 The laws of reflection......Page 55
2.1.2 Snell’s law of refraction......Page 56
2.1.3 Fermat’s principle......Page 57
2.1.4 Simple imaging......Page 58
2.1.5 Deviation of light by a triangular prism......Page 59
2.2 Total internal reflection......Page 60
2.2.1 Constant deviation prism......Page 61
2.2.3 Corner cube reflector......Page 62
2.2.4 Pulfrich refractometer......Page 63
2.3 Optical fibre......Page 64
3.1 Introduction......Page 68
3.1.1 Cartesian sign convention......Page 69
3.2 Spherical mirrors......Page 70
3.2.1 Ray tracing for mirrors......Page 72
3.3 Refraction at a spherical interface......Page 73
3.4 Thin lens equation......Page 74
3.4.1 Ray tracing for lenses......Page 76
3.6 Matrix methods for paraxial optics......Page 78
3.7 Aberrations......Page 81
3.7.1 Monochromatic aberrations......Page 82
3.7.2 Spherical aberration......Page 84
3.7.3 Coma......Page 85
3.7.4 Astigmatism......Page 86
3.7.5 Field curvature......Page 87
3.7.6 Distortion......Page 89
3.7.7 Chromatic aberration......Page 90
3.8 Further reading......Page 93
4.2 The refracting telescope......Page 96
4.2.1 Field of view......Page 98
4.2.2 Etendue......Page 100
4.3 Telescope objectives and eyepieces......Page 102
4.4 The microscope......Page 103
4.5 Cameras......Page 106
4.5.1 Camera lens design......Page 108
4.5.2 SLR camera features......Page 109
4.5.3 Telecentric lenses......Page 111
4.5.5 Zoomlenses......Page 112
4.6 Graded index lenses......Page 114
4.8 Fresnel lenses......Page 115
5.1 Introduction......Page 118
5.3 Young’s two slit experiment......Page 119
5.3.1 Fresnel’s analysis......Page 120
5.3.2 Interference by amplitude division......Page 124
5.4 Michelson’s interferometer......Page 126
5.4.1 The constancy of c......Page 127
5.5 Coherence and wavepackets......Page 128
5.5.1 The frequency content of wavepackets......Page 132
5.5.2 Optical beats......Page 133
5.5.3 Coherence area......Page 135
5.7 Interferometry......Page 136
5.7.1 The Twyman–Green interferometer......Page 137
5.7.2 The Fizeau interferometer......Page 138
5.7.4 The Sagnac interferometer......Page 139
5.8 Standing waves......Page 141
5.9 The Fabry–Perot interferometer......Page 144
6.1 Introduction......Page 152
6.3 Single slit Fraunhofer diffraction......Page 153
6.4 Diffraction at a rectangular aperture......Page 156
6.5 Diffraction from multiple identical slits......Page 157
6.6 Babinet’s principle......Page 160
6.7 Fraunhofer diffraction at a circular hole......Page 161
6.8 Diffraction gratings......Page 162
6.9 Spectrometers and spectroscopes......Page 164
6.9.1 Grating structure......Page 165
6.9.2 Etendue......Page 166
6.9.3 Czerny–Turner spectrometer......Page 167
6.9.5 Echelle grating......Page 168
6.10 Fresnel and Fraunhofer diffraction......Page 169
6.11 Single slit Fresnel diffraction......Page 170
6.12 Fresnel diffraction at screens with circular symmetry......Page 174
6.12.1 Zone plates......Page 176
6.13 Microprocessor lithography......Page 177
6.14 Near field diffraction......Page 178
6.15 Gaussian beams......Page 179
6.15.1 Matrix methods......Page 183
7.2 Fourier analysis......Page 186
7.2.1 Diffraction and convolution......Page 189
7.3 Coherence and correlations......Page 192
7.3.1 Power spectra......Page 194
7.3.2 Fourier transform spectrometry......Page 199
7.4 Image formation and spatial transforms......Page 204
7.5 Spatial filtering......Page 205
7.6 Acoustooptic Bragg gratings......Page 207
7.7 Holography......Page 209
7.7.1 Principles of holography......Page 211
7.7.2 Hologram preparation......Page 212
7.7.3 Motion and vibration analysis......Page 213
7.7.4 Thick holograms......Page 214
7.8.1 The 4f architecture......Page 215
7.8.2 Data storage and retrieval......Page 217
8.1 Introduction......Page 220
8.2 Telescope design......Page 221
8.2.1 Auxiliary equipment......Page 225
8.3 Schmidt camera......Page 226
8.4 Atmospheric turbulence......Page 227
8.5 Adaptive optics......Page 228
8.6 Michelson’s stellar interferometer......Page 231
8.7 Modern interferometers......Page 233
8.8 Aperture synthesis......Page 235
8.9 Aperture arrays......Page 239
8.10 Image recovery......Page 240
8.11 Comparisons with radioastronomy......Page 242
8.12 Gravitational wave detectors......Page 243
8.12.1 Lasercavity locking......Page 245
8.12.2 Noise sources......Page 247
8.13 Gravitational imaging......Page 248
9.1 Introduction......Page 252
9.2 Maxwell’s equations......Page 253
9.3 The wave equation......Page 257
9.3.1 Energy storage and energy flow......Page 259
9.4 Electromagnetic radiation......Page 262
9.5 Reflection and refraction......Page 264
9.6 Fresnel’s equations......Page 268
9.7.1 Analysis of multiple layers......Page 272
9.7.2 Beam splitters......Page 277
9.8 Modes of the electromagnetic field......Page 278
9.8.1 Mode counting......Page 280
9.9 Planar waveguides......Page 282
9.9.1 The prism coupler......Page 284
10.2 States of polarization......Page 288
10.3 Dichroism and Malus’ law......Page 291
10.4 Birefringence......Page 292
10.4.1 Analysis of birefringence......Page 294
10.4.2 The index ellipsoid......Page 296
10.4.3 Energy flow and rays......Page 297
10.4.4 Huygens’ construction......Page 298
10.5 Wave plates......Page 299
10.5.1 Jones vectors and matrices......Page 300
10.5.2 Prism separators......Page 301
10.5.3 Polarizing beam splitters and DVD readers......Page 303
10.6 Optical activity......Page 304
10.7 Effects of applied electromagnetic fields......Page 305
10.7.1 Pockels effect and modulators......Page 306
10.7.2 Kerr effect......Page 309
10.7.3 Faraday effect......Page 310
10.8 Liquid crystals......Page 311
10.8.1 The twisted nematic LCD......Page 312
10.8.2 Inplane switching......Page 314
10.8.3 Polymer dispersed liquid crystals (PDLC)......Page 316
10.8.4 Ferroelectric liquid crystals (FELC)......Page 317
10.9 Further reading......Page 318
11.1 Introduction......Page 320
11.2 Rayleigh scattering......Page 321
11.2.1 Coherent scattering......Page 323
11.3 Mie scattering......Page 324
11.4 Absorption......Page 326
11.5 Dispersion and absorption......Page 327
11.5.1 The atomic oscillator model......Page 328
11.6 Absorption by, and reflection off metals......Page 332
11.6.1 Plasmas in metals......Page 337
11.6.2 Group and signal velocity......Page 339
11.6.3 Surface plasma waves......Page 342
11.7 Further reading......Page 345
12.1 Introduction......Page 348
12.2 The black body spectrum......Page 349
12.3 The photoelectric effect......Page 353
12.4 The Compton effect......Page 356
12.5 de Broglie’s hypothesis......Page 358
12.6 The Bohr model of the atom......Page 359
12.6.1 Beyond hydrogen......Page 363
12.7 Wave–particle duality......Page 364
12.8 The uncertainty principle......Page 367
12.9 Which path information......Page 370
12.10.1 Etendue......Page 372
12.11 Afterword......Page 373
12.12 Further reading......Page 374
13.1 Introduction......Page 376
13.2 An outline of quantum mechanics......Page 377
13.3 Schroedinger’s equation......Page 378
13.3.1 The square potential well......Page 380
13.4 Eigenstates......Page 382
13.4.1 Orthogonality of eigenstates......Page 383
13.5 Expectation values......Page 384
13.5.2 Compatible, or simultaneous observables......Page 386
13.6 The harmonic oscillator potential......Page 388
13.7 The hydrogen atom......Page 389
13.8 The Stern–Gerlach experiment......Page 393
13.9 Electron spin......Page 394
13.10 Multielectron atoms......Page 396
13.10.2 Atoms in constant fields......Page 398
13.11 Photon momentum and spin......Page 399
13.12 Quantumstatistics......Page 402
13.13 Line widths and decay rates......Page 403
13.14 Further reading......Page 404
14.1 Introduction......Page 406
14.2 The Einstein coefficients......Page 407
14.3 Prerequisites for lasing......Page 409
14.4 The He:Ne laser......Page 411
14.4.2 Gain......Page 413
14.4.3 Cavity modes......Page 416
14.4.4 Hole burning......Page 418
14.4.6 Optical beats......Page 420
14.6 Organic dye lasers......Page 421
14.6.1 Saturation spectroscopy......Page 423
14.6.2 Cavity ringdown spectroscopy......Page 425
14.6.3 A heterodyne laser interferometer......Page 427
14.7 Introducing semiconductors......Page 431
14.7.1 DH lasers......Page 435
14.7.2 DFB lasers......Page 437
14.7.3 Limiting line widths......Page 438
14.8 Quantumwell lasers......Page 439
14.8.1 Vertical cavity lasers......Page 441
14.9 Nd:YAG and Nd:glass lasers......Page 443
14.9.1 Q switching......Page 444
14.10 Ti:sapphire lasers......Page 446
14.11 Optical Kerr effect and mode locking......Page 448
14.11.1 Mode locking......Page 449
14.12 Frequency combs......Page 451
14.12.1 Optical frequency measurement......Page 453
14.13 Extreme energies......Page 454
14.14 Second order nonlinear effects......Page 455
14.14.1 Raman scattering......Page 459
14.14.2 Brillouin scattering......Page 460
14.14.3 Stimulated Raman and Brillouin scattering......Page 461
14.15 Further reading......Page 462
15.1 Introduction......Page 464
15.2 Photoconductors......Page 465
15.3 Photodiodes......Page 468
15.3.1 Dark current......Page 471
15.4 Photodiode response......Page 472
15.4.1 Speed of response......Page 474
15.4.2 Noise......Page 475
15.4.3 Amplifiers......Page 477
15.4.4 Solar cells......Page 479
15.5 Avalanche photodiodes......Page 480
15.6 Schottky photodiodes......Page 482
15.7 Imaging arrays......Page 484
15.7.1 Quantum efficiency and colour......Page 486
15.7.2 CCD readout......Page 487
15.7.4 CMOS arrays......Page 489
15.8 Photomultipliers......Page 490
15.8.1 Counting and timing......Page 492
15.9 Microchannel plates and image intensifiers......Page 494
15.10 Further reading......Page 495
16.1 Introduction......Page 498
16.2 Attenuation in optical fibre......Page 500
16.3 Guided waves......Page 501
16.4 Fibre types and dispersion properties......Page 505
16.5 Signalling......Page 509
16.6 Sources and detectors......Page 511
16.7 Connectors and routing devices......Page 513
16.7.1 Directional couplers......Page 514
16.7.3 MMI devices......Page 515
16.8 Link noise and power budget......Page 516
16.9 Long haul links......Page 519
16.9.1 Fibre amplifiers......Page 520
16.9.2 Dispersion compensation......Page 522
16.10.1 Thin film filters and Bragg gratings......Page 523
16.10.2 Array waveguide gratings......Page 525
16.10.3 MEMS......Page 528
16.11 Solitons......Page 529
16.11.1 Communication using solitons......Page 531
16.12.1 Fibre Bragg sensors......Page 532
16.12.2 The fibre optic gyroscope......Page 533
16.13 Optical current transformer......Page 536
16.14 Photonic crystal fibres......Page 538
16.15 Further reading......Page 540
17.1 Introduction......Page 542
17.2 Transition rates......Page 544
17.2.1 Selection rules......Page 549
17.2.2 Electric susceptibility......Page 551
17.3 Rabi oscillations......Page 553
17.4 Dressed states......Page 556
17.5 Electromagnetically induced transparency......Page 561
17.5.1 Slow light......Page 565
17.6 Trapping and cooling ions......Page 567
17.7 Shelving......Page 571
17.8 Optical clocks......Page 573
17.9 Further reading......Page 575
18.1 Introduction......Page 578
18.2 Second quantization......Page 579
18.2.1 Continuous variables......Page 585
18.3 First order coherence......Page 586
18.4 Second order coherence......Page 588
18.5 Laser light and thermal light......Page 590
18.5.1 Coherent (laserlike) states of the electromagnetic field......Page 591
18.5.2 Thermal light......Page 593
18.6 Observations of photon correlations......Page 596
18.6.1 Stellar correlation interferometer......Page 597
18.7 Entangled states......Page 598
18.7.2 Spontaneous parametric down conversion......Page 599
18.8 The HOM interferometer......Page 602
18.9 Franson–Chiao interferometry......Page 604
18.10 Complementarity......Page 606
18.10.1 Delayed choice and quantum erasure......Page 607
18.11 Further reading......Page 609
A: Physical constants and parameters......Page 612
B: Appendix: Cardinal points and planes of lens systems......Page 614
C: Appendix: Kirchhoff’s analysis of wave propagation at apertures......Page 616
D: Appendix: The nonlinear Schroedinger equation......Page 620
E: Appendix: State vectors......Page 624
F: Appendix: Representations......Page 628
G: Appendix: Fermi’s golden rule......Page 630
H: Appendix: Solutions......Page 632
C......Page 648
F......Page 649
L......Page 650
P......Page 651
T......Page 652
Z......Page 653
T H E L I G H T FA N TA S T I C
This page intentionally left blank
The Light Fantastic A Modern Introduction to Classical and Quantum Optics I. R. Kenyon School of Physics and Astronomy University of Birmingham
1
3
Great Clarendon Street, Oxford OX2 6DP Oxford University Press is a department of the University of Oxford. It furthers the University’s objective of excellence in research, scholarship, and education by publishing worldwide in Oxford New York Auckland Cape Town Dar es Salaam Hong Kong Karachi Kuala Lumpur Madrid Melbourne Mexico City Nairobi New Delhi Shanghai Taipei Toronto With ofﬁces in Argentina Austria Brazil Chile Czech Republic France Greece Guatemala Hungary Italy Japan Poland Portugal Singapore South Korea Switzerland Thailand Turkey Ukraine Vietnam Oxford is a registered trade mark of Oxford University Press in the UK and in certain other countries Published in the United States by Oxford University Press Inc., New York © Oxford University Press 2008 The moral rights of the author have been asserted Database right Oxford University Press (maker) First published 2008 All rights reserved. No part of this publication may be reproduced, stored in a retrieval system, or transmitted, in any form or by any means, without the prior permission in writing of Oxford University Press, or as expressly permitted by law, or under terms agreed with the appropriate reprographics rights organization. Enquiries concerning reproduction outside the scope of the above should be sent to the Rights Department, Oxford University Press, at the address above You must not circulate this book in any other binding or cover and you must impose the same condition on any acquirer British Library Cataloguing in Publication Data Data available Library of Congress Cataloging in Publication Data Data available Printed in Great Britain on acidfree paper by Antony Rowe Ltd., Chippenham ISBN 978–0–19–856645–8 (Hbk) ISBN 978–0–19–856646–5 (Pbk) 10 9 8 7 6 5 4 3 2 1
Preface This book deals primarily with the properties and uses of electromagnetic waves and photons of visible light. Other regions of the electromagnetic spectrum are only treated where appropriate: for example there is coverage of optical ﬁbre communication using near infrared radiation. Modern quantum theory originated in the observations of the quantum behaviour of electromagnetic radiation, and now, a century later the quantum behaviour of light oﬀers tantalizing possibilities for computing and encryption. During that period a deeper understanding of electromagnetic radiation in terms of waves and photons has made possible the invention of lasers, optical ﬁbre communication, spacebased telescopes, the world wide web and digital cameras. The optoelectronics industry, undreamt of even forty years ago, has grown to be a major employer of scientists and engineers. Even crude measures, such as the hundred million solid state lasers made annually, the millions of kilometres of optical ﬁbre installed, and the widespread availability of megapixel digital cameras and of DVDs give a sense of this industry’s economic and cultural impact. Studies of the subtle features of quantum theory, such as entangled states, have been facilitated by research tools dependent on the technological advances in optoelectronics, which illustrates the truism that technology and pure science go forward handinhand. Clearly there is a necessity for a wide range of scientists and technologists to possess an uptodate understanding of waves and photons so that they can make use of the theoretical, experimental and technological tools now available. The main objective of this text is to provide that basic understanding, which will be important if the reader is to follow future developments in this rapidly expanding ﬁeld. The text is designed to be comprehensive and uptodate so that students at universities and colleges of technology should ﬁnd this volume useful throughout their degree programme. Following an introductory chapter in which basic concepts and facts are presented, the book is divided into three sections: the ﬁrst section (Chapters 2–4) covers ray optics, the second section (Chapters 5–11) wave optics, and the ﬁnal section (Chapters 12–18) quantum optics. Huygen’s principle is used to derive laws of propagation at interfaces in Chapter 2. On this basis the geometric optics of mirrors and lenses
vi Preface
is treated in Chapter 3. Then the principles and design of optical instruments including microscopes, telescopes and cameras are outlined in Chapter 4. Aberrations and the simpler techniques for reducing them to tolerable levels are also described in Chapters 3 and 4. The section on wave optics starts in Chapter 5 with the superposition rule for electromagnetic waves and its application to interference eﬀects such as those seen in Young’s crucial two slit experiment and the Michelson interferometer. Coherence and the relation to atomic wavepackets are both introduced in these simple examples. Diﬀraction eﬀects are considered in Chapter 6. Fourier transforms, of which diﬀraction patterns are an example, are treated formally in Chapter 7. This allows the connection between Michelson interferograms and the source spectrum to be exploited in extracting spectra with standard infrared Fourier transform (FTIR) spectrometers. Chapter 8 pulls together themes in optical instrument design in describing the design of optical mirror telescopes and radio telescopes, and goes on to compare their performance. Electromagnetic wave theory rests on Maxwell’s equations and Poynting’s theorem for the energy in electromagnetic waves. The electromagnetic wave equation and the laws of propogation of light at interfaces (Fresnel’s laws) are derived directly from classical electromagnetic theory in Chapter 9. The use of evanescent waves in optical ﬁbres and other applications are described. Chapter 10 carries the description of polarization forward to include circular polarization which is revealingly the polarization state of individual photons. Electromagnetic interactions with matter in semiclassical terms are discussed in Chapter 11: dispersion, absorption and scattering are described and shown to be related. This completes two sections devoted to the purely classical behaviour of light. An account of the fundamental experiments that underpin the quantum theory of electromagnetic radiation in Chapter 12 opens the section on quantum optics. In Chapter 13 the dual wave–particle nature of electromagnetic radiation and the Heisenberg uncertainty principle are examined at length. The principles underlying laser operation, as well as gas, solid state and semiconductor lasers, and their applications are treated in Chapter 14. Detectors of radiation in the visible and near infrared are described in Chapter 15: these include the CCD and CMOS arrays used in digital cameras. Optical ﬁbre based communication principles, devices and systems, as well as optical ﬁbre sensors are described in Chapter 16. Chapter 17 introduces the semiclassical calculation of decay rates and the behaviour of atoms in the resonant and near resonant laser beams. Eﬀects including electromagnetically induced transparency and slow light are introduced. After this the developments leading to the fabrication of optical clocks are described. Chapter 18 starts by introducing the formal treatment of electromagnetic ﬁelds as quantum mechanical operators (second quantization). This is followed by a description of the study of correlations between photons, ﬁrst observed by Hanbury Brown and Twiss. Then the theory and experimental methods
vii
for generating entangled photons are described; experimental studies of two photon correlations in interferometers showing delayed choice and quantum erasure close the chapter. The text has been designed so that subsets of chapters are selfcontained and well suited to accompany focused optics courses, while the complete text provides compact coverage for courses that extend through three or four years. Chapters 1–4 cover geometric optics; Chapters 5–11 cover classical wave optics; and Chapters 12–18 cover quantum optics, including individual chapters on lasers and modern detectors. A suggested reduced course could include all the chapters and sections listed here: • • • • • • • • • • • •
Introduction and ray optics: Chapters 1 and 2; Lenses without abberations: Sections 3.1 to 3.6.1; Optical instruments: Sections 4.1 to 4.5.2, and 4.8; Wave optics and interferometers: Sections 5.1 to 5.7.1, and 5.8 to 5.9; Diﬀraction and gratings: Sections 6.1 to 6.9; Astronomical telescopes: Sections 8.1 to 8.3; Electromagnetic theory and Fresnel’s laws: Sections 9.1 and 9.4 to 9.6, and 9.8 to 9.8.1; Polarization phenomena: Sections 10.1 to 10.4, and 10.5, and 10.5.2 to 10.7.1, and 10.8 to 10.8.3; Light in matter: Sections 11.1 to 11.6.2; Quantum behaviour of light: Chapter 12; Sections 13.1 to 13.5.2, and 13.11 to 13.13; Lasers and detectors: Sections 14.1 to 14.4, and 14.4.3 to 14.6; Sections 15.1 to 15.3.1, and 15.7 to 15.9; Optical ﬁbre communication: Sections 16.1 to 16.2, and 16.4 to 16.6, and 16.9 to 16.10.1, and 16.13 to 16.14.
This page intentionally left blank
Acknowledgements I would like to thank two Heads of the School of Physics and Astronomy at Birmingham University, Professors John Nelson and Mike Gunn, and also Professor Peter Watkins, Head of the Elementary Particle Physics Group at Birmingham, for their support and encouragement during the more than four years’ preparation of this textbook. Dr Sonke Adlung, the senior science editor at Oxford University Press, has always been unfailingly helpful and courteous in dealing with the many aspects of the preparation, and my thanks go to him for making my path easier. I am also grateful to Jonathan Rubery, Chloe Plummer and Lynsey Livingston at Oxford University Press for the smooth management of copy editing, layout, production and publicity. Many colleagues have been more than generous in ﬁnding time in busy lives to read and comment on material for which they have a particular interest and expertise. First I want to thank Professor Ken Strain of the Institute for Gravitational Research, Glasgow University and the GEO600 team for reading and commenting on the material on gravitational wave detection and Professor Peter Tuthill of the Astronomy Department, University of Sydney, for reading and commenting on the sections concerning aperture synthesis with telescopes. I am indebted to Professor Chris Haniﬀ of the Astrophysics Group at the Cavendish Laboratory, Cambridge University who read through the material on modern interferometry with telescopes and aperture synthesis, and made extensive valuable comments. Also I wish to thank Professor Helen Gleeson of the University of Manchester for reading and commenting on the polarization chapter, particularly the section on liquid crystals. Dr Peter Norreys, Group Leader at the Central Laser Facility, Rutherford Appleton Laboratory, helped me by checking the material relating to extreme energy lasers. I am indebted to Dr Peter Pool of EEV CCD Sensors who patiently answered my many questions about CCD stucture and readout. Ian Bennion, Professor of Optoelectronics at the University of Aston, was kind enough to look over the material on optical ﬁbres and made some very useful suggestions for improvement; I extend my thanks to him. I am particularly grateful to Lene Hau, Mallinckrodt Professor of Physics and Applied Physics at Harvard University, who made comments on the sections concerning electromagnetically induced transparency, and to Professor David Wineland of the Time and Fre
x Acknowledgements
quency Division of the National Institute for Standards and Technology, Boulder for looking over the section on optical clocks. I am extremely grateful to both Ulf Leonhardt, Professor in Theoretical Physics at the University St Andrews University, and Rodney Loudon, Research Professor at the University of Essex, who guided me around several diﬃculties in the theory of the quantized electromagmetic ﬁeld, and to Professor Loudon for a very careful reading the ﬁnal chapter. Turning to my Birmingham colleagues, I want to ﬁrst thank Dr John Griﬃth whose allround, enthusiastic knowledge of optics was always freely available over the years before I started work on this book. He also read and commented on an very early draft of the classical optics component of the book. Next I wish to thank Professor Mike Gunn, who took on the task of reading the description of semiclassical interactions between radiation and atoms. His expert advice and comments were very valuable. I especially thank Dr Ken Elliott for producing the spectra of helium gas, nitrogen gas, a ﬂuorescent lamp and the Sun which appear in ﬁgures 1.8–1.15. My thanks also go to Professor Ted Forgan who was kind enough to read the chapter dealing with scattering, absorption and dispersion: his guidance and suggestions on particular points were very helpful. I wish to thank Professor Yvonne Elsworth for taking the time to read and comment on the chapter presenting Fourier optics, and her comments. I would like to thank Dr. Somak Raychaudry for reading and commenting on the chapter relating to astronomical telescopes. Dr Alastair Rae read and made many useful comments on the content of the chapters which introduce quantum concepts: Alastair’s advice was particularly helpful in balancing the material. Dr Garry Tungate was brave enough to read the long laser chapter and shared with me the fruits of his experience with lasers, for which I am very much indebted to him. Thanks too to Dr Chris Eyles who took on the reading and commenting on the chapter dealing with detectors; he, as well as Dr Peter Pool of EEV CCD Sensors, patiently answered many questions about CCD stucture and readout. Thanks too to Dr Ray Jones for comments on the early version of the ﬁrst two chapters. I am grateful to Dr David Evans for carefully reading through the chapter on classical electromagnetism, and also that on scattering, absorption and dispersion, and to Dr Goronwy Jones for reading through and making lively and useful comments on the chapter presenting the initial development of quantum mechanics. I am grateful to Dr. Costas Mylonas for patiently answering a string of questions on the analysis of electromagnetically induced transparency. Dr. Chris Mayhew kindly oﬀered to read the whole of the ﬁnal draft and I want to thank him for his useful and helpful comments. In addition I would like to heartily thank the reviewer, Dr. Geoﬀrey Brooker of Oxford University, for his thorough and detailed report. His incisive and enlightening comments proved extremely beneﬁcial. The input from all these colleagues removed misunderstandings on
xi
my part, helped me to clarify arguments and brought points to my attention that I would otherwise have missed. The responsibility for any remaining errors should be laid at my door. My thanks also go to the authors and publishers who have allowed me to use published ﬁgures, or adaptations of ﬁgures, or tables, each acknowledged individually in the text. I am grateful to Taylor and Francis, owners of the publications of the CRC Publishing Company regarding Table 12.1 and ﬁgure 11.5; the American Physical Society regarding ﬁgures 12.4, 12.7, 12.12, 13.16, 14.12, 16.30, 18.10, 18.12 and 18.14; the Astronomical Society of the Paciﬁc regarding ﬁgure 8.20; Elsevier Publishing regarding ﬁgure 18.1; Pearson Education Ltd. regarding ﬁgure 15.1; Schott Glas regarding ﬁgure 3.33; and the Institution of Engineering and Technology regarding ﬁgure 16.1. In addition thanks to Dr L. Gardner, program manager at Lambda Research Corporation, USA, for R supplying a lens design analysis from the OSLO system; Dr Bernnd Lingelbach of HTW, Aalen and Institut fuer Augenoptik, Abtsgemuend, Germany, for supplying a ﬁgure of a mediaeval aspheric lens; and Dr Mike Rietveld of the EISCAT Scientiﬁc Association, N9027 Ramfjordmoen, Norway, for supplying a ﬁgure of delay data from the EISCAT radiosonde. In producing the 450 or so diagrams I have made almost exclusive use of the ROOT package developed by Dr. Rene Brun and Dr. Fons Rademakers and described in ROOT – An Object Oriented Data Analysis Framework which appeared in the Proceedings of AIHENP’96 Workshop, Lausanne, Nuclear Instruments and Methods in Physics Research A389(1997)816. ROOT can be accessed at http://root.cern.ch. I would like to thank Dr. Brun for help whilst learning to make use of this sophisticated tool.
To my wife Valerie, without whose support I would not have attempted, let alone completed this book.
Contents 1 Introduction 1.1 Aims and contents . . . . . . . . . . . . 1.2 Electromagnetic waves . . . . . . . . . . 1.3 The velocity of light . . . . . . . . . . . 1.4 A sketch of electromagnetic wave theory 1.4.1 More general waveforms . . . . . 1.5 The electromagnetic spectrum . . . . . . 1.5.1 Visible spectra . . . . . . . . . . 1.6 Absorption and dispersion . . . . . . . . 1.7 Radiation terminology . . . . . . . . . . 1.8 Black body radiation . . . . . . . . . . 1.9 Doppler shift . . . . . . . . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
1 1 6 8 10 14 16 18 19 22 26 27
2 Reﬂection and refraction at plane surfaces 2.1 Light rays and Huygens’ principle . . . . . . . . 2.1.1 The laws of reﬂection . . . . . . . . . . 2.1.2 Snell’s law of refraction . . . . . . . . . 2.1.3 Fermat’s principle . . . . . . . . . . . . 2.1.4 Simple imaging . . . . . . . . . . . . . . 2.1.5 Deviation of light by a triangular prism 2.2 Total internal reﬂection . . . . . . . . . . . . . 2.2.1 Constant deviation prism . . . . . . . . 2.2.2 Porro prisms . . . . . . . . . . . . . . . 2.2.3 Corner cube reﬂector . . . . . . . . . . . 2.2.4 Pulfrich refractometer . . . . . . . . . . 2.3 Optical ﬁbre . . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . .
. . . . . . . . . . . .
. . . . . . . . . . . .
. . . . . . . . . . . .
. . . . . . . . . . . .
. . . . . . . . . . . .
31 31 32 33 34 35 36 37 38 39 39 40 41
3 Spherical mirrors and lenses 3.1 Introduction . . . . . . . . . . . . . 3.1.1 Cartesian sign convention . 3.2 Spherical mirrors . . . . . . . . . . 3.2.1 Ray tracing for mirrors . . 3.3 Refraction at a spherical interface . 3.4 Thin lens equation . . . . . . . . . 3.4.1 Ray tracing for lenses . . . 3.5 Magniﬁers . . . . . . . . . . . . . .
. . . . . . . .
. . . . . . . .
. . . . . . . .
. . . . . . . .
. . . . . . . .
. . . . . . . .
45 45 46 47 49 50 51 53 55
. . . . . . . .
. . . . . . . .
. . . . . . . .
. . . . . . . . . . .
. . . . . . . .
. . . . . . . . . . .
. . . . . . . .
. . . . . . . . . . .
. . . . . . . .
. . . . . . . .
xiv CONTENTS
3.6
. . . . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
55 58 58 59 61 62 63 64 66 67 70
. . . . . . . . . . . . . . .
. . . . . . . . . . . . . . .
. . . . . . . . . . . . . . .
. . . . . . . . . . . . . . .
. . . . . . . . . . . . . . .
. . . . . . . . . . . . . . .
. . . . . . . . . . . . . . .
. . . . . . . . . . . . . . .
. . . . . . . . . . . . . . .
. . . . . . . . . . . . . . .
. . . . . . . . . . . . . . .
. . . . . . . . . . . . . . .
73 73 73 75 77 79 80 83 85 86 88 89 89 91 92 92
5 Interference eﬀects and interferometers 5.1 Introduction . . . . . . . . . . . . . . . . . . . 5.2 The superposition principle . . . . . . . . . . 5.3 Young’s two slit experiment . . . . . . . . . . 5.3.1 Fresnel’s analysis . . . . . . . . . . . . 5.3.2 Interference by amplitude division . . 5.4 Michelson’s interferometer . . . . . . . . . . . 5.4.1 The constancy of c . . . . . . . . . . . 5.5 Coherence and wavepackets . . . . . . . . . . 5.5.1 The frequency content of wavepackets 5.5.2 Optical beats . . . . . . . . . . . . . . 5.5.3 Coherence area . . . . . . . . . . . . 5.6 Stokes’ relations . . . . . . . . . . . . . . . . 5.7 Interferometry . . . . . . . . . . . . . . . . . 5.7.1 The Twyman–Green interferometer . . 5.7.2 The Fizeau interferometer . . . . . . . 5.7.3 The Mach–Zehnder interferometer . . 5.7.4 The Sagnac interferometer . . . . . . 5.8 Standing waves . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . .
95 95 96 96 97 101 103 104 105 109 110 112 113 113 114 115 116 116 118
3.7
3.8
Matrix methods for paraxial optics 3.6.1 The equivalent thin lens . . Aberrations . . . . . . . . . . . . . . 3.7.1 Monochromatic aberrations . 3.7.2 Spherical aberration . . . . . 3.7.3 Coma . . . . . . . . . . . . . 3.7.4 Astigmatism . . . . . . . . . 3.7.5 Field curvature . . . . . . . . 3.7.6 Distortion . . . . . . . . . . . 3.7.7 Chromatic aberration . . . . Further reading . . . . . . . . . . . .
4 Optical instruments 4.1 Introduction . . . . . . . . . . . . . 4.2 The refracting telescope . . . . . . 4.2.1 Field of view . . . . . . . . 4.2.2 Etendue . . . . . . . . . . 4.3 Telescope objectives and eyepieces 4.4 The microscope . . . . . . . . . . 4.5 Cameras . . . . . . . . . . . . . . . 4.5.1 Camera lens design . . . . . 4.5.2 SLR camera features . . . . 4.5.3 Telecentric lenses . . . . . . 4.5.4 Telephoto lenses . . . . . . 4.5.5 Zoom lenses . . . . . . . . 4.6 Graded index lenses . . . . . . . . 4.7 Aspheric lenses . . . . . . . . . . . 4.8 Fresnel lenses . . . . . . . . . . . .
. . . . . . . . . . . . . . .
CONTENTS xv
5.9
The Fabry–Perot interferometer
. . . . . . . . . . . . . . 121
6 Diﬀraction 6.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . 6.2 Huygens–Fresnel analysis . . . . . . . . . . . . . . . 6.3 Single slit Fraunhofer diﬀraction . . . . . . . . . . . 6.4 Diﬀraction at a rectangular aperture . . . . . . . . . 6.5 Diﬀraction from multiple identical slits . . . . . . . 6.6 Babinet’s principle . . . . . . . . . . . . . . . . . . . 6.7 Fraunhofer diﬀraction at a circular hole . . . . . . . 6.8 Diﬀraction gratings . . . . . . . . . . . . . . . . . . . 6.9 Spectrometers and spectroscopes . . . . . . . . . . . 6.9.1 Grating structure . . . . . . . . . . . . . . . . 6.9.2 Etendue . . . . . . . . . . . . . . . . . . . . . 6.9.3 Czerny–Turner spectrometer . . . . . . . . . 6.9.4 Littrow mounting . . . . . . . . . . . . . . . 6.9.5 Echelle grating . . . . . . . . . . . . . . . . . 6.9.6 Automated spectrometers . . . . . . . . . . . 6.10 Fresnel and Fraunhofer diﬀraction . . . . . . . . . . 6.11 Single slit Fresnel diﬀraction . . . . . . . . . . . . . . 6.11.1 Lunar occultation . . . . . . . . . . . . . . . 6.12 Fresnel diﬀraction at screens with circular symmetry 6.12.1 Zone plates . . . . . . . . . . . . . . . . . . . 6.13 Microprocessor lithography . . . . . . . . . . . . . . 6.14 Near ﬁeld diﬀraction . . . . . . . . . . . . . . . . . . 6.15 Gaussian beams . . . . . . . . . . . . . . . . . . . . 6.15.1 Matrix methods . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . . . . . . .
129 129 130 130 133 134 137 138 139 141 142 143 144 145 145 146 146 147 151 151 153 154 155 156 160
7 Fourier optics 7.1 Introduction . . . . . . . . . . . . . . . . 7.2 Fourier analysis . . . . . . . . . . . . . . 7.2.1 Diﬀraction and convolution . . . 7.3 Coherence and correlations . . . . . . . 7.3.1 Power spectra . . . . . . . . . . 7.3.2 Fourier transform spectrometry 7.4 Image formation and spatial transforms 7.5 Spatial ﬁltering . . . . . . . . . . . . . . 7.5.1 Schlieren photography . . . . . . 7.5.2 Apodization . . . . . . . . . . . . 7.6 Acoustooptic Bragg gratings . . . . . . 7.6.1 Microwave spectrum analysis . . 7.7 Holography . . . . . . . . . . . . . . . . 7.7.1 Principles of holography . . . . . 7.7.2 Hologram preparation . . . . . . 7.7.3 Motion and vibration analysis . . 7.7.4 Thick holograms . . . . . . . . . 7.8 Optical information processing . . . . . 7.8.1 The 4f architecture . . . . . . .
. . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . .
163 163 163 166 169 171 176 181 182 184 184 184 186 186 188 189 190 191 192 192
. . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . .
xvi CONTENTS
7.8.2
Data storage and retrieval . . . . . . . . . . . . . . 194
8 Astronomical telescopes 8.1 Introduction . . . . . . . . . . . . . 8.2 Telescope design . . . . . . . . . . 8.2.1 Auxiliary equipment . . . . 8.3 Schmidt camera . . . . . . . . . . . 8.4 Atmospheric turbulence . . . . . . 8.5 Adaptive optics . . . . . . . . . . . 8.6 Michelson’s stellar interferometer 8.7 Modern interferometers . . . . . . 8.8 Aperture synthesis . . . . . . . . . 8.9 Aperture arrays . . . . . . . . . . . 8.10 Image recovery . . . . . . . . . . . 8.11 Comparisons with radioastronomy 8.12 Gravitational wave detectors . . . 8.12.1 Lasercavity locking . . . . 8.12.2 Noise sources . . . . . . . . 8.13 Gravitational imaging . . . . . . .
. . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . .
197 197 198 202 203 204 205 208 210 212 216 217 219 220 222 224 225
9 Classical electromagnetic theory 9.1 Introduction . . . . . . . . . . . . . . . 9.2 Maxwell’s equations . . . . . . . . . . 9.3 The wave equation . . . . . . . . . . . 9.3.1 Energy storage and energy ﬂow 9.4 Electromagnetic radiation . . . . . . . 9.5 Reﬂection and refraction . . . . . . . 9.6 Fresnel’s equations . . . . . . . . . . . 9.7 Interference ﬁlters . . . . . . . . . . . 9.7.1 Analysis of multiple layers . . 9.7.2 Beam splitters . . . . . . . . . 9.8 Modes of the electromagnetic ﬁeld . . 9.8.1 Mode counting . . . . . . . . . 9.9 Planar waveguides . . . . . . . . . . . 9.9.1 The prism coupler . . . . . . .
. . . . . . . . . . . . . .
. . . . . . . . . . . . . .
. . . . . . . . . . . . . .
. . . . . . . . . . . . . .
. . . . . . . . . . . . . .
. . . . . . . . . . . . . .
. . . . . . . . . . . . . .
. . . . . . . . . . . . . .
. . . . . . . . . . . . . .
. . . . . . . . . . . . . .
. . . . . . . . . . . . . .
229 229 230 234 236 239 241 245 249 249 254 255 257 259 261
10 Polarization 10.1 Introduction . . . . . . . . . . . . . 10.2 States of polarization . . . . . . . . 10.3 Dichroism and Malus’ law . . . . . 10.4 Birefringence . . . . . . . . . . . . 10.4.1 Analysis of birefringence . . 10.4.2 The index ellipsoid . . . . 10.4.3 Energy ﬂow and rays . . . . 10.4.4 Huygens’ construction . . . 10.5 Wave plates . . . . . . . . . . . . . 10.5.1 Jones vectors and matrices 10.5.2 Prism separators . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
265 265 265 268 269 271 273 274 275 276 277 278
. . . . . . . . . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
CONTENTS xvii
10.5.3 Polarizing beam splitters and DVD readers 10.6 Optical activity . . . . . . . . . . . . . . . . . . . 10.7 Eﬀects of applied electromagnetic ﬁelds . . . . . . 10.7.1 Pockels eﬀect and modulators . . . . . . . 10.7.2 Kerr eﬀect . . . . . . . . . . . . . . . . . . 10.7.3 Faraday eﬀect . . . . . . . . . . . . . . . . 10.8 Liquid crystals . . . . . . . . . . . . . . . . . . . . 10.8.1 The twisted nematic LCD . . . . . . . . . . 10.8.2 Inplane switching . . . . . . . . . . . . . . 10.8.3 Polymer dispersed liquid crystals (PDLC) 10.8.4 Ferroelectric liquid crystals (FELC) . . . . 10.9 Further reading . . . . . . . . . . . . . . . . . . . . 11 Scattering, absorption and dispersion 11.1 Introduction . . . . . . . . . . . . . . . . 11.2 Rayleigh scattering . . . . . . . . . . . . 11.2.1 Coherent scattering . . . . . . . 11.3 Mie scattering . . . . . . . . . . . . . . . 11.4 Absorption . . . . . . . . . . . . . . . . 11.5 Dispersion and absorption . . . . . . . . 11.5.1 The atomic oscillator model . . 11.6 Absorption by, and reﬂection oﬀ metals 11.6.1 Plasmas in metals . . . . . . . . 11.6.2 Group and signal velocity . . . . 11.6.3 Surface plasma waves . . . . . . 11.7 Further reading . . . . . . . . . . . . . .
. . . . . . . . . . . .
. . . . . . . . . . . .
. . . . . . . . . . . .
. . . . . . . . . . . .
280 281 282 283 286 287 288 289 291 293 294 295
. . . . . . . . . . . .
. . . . . . . . . . . .
. . . . . . . . . . . .
. . . . . . . . . . . .
. . . . . . . . . . . .
. . . . . . . . . . . .
. . . . . . . . . . . .
. . . . . . . . . . . .
. . . . . . . . . . . .
. . . . . . . . . . . .
297 297 298 300 301 303 304 305 309 314 316 319 322
quantum nature of light and matter Introduction . . . . . . . . . . . . . . . . . The black body spectrum . . . . . . . . . The photoelectric eﬀect . . . . . . . . . . The Compton eﬀect . . . . . . . . . . . . de Broglie’s hypothesis . . . . . . . . . . . The Bohr model of the atom . . . . . . . 12.6.1 Beyond hydrogen . . . . . . . . . . 12.6.2 Weaknesses of the Bohr model . . 12.7 Wave–particle duality . . . . . . . . . . . 12.8 The uncertainty principle . . . . . . . . . 12.9 Which path information . . . . . . . . . . 12.10 Wavepackets and modes . . . . . . . . . 12.10.1 Etendue . . . . . . . . . . . . . . 12.11 Afterword . . . . . . . . . . . . . . . . . 12.12 Further reading . . . . . . . . . . . . . .
. . . . . . . . . . . . . . .
. . . . . . . . . . . . . . .
. . . . . . . . . . . . . . .
. . . . . . . . . . . . . . .
. . . . . . . . . . . . . . .
. . . . . . . . . . . . . . .
. . . . . . . . . . . . . . .
. . . . . . . . . . . . . . .
. . . . . . . . . . . . . . .
325 325 326 330 333 335 336 340 341 341 344 347 349 349 350 351
12 The 12.1 12.2 12.3 12.4 12.5 12.6
13 Quantum mechanics and the atom 13.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . 13.2 An outline of quantum mechanics . . . . . . . . . . . . . 13.3 Schroedinger’s equation . . . . . . . . . . . . . . . . . .
353 . 353 . 354 . 355
xviii CONTENTS
13.3.1 The square potential well . . . . . . . . 13.4 Eigenstates . . . . . . . . . . . . . . . . . . . . 13.4.1 Orthogonality of eigenstates . . . . . . . 13.5 Expectation values . . . . . . . . . . . . . . . . 13.5.1 Collapse of the wavefunction . . . . . . 13.5.2 Compatible, or simultaneous observables 13.6 The harmonic oscillator potential . . . . . . . 13.7 The hydrogen atom . . . . . . . . . . . . . . . 13.8 The Stern–Gerlach experiment . . . . . . . . . 13.9 Electron spin . . . . . . . . . . . . . . . . . . . 13.10 Multielectron atoms . . . . . . . . . . . . . . 13.10.1 Resonance ﬂuorescence . . . . . . . . . . 13.10.2 Atoms in constant ﬁelds . . . . . . . . . 13.11 Photon momentum and spin . . . . . . . . . . 13.12 Quantum statistics . . . . . . . . . . . . . . . 13.13 Line widths and decay rates . . . . . . . . . . 13.14 Further reading . . . . . . . . . . . . . . . . . 14 Lasers 14.1 Introduction . . . . . . . . . . . . . . . . . 14.2 The Einstein coeﬃcients . . . . . . . . . 14.3 Prerequisites for lasing . . . . . . . . . . . 14.4 The He:Ne laser . . . . . . . . . . . . . . 14.4.1 Three and four level lasers . . . . . 14.4.2 Gain . . . . . . . . . . . . . . . . 14.4.3 Cavity modes . . . . . . . . . . . 14.4.4 Hole burning . . . . . . . . . . . . 14.4.5 Laser speckles . . . . . . . . . . . . 14.4.6 Optical beats . . . . . . . . . . . . 14.5 The CO2 gas laser . . . . . . . . . . . . . 14.6 Organic dye lasers . . . . . . . . . . . . . 14.6.1 Saturation spectroscopy . . . . . . 14.6.2 Cavity ringdown spectroscopy . . 14.6.3 A heterodyne laser interferometer 14.7 Introducing semiconductors . . . . . . . . 14.7.1 DH lasers . . . . . . . . . . . . . . 14.7.2 DFB lasers . . . . . . . . . . . . . 14.7.3 Limiting line widths . . . . . . . . 14.8 Quantum well lasers . . . . . . . . . . . . 14.8.1 Vertical cavity lasers . . . . . . . . 14.9 Nd:YAG and Nd:glass lasers . . . . . . . . 14.9.1 Q switching . . . . . . . . . . . . . 14.10 Ti:sapphire lasers . . . . . . . . . . . . . 14.11 Optical Kerr eﬀect and mode locking . . 14.11.1 Mode locking . . . . . . . . . . . . 14.12 Frequency combs . . . . . . . . . . . . . . 14.12.1 Optical frequency measurement . 14.13 Extreme energies . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . .
357 359 360 361 363 363 365 366 370 371 373 375 375 376 379 380 381
. . . . . . . . . . . . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . . . . . . . . . . . .
383 383 384 386 388 390 390 393 395 397 397 398 398 400 402 404 408 412 414 415 416 418 420 421 423 425 426 428 430 431
CONTENTS xix
14.14 Second order nonlinear eﬀects . . . . . . . . . . 14.14.1 Raman scattering . . . . . . . . . . . . . . 14.14.2 Brillouin scattering . . . . . . . . . . . . . . 14.14.3 Stimulated Raman and Brillouin scattering. 14.15 Further reading . . . . . . . . . . . . . . . . . . . 15 Detectors 15.1 Introduction . . . . . . . . . . . . . . . . . 15.2 Photoconductors . . . . . . . . . . . . . . 15.3 Photodiodes . . . . . . . . . . . . . . . . . 15.3.1 Dark current . . . . . . . . . . . . 15.4 Photodiode response . . . . . . . . . . . . 15.4.1 Speed of response . . . . . . . . . . 15.4.2 Noise . . . . . . . . . . . . . . . . 15.4.3 Ampliﬁers . . . . . . . . . . . . . . 15.4.4 Solar cells . . . . . . . . . . . . . . 15.5 Avalanche photodiodes . . . . . . . . . . . 15.6 Schottky photodiodes . . . . . . . . . . . 15.7 Imaging arrays . . . . . . . . . . . . . . . 15.7.1 Quantum eﬃciency and colour . . 15.7.2 CCD readout . . . . . . . . . . . . 15.7.3 Noise and dynamic range . . . . . 15.7.4 CMOS arrays . . . . . . . . . . . . 15.8 Photomultipliers . . . . . . . . . . . . . . 15.8.1 Counting and timing . . . . . . . . 15.9 Microchannel plates and image intensiﬁers 15.10 Further reading . . . . . . . . . . . . . .
. . . . .
. . . . .
. . . . .
. . . . .
432 436 437 438 439
. . . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . . .
441 441 442 445 448 449 451 452 454 456 457 459 461 463 464 466 466 467 469 471 472
16 Optical ﬁbres 16.1 Introduction . . . . . . . . . . . . . . . . . . 16.2 Attenuation in optical ﬁbre . . . . . . . . . 16.3 Guided waves . . . . . . . . . . . . . . . . 16.4 Fibre types and dispersion properties . . . 16.5 Signalling . . . . . . . . . . . . . . . . . . . 16.6 Sources and detectors . . . . . . . . . . . . 16.7 Connectors and routing devices . . . . . . . 16.7.1 Directional couplers . . . . . . . . . 16.7.2 Circulators . . . . . . . . . . . . . . 16.7.3 MMI devices . . . . . . . . . . . . . 16.8 Link noise and power budget . . . . . . . . 16.9 Long haul links . . . . . . . . . . . . . . . . 16.9.1 Fibre ampliﬁers . . . . . . . . . . . 16.9.2 Dispersion compensation . . . . . . 16.10 Multiplexing . . . . . . . . . . . . . . . . . 16.10.1 Thin ﬁlm ﬁlters and Bragg gratings 16.10.2 Array waveguide gratings . . . . . . 16.10.3 MEMS . . . . . . . . . . . . . . . . . 16.11 Solitons . . . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . .
. . . . . . . . . . . . . . . . . . .
475 475 477 478 482 486 488 490 491 492 492 493 496 497 499 500 500 502 505 506
xx CONTENTS
16.11.1 Communication using solitons 16.12 Fibre optic sensors . . . . . . . . . . 16.12.1 Fibre Bragg sensors . . . . . 16.12.2 The ﬁbre optic gyroscope . . 16.13 Optical current transformer . . . . . 16.14 Photonic crystal ﬁbres . . . . . . . . 16.15 Further reading . . . . . . . . . . .
. . . . . . .
. . . . . . .
. . . . . . .
. . . . . . .
. . . . . . .
. . . . . . .
. . . . . . .
. . . . . . .
. . . . . . .
. . . . . . .
. . . . . . .
. . . . . . .
508 509 509 510 513 515 517
17 Quantum interactions 17.1 Introduction . . . . . . . . . . . . . . . . . 17.2 Transition rates . . . . . . . . . . . . . . 17.2.1 Selection rules . . . . . . . . . . . 17.2.2 Electric susceptibility . . . . . . . 17.3 Rabi oscillations . . . . . . . . . . . . . . 17.4 Dressed states . . . . . . . . . . . . . . . . 17.5 Electromagnetically induced transparency 17.5.1 Slow light . . . . . . . . . . . . . . 17.6 Trapping and cooling ions . . . . . . . . . 17.7 Shelving . . . . . . . . . . . . . . . . . . . 17.8 Optical clocks . . . . . . . . . . . . . . . . 17.9 Further reading . . . . . . . . . . . . . . .
. . . . . . . . . . . .
. . . . . . . . . . . .
. . . . . . . . . . . .
. . . . . . . . . . . .
. . . . . . . . . . . .
. . . . . . . . . . . .
. . . . . . . . . . . .
. . . . . . . . . . . .
. . . . . . . . . . . .
519 519 521 526 528 530 533 538 542 544 548 550 552
. . . . . .
555 555 556 562 563 565 567
. . . . . . . . . . . .
568 570 573 574 575 576 576 579 581 583 584 586
18 The quantized electromagnetic ﬁeld 18.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . 18.2 Second quantization . . . . . . . . . . . . . . . . . . . . 18.2.1 Continuous variables . . . . . . . . . . . . . . . 18.3 First order coherence . . . . . . . . . . . . . . . . . . . . 18.4 Second order coherence . . . . . . . . . . . . . . . . . . 18.5 Laser light and thermal light . . . . . . . . . . . . . . . 18.5.1 Coherent (laserlike) states of the electromagnetic ﬁeld . . . . . . . . . . . . . . . . . . . . . . . . . 18.5.2 Thermal light . . . . . . . . . . . . . . . . . . . . 18.6 Observations of photon correlations . . . . . . . . . . . . 18.6.1 Stellar correlation interferometer . . . . . . . . . 18.7 Entangled states . . . . . . . . . . . . . . . . . . . . . . 18.7.1 Beam splitters . . . . . . . . . . . . . . . . . . . 18.7.2 Spontaneous parametric down conversion . . . . 18.8 The HOM interferometer . . . . . . . . . . . . . . . . . 18.9 Franson–Chiao interferometry . . . . . . . . . . . . . . . 18.10Complementarity . . . . . . . . . . . . . . . . . . . . . . 18.10.1 Delayed choice and quantum erasure . . . . . . 18.11Further reading . . . . . . . . . . . . . . . . . . . . . . . A Physical constants and parameters
589
B Appendix: Cardinal points and planes of lens systems 591 C Appendix: Kirchhoﬀ ’s analysis of wave propagation at apertures 593
CONTENTS xxi
D Appendix: The nonlinear Schroedinger equation
597
E Appendix: State vectors
601
F Appendix: Representations
605
G Appendix: Fermi’s golden rule
607
H Appendix: Solutions
609
Index
625
This page intentionally left blank
Introduction 1.1
Aims and contents
An understanding of the properties and manipulation of electromagnetic radiation in the visible and near visible part of the spectrum are of great importance in the physical sciences and related branches of engineering. It was, of course, in the study of electromagnetic phenomena that quantum eﬀects were ﬁrst encountered, while today optical experiments continue to reveal subtleties of the quantum behaviour of radiation. On the practical side optical sources, devices and detectors are crucial to both research and industry. A few examples will illustrate their value, starting with applications in pure research: ground and space based telescopes look back into the remote past of the universe; kilometre sized interferometers are being used to search for gravitational waves; while laser cooling enables researchers to bring ions to rest so that their transitions provide the ultimate in stable reference frequencies for optical clocks. In the commercial ﬁeld applications also abound and are of increasing importance: the millions of kilometres of optical ﬁbre now installed provide the modern data highway; semiconductor lasers are used in transmitting this data over ﬁbres, other lasers are used to weld retinas while kilowatt power lasers routinely drill holes in inchthick steel plates. Among consumer goods DVD readers and LCD screens function thanks to our ability to manipulate polarized light. The techniques and interests of fundamental research and industry overlap a good deal. For example charged coupled devices (CCDs) with several million pixels on a single chip are the preferred detectors in astronomy and are also found in digital cameras. A more exotic example is the interest of solitons to both pure and applied research: these are waveforms which can be transmitted over thousands of kilometres of optical ﬁbre without change of shape. Finally entangled states of photons are studied to investigate questions about the logical structure of quantum mechanics, and at the same time these states oﬀer the possibility of a form of secure quantum key distribution for cryptography when exchanging sensitive information. It is to introduce students of science and engineering to this intellectually stimulating and industrially signiﬁcant subject that the text aims. In the later sections of this chapter some background material is provided to set the scene. This includes the following topics: a sketch of electromagnetic wave theory; a review of the electromagnetic spec
2 Introduction
trum; basic properties of optical materials; radiation terminology; and the Doppler shift. Thereafter the text is divided into three parts corresponding to the natural division of optics into ray optics, wave optics and quantum optics. These three parts occupy three, seven and seven chapters respectively. Topics that are most naturally dealt with after the introduction of quantum concepts include lasers, detectors and ﬁbre optics. The emphasis is on visible radiation and the near infrared radiation used in optical ﬁbre communications, with coverage of ultraviolet radiation where this is relevant to the main themes discussed. A brief review of the contents of these chapters now follows. The apertures of lens and mirror systems are usually much larger than the wavelength of light; in such conditions it is an excellent and productive approximation to regard light as travelling in straight lines, rays, between and within the optical components. Huygens’ principle for wavefront construction is made the starting point of ray optics in Chapter 2, and immediately yields the laws of reﬂection and refraction. The ray optics of light incident on plane and spherical mirrors and interfaces is presented in Chapters 2 and 3 and applied to prisms and lenses. Matrix methods of analyzing paraxial ray propogation are included. The results are used in Chapter 4 to explain the principles and design of optical instruments. The function of optical instruments can be compromised by the aberrations of lenses and mirrors, so these aberrations are discussed and the corrective design techniques reviewed in Chapters 3 and 4. The wave optics section of the text begins in Chapter 5 with the discussion of the superposition of electromagnetic waves and resulting interference eﬀects. Young’s twoslit experiment, which provided the key step in the development of wave theory, and the Michelson interferometer illustrate two complementary techniques for observing interference. It will be shown that interference eﬀects between beams only appear when the beams have the same wavelength, polarization and a constant phase relationship. When beams have the same wavelength and a constant phase relation they are said to be coherent, and if they are, or are brought into the same polarization state, they will interfere. The coherence of beams as they are emitted from a source is directly related to the emission process. Chapters 6 contains material on diﬀraction, which is basically interference involving any pattern of apertures, as for example a broad slit or multiple slits. The emphasis in Chapter 6 is on Fraunhofer diﬀraction, for which the source and detector are both far from the aperture array. Diﬀraction produced by circular apertures and by gratings are considered, the latter being the basis for spectrometers. Both diagrammatic (phasor) and analytic techniques for handling interference and diﬀraction are described because the ﬂexibility gained is useful in practice. The usual Gaussian beams of lasers are described and
1.1
matrix methods for paraxial propagation of rays are extended to these beams. Near ﬁeld Fresnel diﬀraction produced by slits and circular apertures/absorbers are discussed, and how Fresnel diﬀraction connects with Fraunhofer diﬀraction. Fraunhofer diﬀraction patterns are examples of Fourier transforms in which the aperture pattern is the input and the diﬀraction pattern is the output. In Chapter 7 this connection is treated formally, and the chapter also contains a description of how the Michelson interferogram can be Fourier analysed to give the source spectrum, which is a standard procedure in Fourier transform infrared (FTIR) spectrometers. Gaussian and Lorentzian distributions, which are the commonly met shapes of spectral lines, are Fourier analysed, and the bandwidth theorem deduced. Optical telescopes such as the Hubble Space Telescope (HST), and telescope arrays are described in Chapter 8. The modern techniques that have extended the usefulness of telescopes, such as adaptive optics, aperture synthesis and interferometry are discussed. The more exotic gravitational wave detectors and gravitational imaging are treated brieﬂy. In Chapters 9 and 10 the classical treatment of radiation is developed from Maxwell’s equations and Poynting’s theorem. It is then possible to calculate the fractions of radiation reﬂected and transmitted at interfaces between media and how these fractions depend on the polarization state of the electromagnetic radiation. It is a surprise to ﬁnd also that in total internal reﬂection an evanescent wave travels along the surface of the second medium; a feature of importance for wave propogation in optical ﬁbres. There follows a description of interference ﬁlters and beam splitters: these widely used devices rely on interference between light reﬂected from multiple layers of dielectrics. Modes of electromagnetic radiation and their propagation along planar waveguides and prism couplers are the ﬁnal topics covered in Chapter 9. Chapter 10 is used to relate plane and circular polarization, and to describe eﬀects such as dichroism, birefringence and the optical activity of materials under applied ﬁelds. Birefringence of uniaxial crystals is analysed using the index ellipsoid. Techniques and materials for polarizing and manipulating polarization are widely used in modern technology: wave plates, Kerr cells, Faraday isolators, and more complex devices including optical modulators, DVD readers and liquid crystal displays are discussed in detail. The wave velocity in a material depends on the radiation’s wavelength and this variation is known as dispersion. A wellknown example is the dispersion of white light into a coloured spectrum by a prism. Dispersion in materials and the related phenomena of absorption and scattering are discussed in Chapter 11 in terms of the underlying atomic and molecular response to radiation. Scattering is strongly dependent on wavelength. The behaviour for wavelengths both short and long (Rayleigh scattering from the sky) compared to the size of scatterer are described. Absorp
Aims and contents 3
4 Introduction
tion and dispersion are coupled and their behaviour in dielectrics when incident radiation induces a resonant response of electrons bound in atoms is analysed. Drude’s model of free electrons is used to account for the simple optical properties of metals and plasmas. The energy and information carried by electromagnetic radiation travel with the group velocity of the wavepackets rather than the wave velocity of the ideal but unattainable inﬁnite plane sinusoidal waves. This diﬀerence is strikingly exhibited using the example of propagation of electromagnetic waves in the ionosphere. Surface plasma waves and one of their uses in medicine are described. The quantum optics section of the book commences in Chapter 12 with a review of the key pieces of experimental evidence from early in the twentieth century that show that electromagnetic radiation possesses particle as well as a wave properties. The chapter covers de Broglie’s insight that material particles, for example electrons, also have a wave– particle nature, and the data that revealed electron diﬀraction. Bohr’s model of an atom with quantized orbits, contemporary to these developments, is used as a ﬁrst step in the approach to the modern quantum model of the atom. The reconciliation of the wave and particle nature of electromagnetic radiation and material particles is presented, and is known as waveparticle duality. The fundamental step was to recognize that the intensity of the classical wave over a region of spacetime is directly proportional to the probability for ﬁnding a quantum of radiation, a photon, within that region. Heisenberg’s uncertainty principle and its implications for the precision of measurements are discussed. Chapter 13 is used to introduce quantum mechanics and its highly successful predictions about atomic structure and spectra. Eigenstates, expectation values, compatible observables and the collapse of the wavefunction are all discussed. Schroedinger’s nonrelativistic equation for material particles is presented and solved for the electron motion in the hydrogen atom. Quantization rules for optical transitions are noted. Intrinsic angular momentum of particles, spin, was introduced by Pauli and the experiments revealing the spins of electrons and photons are described. Finally the relationship between the width of a spectral lines and the lifetime of the decaying state is looked at. There follow three chapters on applications of quantum phenomena in research and industry. Chapter 14 contains descriptions of lasing, lasers types and laser applications. The common helium–neon gas laser is used as the introductory example, and the treatment covers lasing prerequisites, laser modes, the calculation of gain, hole burning and speckles. Dye lasers are described next, the ﬁrst tunable lasers. Then follow applications in interferometry and spectroscopy where the unprecedented narrow width of a laser emission line is of ﬁrst importance. Semiconductor lasers are convenient sources for injecting light into optical ﬁbre because of their compact beams, and compatibility with modern electronics. Several types of semiconductor lasers are described including
1.1
vertical cavity and quantum well lasers. Solid state lasers have a crystal such as Nd:YAG or Ti:sapphire as the active lasing material. They can achieve high powers and their applications in pulsed mode, with pulses as short as femtoseconds, are described. Nonlinear eﬀects in certain crystals produced by high intensity beams are of increasing importance: harmonic generation, parametric ampliﬁers, Raman scattering and Brillouin scattering are all discussed. Then an account of electronic detectors of visible and near infrared radiation is given in Chapter 15. The common detectors are semiconductor photodiodes and photomultipliers, in which a photons are absorbed and electrons freed. An applied electrical potential produces a current that may be ampliﬁed within the device, as in avalanche photodiodes and photomultipliers, or ampliﬁed externally. Ideally the detector current should be proportional to the light intensity over a wide range. Thresholds, eﬃciency, sensitivity, linearity and noise are all discussed. Various types of photodiodes including Schottky and avalanche photodiodes are described. The features of CCD imaging arrays met in cameras are explained. Photomultipliers and the hybrid image intensiﬁers met in night sights are also discussed. With the description of detectors and laser sources completed, optical ﬁbre communications and sensors are discussed in context in Chapter 16. Single mode optical ﬁbre along which only a single optical mode can propagate has especially low dispersion and absorption in the near infrared part of the spectrum. As a result faithful transmission of data at high rates over intercontinental distances on single mode ﬁbre has become commonplace. Mode transmission and dispersion of optical ﬁbre are analysed in depth in Chapter 16. Optical link techniques, devices, and power budget are all discussed. The methods deployed on intercontinental links to regenerate attenuated signals optically and to compensate residual dispersion are highly developed and are treated in some detail. Multiple laser beams are used to multiplex independent data streams on a single ﬁbre: this wavelength division multiplexing is also described. Soliton propagation on optical ﬁbre is now commercially feasible and the underlying physical theory is presented. Optical ﬁbre is increasingly used in a range of sensors. Representative examples are also outlined, including grating sensors, gyros and current transformers. In the last section of the chapter an account of the properties and uses of microstructured optical ﬁbres is given. The remaining two chapters are used to introduce modern topics in quantum optics. Chapter 17 provides an introduction to the semiclassical analysis of the interaction of electromagnetic radiation with atoms. Radiation is treated as waves while the electrons in the atoms are described by Schroedinger’s equation. This has proved to be an adequate approach to explain many simple phenomena. The rate of decay of hydrogen from the excited 2p to the 1s ground state is calculated
Aims and contents 5
6 Introduction
as an example of an electric dipole transition. Selection rules for such transitions, already stated in Chapter 13, are discussed in more detail here. Then the susceptibility of a gas to radiation at frequencies around a transition frequency is calculated. When a gas is exposed to a beam from a laser tuned to a resonance frequency of the atom, the atoms are pumped cyclically between the ground and excited state. Such Rabi oscillations and other related eﬀects are analysed. One curious eﬀect discussed is electromagnetically induced transparency. This occurs when a gas opaque to a laser beam on resonance becomes transparent to this beam when a second laser excites a diﬀerent atomic transition. A further unexpected result is that the ﬁrst laser beam can be slowed dramatically and even, in a sense, brought to rest. The cooling and trapping of ions is described to illustrate the usefulness of laser cooling. Atomic transitions are then Doppler free and oﬀer the ultimate in precise frequency references. The description of an optical clock based on a single mercury ion closes Chapter 17. A ﬁnal step needed in developing a fully consistent quantum theory of matter and radiation is to replace ﬁelds by operators. In the case of radiation the electromagetic ﬁelds become operators, while in the case of material particles the wavefunctions become operators. Chapter 18 is used to present the procedure for this second quantization applied to the electromagnetic ﬁelds. A parallel procedure can be applied to material particles, but will not be attempted here because it would lead too far aﬁeld from the topics of interest in optics. The operators of the electromagnetic ﬁelds create and annihilate individual photons. Correlations and degrees of coherence will be reexpressed in terms of the expectation values of these operators. This approach provides a framework within which to explain many surprising quantum eﬀects. The groundbreaking example in which photon bunching from an incoherent source was seen in the experiments by Hanbury Brown and Twiss is analysed. Photons (and other particles) can be created in entangled states: symbolically a state 1a 2b + 1b 2a describes a pair of photons, 1 and 2, which are entangled in states a and b. If photon1 is measured later and found to be in state a/b, then photon2 is inevitably in state b/a. Experimental methods of creating entangled pairs of photons are described. Interferometers used to measure photon correlations in entangled states have permitted the exploration of many purely quantum mechanical subtleties such as complementarity, delayed choice and quantum erasure. Representative experiments are described in the closing sections of Chapter 18.
1.2
Electromagnetic waves
Electromagnetic waves have properties resembling those of the more familiar mechanical waves. Waves on a stretched string or on the surface of deep water show local movement and yet the string or the water is
1.2
not carried forward with the waves. What is carried forward is energy, which in the case of a waves on the sea is readily apparent to someone in a ship tossed by waves. In the case of electromagnetic waves the electric and magnetic ﬁelds at every point in the wave’s path are the things that oscillate. Energy carried by the the electromagnetic wave is detected in ways that depend on its wavelength. For example as you sit at your desk you might feel the Sun’s rays warming you, or use a radio receiver to convert modulation of radio waves to sound, or receive light waves from this sentence on this page so that you can share my thoughts. The electromagnetic ﬁelds are vector quantities which generally point transverse to the direction of wave motion so that the electric ﬁeld, the magnetic ﬁeld and the direction of motion of the wave are mutually perpendicular. This assertion is exactly true in free space; where free space is the vacuum uncluttered by material objects. This is similar to 1.5
1
E field
0.5
k 0
0.5
B field 1
1.5
0
0.2
0.4
0.6
0.8
1
1.2
1.4
Distance in wavelengths OR time in periods Fig. 1.1 The variation of the electric and magnetic ﬁelds (a) with position along the direction of wave motion for a plane electromagnetic wave with linear polarization; or (b) with time at a ﬁxed point in space.
the transverse displacement of a string when a wave travels along it. Suppose that an electromagnetic wave travels in the zdirection and the electric ﬁeld vector points in the xdirection, then the magnetic ﬁeld vector points in the ydirection. The deﬁnition of the units rests on the measurement of the force exerted on a charge: the force on a charge q
Electromagnetic waves 7
8 Introduction
coulombs moving with velocity v is given by the Lorentz relation F = q [ E + v ∧ B ],
(1.1)
where the force F is in newtons when the electric ﬁeld E is expressed in volts per metre (V m−1 ) and the magnetic ﬁeld B is in teslas (T). Figure 1.1 shows the electric and magnetic ﬁelds in a sinusoidal wave moving along the arrowed path, labelled k. Representative ﬁeld vectors are drawn at three points along the path, the remainder are indicated by shading. The diagram is a snapshot in time, and at a later moment the whole wave will have moved to the right or to the left, depending on its direction of travel. The distance between successive identical features (e.g. peaks) is the wavelength with the symbol λ. Alternatively we can regard ﬁgure 1.1 as showing the time variation of the wave at a ﬁxed point in space. In this case the time between successive identical features is the period with the symbol τ . Then the number of peaks passing a ﬁxed point per unit time is 1/τ , which is called the frequency, measured in hertz (Hz), and given the symbol f . Therefore, with f waves of length λ passing per second, the wave velocity is v = f λ.
Io
Io Jupiter
This is also known as the phase velocity. Electromagnetic waves can have any wavelength, with light being the em waves detectable by the eye with wavelengths between 400 and 700 nm. The electromagnetic waves shown in ﬁgure 1.1 are called plane polarized or linearly polarized because each ﬁeld vector remains in a plane as the wave propogates.
1.3 Romer starts timing eclipses 6 months later Romer is here Fig. 1.2 Rømer’s method of determining the speed of light.
(1.2)
The velocity of light
Compared to the velocity of mechanical waves the velocity of light is extremely large; a distant lightning ﬂash is seen well before the thunderclap is heard. However this simple observation does not allow any inference as to whether light arrives instantaneously, or whether its velocity is large but ﬁnite. Rømer in 1676 was the ﬁrst to deduce that the velocity of light is ﬁnite: he had been timing a sequence of eclipses of the moon Io by its parent planet Jupiter. Jupiter orbits the Sun once every 11.8 years with Io in close attendance, Io orbiting Jupiter once every 1.77 days. The motion of the Earth around the Sun and of Io around Jupiter are so much more rapid than that of Jupiter that we can ignore the motion of Jupiter in following Rømer’s argument. Rømer noted that, over a halfyear during which the Earth–Jupiter separation was continuously increasing, the period of rotation of Io round Jupiter appeared to lengthen also. The eclipses of Io by Jupiter were ﬁnally twentytwo minutes later than was to be expected if Io’s period of revolution around Jupiter was constant. Rømer correctly interpreted this delay as the time taken by light from Io to travel the distance that the Earth had moved away from Io over that half year. This eﬀect is illustrated in ﬁgure 1.2. Later Bradley used another astronomical method to obtain a value for
Late in the nineteenth century Michelson discovered that the velocity of light is independent of the motion of the source and of the observer; his measurement will be discussed in Chapter 5. This result is quite diﬀerent from the way the velocity of, for example, sound waves behaves. If an observer at rest measures the velocity of sound as v m s−1 , then on moving toward the source at a velocity u m s−1 the velocity will appear to rise to (v + u) m s−1 , just as you would guess. For electromagnetic radiation the measured velocity is constant whatever the relative motion of source and observer! This experimental fact, the constancy of the velocity of light, is a fundamental feature of nature. Einstein made the constancy of the velocity of light in free space, whatever the motion of the source or observer, one of the two postulates on which he built the special theory of relativity in 1905. By 1984 the value of the velocity of light in vacuum determined from the product of wavelength and frequency
Apparent di
Bradley’s velocity in December
irection
In the 1960s a new approach to measuring the velocity of light took advantage of the relation given above: velocity equals wavelength times frequency. A source emitting a narrow range of frequencies is used, and both the frequency and wavelength of the radiation in vacuum are measured, and then multiplied together to give c. The velocity of electromagnetic waves in vacuum has been measured very precisely by various methods, and it is found to be a constant independent of the wavelength of the radiation.
Actual starlight direction
d Apparent
the velocity of light close to the modern value, 3 108 m s−1 . Bradley observed that distant stars appeared to change position during the year. Looking north for example, all the stars complete an annual circular orbit of 43 arcseconds diameter. (This is twenty times larger than the displacement observed in a year in the relative positions of the nearest stars with respect to the distant stars due to parallax.) Bradley correctly interpreted this aberration of starlight as being due to the motion of the Earth around the Sun. He argued that the direction which the starlight appears to come from is the vector diﬀerence between the velocity of the starlight and that of the Earth. A knowledge of the Earth’s orbital velocity and the aberration yields the velocity of light. This is shown in ﬁgure 1.3. Such astronomical methods were displaced by more precise Earthbased measurements in which the round trip time is measured for light to travel along a measured path to and from a mirror. If this is done in air then very long paths, of order tens of kilometres, can be used, but a small correction is needed to compensate for the diﬀerence between the velocity of light in vacuum and in air at the atmospheric conditions. The round trip time for a total path of 30 km is 0.1 ms so that a precision of 10−9 s (one nanosecond 1 ns) in timing is required to get a precision of 1 part in 105 in the velocity determination. In order to acheive the necessary precision in timing electronically controlled shutters such as the Kerr cell described in Chapter 10 are used. This techniques is commonly used to measure distances in surveying, in which case the velocity of light is the input and the distance the output.
The velocity of light 9
rection
1.3
Bradley’s velocity in June
Fig. 1.3 Aberration of starlight. The apparent direction of the same star is shown at times six months apart.
10 Introduction
measurements was:
c = 299 792 458m s−1 .
(1.3)
The precision in the determination of c depended on the precision of the reference standards of length and time. In the case of the unit of time this was, and is, provided by atomic clocks based on the frequency of a microwave transition in caesium. Such clocks commonly agree to better than parts in 1012 so that atomic clocks are accepted as the primary standards of time (and frequency). The length of the second is deﬁned to be 9 192 631 770 periods of the radiation emitted in a speciﬁed transition of 133 Cs. At that time the standard of length was deﬁned in terms of a wavelength of krypton but with much poorer precision. Scientists therefore chose to deﬁne, once for all, the velocity of light in vacuum at its then best measured value, given above. This is an altogether reasonable approach because the velocity of light in vacuum is a constant of nature, whereas units of length and time are deﬁnitely not. This leaves the unit of length as something that has to be measured; the metre being the distance travelled by light in vacuum in (9 192 631 770/299 792 458) periods of the 133 Cs microwave transition.
1.4
A sketch of electromagnetic wave theory
A wave equation is the equation of motion for the type of waves considered; sound, em, water waves, etc. In the case of electromagnetic waves the wave equation is obtained from Maxwell’s four equations which encapsulate the properties of the electromagnetic ﬁelds. The resulting wave equation in vacuum is extremely compact and applies to both electric and magnetic ﬁelds. In this section three key formulae from electromagnetic theory will be quoted: the wave equation, the equation for energy ﬂow and the equation for energy storage in electromagnetic waves. A full derivation will be given later in Chapter 8. This introduction is restricted to electromagnetic waves travelling in free space: very conveniently it applies to a good approximation, and in many circumstances, to such waves travelling in air. The simplest em wave is a plane wave travelling in what we can choose to be the zdirection. With such a wave E has at any moment the same value over any plane surface perpendicular to the zaxis, and the same is true of B. Taking E to lie along the xdirection, the wave equation for the electric ﬁeld then reduces to ∂ 2 Ex /∂z 2 = ε0 µ0 ∂ 2 Ex /∂t2 ,
(1.4)
where ε0 is a scalar quantity called the permittivity of free space, and µ0 is another scalar called the permeability of free space. The values of physical constants such as ε0 and µ0 are collected in Appendix A. We
1.4
A sketch of electromagnetic wave theory 11
can try a sinusoidal plane wave solution Ex = E0 cos (2πf t − 2πz/λ + φ),
(1.5)
where φ is a phase factor. The word amplitude is customarily deﬁned to mean E0 , which is the maximum value that Ex takes, but it is also widely used for the instantaneous value of Ex in for example discussions of coherence. It will be clear from the context in which way the word is being used. Diﬀerentiating Ex twice with respect to z and t and inserting the results into the wave equation gives −(4π 2 /λ2 )Ex = −ε0 µ0 (4πf 2 )Ex ,
(1.6)
so that the wave equation is satisﬁed provided that f 2 λ2 = 1/ε0 µ0 .
(1.7)
It was shown above that the velocity of sinusoidal waves equals the product of the frequency and wavelength, so the velocity of electromagnetic waves in free space must be √ c = 1/ ε0 µ0 .
(1.8)
On getting this result Maxwell substituted the then measured values of ε0 and µ0 into this equation in order to predict the wave velocity. He found that the predicted velocity agreed with the measured value of the velocity of light to within the precision with which c, ε0 and µ0 were known. Thus electromagnetic theory leads to a wave equation whose solutions, in free space, are waves travelling at exactly the measured speed of light. Maxwell’s equations impose no restriction on the wavelengths or the frequencies possible; so the inference made was that light is just one form of electromagnetic radiation. Soon afterwards, in 1886, Hertz tested the idea that electromagnetic waves can be generated and detected well beyond the visible spectrum. He generated electromagnetic waves of frequencies ∼100 MHz from a spark gap in an oscillatory circuit, and he observed that an oscillatory current was produced in an identical circuit located several metres away. Early in the twentieth century Marconi successfully transmitted radio signals across the Atlantic; and a century later communication via electromagnetic waves has become allpervasive. The magnetic ﬁeld obtained using Maxwell’s equations is perpendicular to the direction of motion and to the electric ﬁeld: By = Ex /c.
(1.9)
An alternative way of writing the ﬁelds using the angular frequency ω = 2πf and the wave number or propogation constant1 k = 2π/λ is: Ex = E0 cos (ωt − kz + φ),
(1.10)
1
Another deﬁnition used in spectroscopy for the wave number is 1/λ.
12 Introduction
(a)
(b)
Fig. 1.4 The variation of the electric ﬁeld with position along the direction of wave motion for (a) a circularly polarized electromagnetic wave and (b) a plane polarized wave of the same wavelength.
and By = (E0 /c) cos (ωt − kz + φ).
(1.11)
Waves can of course travel in any direction. In the more general case ˆ the wave that the wave travels in a direction given by the unit vector k at a point (r, t) in spacetime is E = E0 cos (ωt − k · r + φ),
(1.12)
ˆ is the wave vector and, as before, k is the wave number. where k = k k The shape of the wavefronts of light emitted from a point source would be spherical, but typical sources are generally in containers which restrict the angular range of the wavefront. In the ideal case of an unobscured sinusoidal wave radiating from a point source at the origin, the electric ﬁeld at a point in space time (r, t) would be E = E0 cos (ωt − kr + φ).
(1.13)
Here again ω, k and φ are respectively the angular frequency, the wave number and the phase factor. As the spherical wavefront travels further
1.4
A sketch of electromagnetic wave theory 13
and further from the source it approximates ever more closely to a plane surface over any ﬁxed area. A second simple type of polarization is that known as circular polarization. In this case the electric ﬁeld of a plane sinusoidal wave travelling in the zdirection is E = E0 [ ex cos (ωt − kz) + ey sin (ωt − kz) ],
(1.14)
where ex and ey are unit vectors along the x and ydirections respectively. This electric ﬁeld rotates with angular frequency ω in the xOy plane. The magnetic ﬁeld rotates too, remaining at right angles to the electric ﬁeld. The contrasting behaviour of the electric ﬁeld in plane and circularly polarized waves is shown in ﬁgure 1.4. Light emitted by most sources is a mix of polarizations, changing from instant to instant. If the plane of polarization is entirely random then the source is said to be unpolarized. Lasers are the exception, in that lasers are generally designed to produce beams with plane polarization. The electromagnetic waves carry energy and in vacuum the instantaneous energy density is U = (ε0 E2 + B2 /µ0 )/2,
(1.15)
where U is in joules/metre3 (J m−3 ). In the case of a sinusoidal plane wave in free space, for which B = E/c, this reduces to U = ε0 E 2 .
(1.16)
The time average of the energy density for a wave of the form given by eqns. 1.10 and 1.11 is obtained by taking the average over one cycle of oscillation τ = 2π/ω, τ U = ε0 E02 cos2 (ωt − kz + φ) dt/τ 0
= ε0 E02 /2.
(1.17)
The ﬂow of energy in an electromagnetic wave is the energy crossing unit area per unit time perpendicular to the wave direction. It is therefore a vector quantity, called the Poynting vector N = E ∧ B/µ0 .
(1.18)
This energy ﬂux or power per unit area, N, is measured in watts/metre2 (W m−2 ), and points in the direction the wave is travelling. Its magnitude is N = E 2 /(µ0 c) = E 2 /Z0 . (1.19) where Z0 = µ0 /0 is called the impedance of free space. The time average of the energy ﬂow for a wave of the form given by eqns. 1.10 and 1.11 is N z = ε0 E02 c/2, (1.20)
14 Introduction
which equals the product of the energy density in the ﬁeld and the wave velocity. Many discussions of light travelling through free space and simple materials use only a single vector ﬁeld rather than the two vector ﬁelds, E and B. This simpliﬁcation is permissible because the magnitudes of the electric and magnetic ﬁelds are proportional, and no information is lost thereby. In analyzing light propagating through optical systems with lenses and mirrors the simpliﬁcation can be taken a step further to use a single scalar ﬁeld when there is no dependence of the devices on polarization. Referring back to the expression for the Lorentz force eqn. 1.1 we see that the ratio of the electric/magnetic force is E/[vB] = E/[v(E/c)] = v/c, for an electromagnetic wave in free space. Evidently the magnetic force can be neglected unless the velocity of the electrons approaches the velocity of light, or if the material has a large magnetic permeability. Therefore in many cases where electromagnetic radiation interacts with matter the magnetic ﬁeld may be neglected. It remains true that both ﬁelds are inextricably involved in carrying and storing energy in electromagnetic radiation.
1.4.1
More general waveforms
In the preceding account attention was focused on the sinusoidal (also called harmonic) wave solutions to the wave equation. A more general travelling wave solution of the wave equation is Ex = E0 F (ct ± z) = E0 F (w)
(1.21)
where F is any function whatever of the combination w = (ct ± x).This statement can be readily checked. Firstly, diﬀerentiating twice with respect to z gives ∂Ex /∂z = E0 (dF/dw)(∂w/∂z) = ±E0 (dF/dw),
(1.22)
∂ 2 Ex /∂z 2 = ±E0 (d2 F/dw2 )(∂w/∂z) = E0 (d2 F/dw2 );
(1.23)
and
then diﬀerentiate Ex twice with respect to t ∂ 2 Ex /∂t2 = E0 (d2 F/dw2 )(∂w/∂z)2 = c2 E0 (d2 F/dw2 ).
(1.24)
Hence ∂ 2 Ex /∂z 2 = (1/c2 )∂ 2 Ex /∂t2
(1.25)
as required. Figure 1.5 shows a wave of arbitrary shape at two instants t and (t + ∆t), where the star indicates a reference point (feature) on the wave which moves from z to (z + ∆z) in a time interval ∆t. The value of F (ct − z) is the therefore the same at these two points in spacetime, namely at (z,t) and (z + ∆z, t + ∆t). Therefore ct − z = c(t + ∆t) − (z + ∆z), and ∆z = +c∆t.
(1.26)
1.4
A sketch of electromagnetic wave theory 15
Wave amplitude
1
0.5
Time t
(a)
0
0.5
1
z Space coordinate
Wave amplitude
1
0.5
Time t+∆t
(b)
0
0.5
1
z+∆z Space coordinate
Fig. 1.5 An electromagnetic wave F (ct − z) at two diﬀerent moments t and (t + ∆t).
This means that the feature and hence the wave is moving rightward. The reader can check that F (ct + z) represents a leftward moving wave. There are many possible waveforms of the type F (ct ± z) and some examples are shown in ﬁgure 1.6: (a) is a square wave, (b) is a wave of irregular shape but still a repetitive wave and (c) is a waveform called a pulse or wavepacket which is not repetitive. The reason we can concentrate on sinusoidal waves in the face of these and countless other possibilities is that whatever the waveform it can always be duplicated exactly by a sum of sinusoidal waves with their amplitudes and phase factors suitably chosen. There is a well deﬁned procedure called Fourier analysis which is used to extract these harmonic components from any waveform. Fourier analysis is treated at length in Chapter 6, while here it will be suﬃcient to take note of this underlying simplicity: if results can be proved for harmonic waves then they must apply equally for any waveform.
16 Introduction
1 (a)
0.5 0 0.5 1 Space coordinate
Wave amplitude
1 (b)
0.5 0 0.5 1 Space coordinate 1
(c)
0.5 0 0.5 1 Space coordinate
Fig. 1.6 Examples of electromagnetic waveforms: (a) a square repetitive wave; (b) an irregular repetitive wave; (c) a wavepacket or pulse.
1.5
The electromagnetic spectrum
Electromagnetic waves in free space can have any wavelength. At one extreme the wavelength of radiation at the mains frequency (EU 50 Hz, US 60 Hz) is huge (6 106 m, 5 106 m); and at the other extreme the gamma rays emitted by a decaying π 0 meson have wavelength 2.9 10−15 m. The way that em waves interact with matter depends on their wavelength and this variation has aﬀected how and when the diﬀerent parts of the electromagnetic spectrum were ﬁrst discovered, how they are named and in what ways the radiation in each part of the spectrum can be used. On ﬁgure 1.7 the principal regions of the spectrum, the primary sources and important uses of electromagnetic radiation are all indicated. Radio and TV stations transmit waves of preassigned wavelengths which lie in the range from about 1 km to a 1 m. Cable TV (CATV) and satellite TV use shorter wavelengths. Electromagnetic waves of very long wavelength (VLF and ELF) are used to communicate with submarines because radiation of shorter wavelengths is strongly absorbed by sea wa
1.5
10 15 10
13
Elementary particle
γ rays
10 21
Nuclear
10 11
Xrays
Inner atomic shell
10 9 10
7
10 3 10
1
10
+1
10
+3
10
+5
10 19 10 17
UV Visible
10 5
10 23
Infrared
Outer atomic shell Fibre optics Molecular
10 15 10 13
Twaves
11
Transitions µwaves CMB Radar µwave oven Cell phone CATV FM Radio AM
10
10
7
VLF
10
5
ELF
Wavelength in metres
10 9
Submarine communication
10 3 Frequency in Hz
Fig. 1.7 The electromagnetic spectrum.
ter. Microwaves are electromagnetic waves with wavelength from about 1 m to 0.1 mm. Radar, microwave transmitters for mobile phones as well as microwave ovens operate in this part of the spectrum. Microwave detectors have been used to observe the spectrum of the relic radiation from the early universe, the cosmic microwave background (CMB), whose intensity peaks at a wavelength around 2 mm. The region from 1 mm wavelength down to the red end of the visible spectrum is called the infrared, and overlaps the microwave region. An electric ﬁre radiates most of its energy in the infrared, while the preferred wavelengths for telecom optical ﬁbre links are in the near infrared (near, that is, to the visible spectrum). At these latter wavelengths the absorption of electromagnetic waves in glass ﬁbre has a broad minimum. The visible part of the spectrum extends from 400 to 700 nm. Shorter wavelength radiation down to around 10 nm is called ultraviolet (UV) radiation. At yet shorter wavelengths the ultraviolet merges into the Xray region, which extends from roughly 10 nm to 1 pm (10−12 m). Excited nuclei emit radiation of very short wavelength from 0.1 nm downward, and these waves are called γ rays. The shortest wavelength γ rays are emitted in the decays of elementary particles such as the π 0 meson. Apart from the visible spectrum, which is deﬁned by the range of wavelengths to which the average eye is sensitive, none of the boundaries mentioned above are at all precise. For the present we should also note that visible radiation interacts very eﬀectively with the atoms whose size is ∼0.1 nm.
The electromagnetic spectrum 17
18 Introduction
Fig. 1.8 Emission spectrum of helium gas. This spectrum and the following emission spectra for nitrogen gas, a ﬂuorescent lamp and the Sun’s spectrum were recorded with a TV SPEC spectrometer with 600 lines per mm, made by Elliott Instruments Ltd., www.elliottinstruments.co.uk. Courtesy Dr K. H. Elliott.
Radiation of wavelength shorter than light can penetrate matter with increasing ease as the wavelength falls. Thus ultraviolet radiation penetrates the skin and can damage cells by ionizing the component DNA, and this penetrating power increases steadily as the wavelength decreases. The longer wavelength UV region is divided into UVA of wavelength from 320 to 400 nm, UVB of wavelength from 280 to 320 nm and UVC of wavelength shorter than 280 nm. UVB radiation damages living tissue at intensities not very much greater than those met on a sunny day in temperate climes. The more dangerous UVC radiation from the Sun is absorbed by oxygen and ozone; but the UVB is only absorbed by ozone, hence the concern about the depletion of the ozone layer in the upper atmosphere. The deep penetrating power of Xrays and γrays makes them useful tools in medical diagnosis and in material science when applied in a controlled manner.
1.5.1
Fig. 1.9 Emission spectrum of helium gas, presented as a histogram of the intensity against wavelength. This spectrum was obtained automatically from the preceding spectrum. Courtesy Dr K. H. Elliott.
Fig. 1.10 Emission spectrum of nitrogen gas, Courtesy Dr K. H. Elliott.
Fig. 1.11 Emission spectrum of nitrogen gas, presented as a histogram of the intensity against wavelength. This spectrum was obtained automatically from the preceding spectrum. Courtesy Dr K. H. Elliott.
Visible spectra
The reader will have seen the spectra produced when sunlight passes through a prism. Light of diﬀerent colours is bent, refracted, through an angle which depends on the colour of the light. This variation of the bending with wavelength is known as dispersion. Light can also be dispersed into a spectrum when it is reﬂected from a grating of parallel, uniformly spaced grooves on a metal surface, the spacing being of order a few micrometres. Such diﬀraction gratings are described in detail in Chapter 6. Figures 1.8 and 1.9 show the same visible spectrum emitted by a gas of excited helium atoms. The source is a slit illuminated by a spectral lamp ﬁlled with pure helium gas, and through which an electric discharge is passed. The spectrum of light dispersed by the diﬀraction grating is shown in ﬁgure 1.8, and its intensity is histogrammed as a function of wavelength in ﬁgure 1.9. If the light were monochromatic, that is to say all of a single wavelength then the spectrum would be a single line in ﬁgure 1.8, at the image of the slit for that particular colour of light. In fact the spectrum contains many lines, each due to light of a particular narrow range of wavelengths. These narrow spectral lines are characteristic of the spectrum of a gas. Notice ﬁrst that this spectrum and those in the following diagrams are truncated by the display software at 300 and 780 nm. Secondly note that for clarity the intensity scales of the histograms are oﬀset by 40 units from zero. The set of spectral lines is unique to the element emitting light in the spectral lamp and we shall see later in Chapter 12 that the spectral lines show regularities in the wavelengths for each element. Such regularities were a puzzle to the physicists who discovered them in the nineteenth century. The regularites arise from the structure of the atoms of each element and can only be understood using quantum theory. The inter
1.6
Absorption and dispersion 19
pretation of spectra, which was a key step in guiding the development of quantum mechanics, will be fully described in Chapters 12 and 13. Figures 1.10 and 1.11 exhibit the emission spectrum of a spectral lamp containing nitrogen. Here the individual spectral lines are grouped into bands which are characteristic of molecular spectra, in this case diatomic nitrogen. The spectrum from a modern ﬂuorescent lamp is more complex. This is shown in ﬁgures 1.12 and 1.13, where the spectral lines are superposed on a continuum. The spectral lines are emitted by the mercury gas within the discharge tube, and are characteristic of that element. Some of the UV radiation emitted by the mercury atoms is absorbed by a powder deposited on the inside wall of the glass envelope. This is a ﬂuorescent material, meaning that its atoms absorb the UV radiation and then reemit visible light with a delay that is less than a microsecond. Now condensed matter, such as the powder used to coat a discharge tube, emits over a range of wavelengths rather than producing line specta. This broadening of the spectral lines into a continuum results from the proximity and strong mutual interaction of atoms in a solid or liquid. By using a ﬂuorescent powder which absorbs the ultraviolet radiation emitted by the mercury atoms and reemits visible light the light yield from the lamp is enhanced considerably. Consequently ﬂuorescent lamps are among the most eﬃcient lamps in converting wall plug power to visible light. The ﬁnal spectrum to be seen in ﬁgures 1.14 and 1.15 is the Sun’s spectrum. This is a continuum marked by dark lines and corresponding notches in the histogram. This continuum is that of a hot body in thermal equilibrium, in this case the outer layers of the Sun at around 6000 K. The dark lines were interpreted by Fraunhofer as due to the absorption of light by cooler gases in the upper reaches of the Sun’s atmosphere. These lines match in wavelength the emission lines of elements found on Earth, and this allows inferences to be made about the gases making up the Sun. The set of Fraunhofer lines due to absorption by helium in the Sun’s outer atmosphere were initially a mystery because at that time helium had not be identiﬁed and was only later isolated on Earth. A comparison of the spectra of the ﬂuorescent lamp and the Sun shows that the modern lamp mimics the shape of the Sun’s spectrum quite well.
1.6
Absorption and dispersion
Electromagnetic waves passing through matter are partly absorbed by the atoms and molecules, and their velocity in matter is less than in vacuum, by a factor called the refractive index. The frequency of the wave is unchanged and hence from eqn. 1.2 it follows that the wavelength in matter is smaller than its value in free space by the value of the refractive index. This point will be made again in Section 9.5. Air
Fig. 1.12 Emission spectrum of a modern ﬂuorescent lamp. Courtesy Dr K. H. Elliott.
Fig. 1.13 Spectrum of a modern ﬂuorescent lamp, presented as a histogram of the intensity against wavelength. This spectrum was obtained automatically from the preceding spectrum. Courtesy Dr K. H. Elliott.
Fig. 1.14 Spectrum of the Sun. Courtesy Dr K. H. Elliott.
Fig. 1.15 Spectrum of the Sun, presented as a histogram of the intensity against wavelength. Courtesy Dr K. H. Elliott.
20 Introduction
2
1.9
Refractive index
1.8 Dense flint
1.7
1.6 Borosilicate crown 1.5 Fused Silica
1.4 MgF 1.3
1.2 0
0.5 1 Wavelength in µ m
1.5
Fig. 1.16 The variation with wavelength of refractive index of optical glass, of fused silica and of MgF2 .
has a refractive index of 1.0003 at normal temperature and pressure (NTP: 20◦ C; 105 Pa) for visible light. Water has a refractive index of 1.333 at NTP for visible light, but the value rises to around 9.0 for microwaves. This diﬀerence indicates that light and microwaves interact very diﬀerently with water; in fact water is quite transparent to visible light while it absorbs microwaves strongly enough to make cooking with microwaves practical. Materials transparent to light have refractive indices ranging from near unity for gases up to 2.47 for diamond, with those for common glasses lying in the range 1.5 and 2.0. The variation of the velocity of electromagnetic radiation with wavelength is known as dispersion and gives rise to eﬀects such as the dispersion of light by a prism. For most materials that are transparent to light the refractive index falls smoothly with wavelength across the visible spectrum. A few materials show anomalous dispersion and for these the refractive index rises from the blue end to the red end of the spectrum. Transmittance is the fraction of radiation transmitted by a material, and falls below unity because there is absorption in the body of material and reﬂection at the entry and exit surface. Internal transmittance is deﬁned as the light intensity reaching the exit surface divided by that which entered the
1.6
material. Figure 1.16 shows the variation of refractive index of several materials commonly used in making lenses or other transparent optical components: borosilicate crown glass and dense ﬂint glass; fused silica which is useful in extending coverage into the UV; magnesium ﬂuoride (MgF) which is useful in extending coverage far into the IR, as well as being used in coating lenses to reduce surface reﬂections. Crown glass is transparent from 350 to 2000 nm; ﬂint glass from 420 to 2300 nm; fused silica from 260 to 2500 nm; and MgF from 120 to 8000 nm. The normal dispersion curves for glasses are well ﬁtted by the Sellmeier empirical formula over the visible spectrum n2 = 1 + B1 /(λ2 − C1 ) + B2 /(λ2 − C2 ) + B3 /(λ2 − C3 ).
(1.27)
The values of the constants Bi and Ci are often tabulated by manufacturers of optical materials.
Log(% eye efficiency)
Log(solar radiance)
Log(water absorption)
Figure 1.17 shows a correlation between optical properties which is crucial for us Earth dwellers. In ﬁgure 1.17(a) the absorption coeﬃcient 6 4 (a)
2 0 2 4 4
3
2
1 0 1 2 3 Log(wavelength in µ m)
4
5
6
5
6
5
6
8 6 (b)
4 2 0 4
3
2
1 0 1 2 3 Log(wavelength in µ m)
4
2
(c)
1
0 4
3
2
1 0 1 2 3 Log(wavelength in µ m)
4
Fig. 1.17 Coincidences: (a) log of absorption coeﬃcient for water in cm−1 ; (b) log of the Sun’s spectral irradiance in W m−2 µm−1 (c) log of the relative spectral eﬃciency of the human eye.
Absorption and dispersion 21
22 Introduction
for water is plotted as a function of the wavelength. An absorption coeﬃcient α is deﬁned such that a beam of radiation will be reduced in intensity by a factor α ds after passing through an inﬁnitesimal layer of thickness ds. That is dI(s)/ds = −αI(s). Integration gives the dependence of the intensity with distance s to be I(s) = I(0) exp (−αs),
(1.28)
which is known as Beer’s law. The absorption for water is at a minimum at approximately 500 nm and rises very fast as the wavelength moves oﬀ that value. In ﬁgure 1.17(b) the spectrum of radiation emitted by the Sun is depicted. It is roughly the radiation spectrum of a black body at 6000 K, and peaks very close to the wavelength at which the absorption of water reaches a minimum. Finally ﬁgure 1.17(c) shows the relative spectral sensitivity of the retina of the human eye, with the peak sensitivity at 550 nm. Evidently the eye is well designed to use the available radiation illuminating the planet. If instead, the retina were designed to ‘see’ at wavelengths just a factor three away (180 nm or 1650 nm) the absorption by water (α now around unity) would be a severe restriction. Radiation at 180 nm, in the ultraviolet, is in any case very damaging to tissue. Attenuation of the ultraviolet component of the Sun’s radiation by water vapour, oxygen and ozone in the atmosphere is essential to protect our eyes and skin. The main process by which the Sun’s energy is converted into a form accessible to living things is plant photosynthesis, which involves photochemical transitions in complex molecules. Plant photosynthesis operates at peak eﬃciency using light in the visible spectrum, which is just the wavelength range of the copious radiation penetrating to ground level. Much longer wavelength radiation cannot initiate these molecular processes, while shorter wavelength ultraviolet radiation would destroy the active molecules. The multiple coincidence presented in ﬁgure 1.17 is fundamental for life as we know it.
1.7
Radiation terminology
The time averaged energy ﬂow of radiation across unit area of a surface is known as the irradiance and also as the intensity. Radiation is not generally directed in a beam but spread over a range of angles, hence a quantity, the radiance, is deﬁned as the radiated energy per unit solid angle per unit area. The total energy radiated by a source or crossing a surface is called the radiant ﬂux, φe . How energy is distributed with wavelength is of importance because the physical eﬀects of radiation depend strongly on the wavelength. The spectral radiant ﬂux is the radiant ﬂux per unit wavelength, φe,λ , and
1.7
naturally the total
φe =
∞
φe,λ dλ.
(1.29)
0
A signiﬁcant property of radiation is whether it produces a visual eﬀect. Therefore visual quantities are deﬁned parallel to the energy/radiation quantities so far described. The fundamental factor used to relate visual and energy quantities is the luminous eﬃcacy, Vλ , so that the spectral luminous ﬂux associated with the spectral radiant ﬂux is φv,λ = Vλ φe,λ ,
(1.30)
where the subscripts e and v refer to energy and visual quantities. The variation of Vλ with wavelength was standardized by the Comite International d’Eclairage (CIE) so as to correspond to the response of the average human eye. In daylight the sensitivity of the human eye is greatest at a wavelength of 555 nm. The unit of luminous ﬂux, the lumen (lm), is deﬁned by taking one watt of radiation at 555 nm to have a visual equivalent of 683 lm. The reason for such a strange conversion factor is that the use of the energy and visual units developed quite independently. Figure 1.18 shows how the sensitivity of the eye varies with wavelength. The solid
1600
Luminous efficacy in lm/W
1400
Scotopic (night)
1200 1000 800 Photopic (day)
600 400 200 0
400
450
500 550 600 650 700 Wavelength in nm. Fig. 1.18 The relative sensitivity of the average human eye as a function of the wavelength of light.
curve shows the variation of Vλ in daylight (photopic vision), while the broken line curve is the corresponding curve for night vision (scotopic vision). Daylight and night vision rely on diﬀerent receptors in the retina: in daylight on the cones and at night on the rods. Cones come in three
Radiation terminology
23
24 Introduction
types which respond to blue, green and yellow light, respectively. The more sensitive rods all have the same spectral response so that night vision is monochrome. Their response peaks at 1700 lm for one watt of radiation at 507 nm wavelength. The luminous ﬂux from a source is φv = φv,λ dλ = φe,λ Vλ dλ,
(1.31)
where the integral need only run over the visible spectrum from 400 to 700 nm. The visual and energy quantities are shown in Table 1.1 with corresponding pairs on the same line. Of the remaining quantites only the visual ones will be discussed because their units are the less obvious. The luminous intensity of a point source in a given direction is Iv = dφv /dΩ,
(1.32)
where dΩ is an element of solid angle, and Iv is measured in lumens per steradian or candelas (cd). If the point source is isotropic then, of course, φv = 4πIv . The illuminance Ev is the luminous ﬂux per unit area, measured in lumens per square metre, or lux (lx). Finally the luminance, Lv , is the luminous ﬂux per unit area per unit solid angle, measured in cd m−2 . Luminance is, in everyday speech, the brightness of an object. At luminances below 0.03 cd m−2 human vision relies on the rods, and at luminances above 3 cd m−2 on the cones. Between these extremes both types of receptor play a part. Typical luminances are 100 cd m−2 for indoor lighting, 104 cd m−2 in full sunlight and 10−3 cd m−2 in starlight. One cd cm−2 may be called a stilb, while in describing the luminance of display panels one cd m−2 is called a nit. A diﬀuse source is one which looks equally bright in all directions. When viewed at an angle θ away from the normal the projected area is reduced by a factor cos θ. It follows that, in order to compensate, the luminance of a diﬀuse source should obey Lambert’s law: Lv (θ) = Lv (0) cos θ.
(1.33)
Integrating over the forward solid angle gives the total luminous ﬂux from unit area of the source 2π π/2 Ev = Lv (θ) sin θ dθ dφ φ=0
θ=0 π/2
= 2π
Lv (θ) sin θ dθ θ=0
π/2
= πLv (0)
sin (2θ) dθ θ=0
= πLv (0).
(1.34)
1.7
Table 1.1 Table of related visual and energy quantities and their units. Energy parameter
Unit
Visual parameter
Unit
Radiant ﬂux Radiant intensity Irradiance/intensity Radiance
W W sr−1 W m−2 W m−2 sr−1
Luminous ﬂux Luminous intensity Illuminance Luminance
lm lm sr−1 lm m−2 lm m−2 sr−1
A rough surface painted matt white is an excellent diﬀuse reﬂecting surface. Light from a compact source can be diﬀused by passing it through a ground glass screen. It is surprising that the brightness of an extended source does not change with its distance from the observer, unlike a point source. This eﬀect is noticable on a bright day: for example when viewing the white cliﬀs of Dover from the crossChannel ferry. The area of cliﬀ face lying within a ﬁxed solid angle at the observer’s eye pupil increases like the distance squared as the cliﬀs recede. At the same time the solid angle presented by the pupil to the cliﬀ face falls oﬀ at the same rate. The two eﬀects cancel and the luminance at the pupil remains the same. In turn this means that the illuminance of the image on the retina remains the same as the ferry moves away. Incandescent light bulbs produce around 15 lm for each watt of electric power drawn from the wall plug; ﬂuorescent lamps and ceramic metalhalide lamps produce ∼ 80 lm W−1 ; high pressure sodium lamps produce ∼ 120 lm W−1 . In these examples lm W−1 is the rate of converting electrical energy from the mains to visible energy expressed as lumens. Thus the eﬃciency of the conversion of electrical to visible energy in the case of high pressure sodium lamps is only 120/683 or around 18%. The other 82% goes to heat the surroundings. When expressing the intensity ratio between the light entering and leaving an optical ﬁbre link the decibel (dB) is often used. If the ratio of the light entering divided by the light emerging is R, then in decibels this becomes n = 10 log10 R. (1.35) Such a ratio unit is convenient when the intensity ratios are very large. If the loss along an optical ﬁbre is n dB km−1 then over s km it is sn dB. Powers of, for example lasers, may be expressed in dBm in which the ratio is the laser power divided by 1 mW. Thus if the laser power is stated to be −5.0 dBm this means that the power, P , is given by −5.0 = 10 log10 [ P/(1 mW) ]. Thus the laser power is 10−0.5 = 0.316 mW.
(1.36)
Radiation terminology
25
26 Introduction
1.8
Black body radiation
This is radiation which can be produced by simple apparatus and whose properties can be analysed simply too. Figure 1.19 shows the black body spectrum at three temperatures. As noted above, the Sun’s emission spectrum, before absorption in the Earth’s atmosphere, is approximately that of a black body at 6000 K. Consider an enclosure whose walls are maintained at a constant and uniform temperature, and which is taken to be evacuated. Suppose a small body is suspended within this volume: in thermal equilibrium it will radiate as much energy as it absorbs. This shows immediately that a good emitter at any wavelength must necessarily also be an equally good absorber of radiation at the same wavelength, and a poor emitter must be a poor absorber. It follows that the radiation in the enclosure at a given temperature is independent of the material in the walls of the cavity. If this were not the case then 0.6
5000K 0.5
3
Energy density in J m µm
1
0.4
0.3
0.2
4000K
0.1 3000K
0 0
0.2
0.4
0.6
0.8 1 1.2 1.4 Wavelength in µm
1.6
1.8
2
Fig. 1.19 Black body radiation spectra at 3000 K, 4000 K and 5000 K.
1.9
energy could be transfered between bodies at the same temperature. The radiation in such an enclosure is known as black body radiation. If an aperture very much smaller than the linear dimensions of the enclosure is left open, then this acts as a source of black body radiation, and is a perfect emitter because it transmits all the radiation coming from inside. This aperture is also a perfect absorber of radiation falling on it because the radiation entering has a negligible chance of being reﬂected back out of the aperture. The spectral emittance λ of any surface can be deﬁned as the fraction of radiation emitted around wavelength λ compared to that emitted by a black body. Also the spectral absorptance aλ can be deﬁned as the fraction of incident radiation that a surface absorbs at the same wavelength. Then the result previously deduced is that λ = aλ ,
(1.37)
and is known as Kirchhoﬀ ’s law. If a large absorptance could be allied to a smaller emittance, then a body made from this anomalous material suspended in the thermal enclosure would form a heat engine that violates the second law of thermodynamics. Practical black body sources are thermally insulated boxes, whose walls contain heaters under thermostatic control, the interior wall surfaces are also corrugated. Commercially available sources running typically at 1000 K have emittances of over 99% across the wavelength range 1 to 30 µm.
1.9
Doppler shift
Everyone is aware that the tone of the siren of an emergency vehicle drops just as the vehicle passes by. When it is approaching the frequency is higher, and when it is receding the frequency is lower than that of the siren at rest. These frequency shifts constitute the Doppler eﬀect, which is observed both for sound and for electromagnetic waves. The upper panel of ﬁgure 1.20 shows sound waves emitted by an approaching source. This source has velocity ve ; the sound has velocity vs , frequency f and wavelength λ. In a short time ∆t the source emits f ∆t waves and the leading one of these waves will travel a distance vs ∆t. In the same time the source moves a distance ve ∆t so that the f ∆t waves are conﬁned to a reduced distance (vs − ve )∆t. Thus the wavelength is compressed to λ = (vs − ve )∆t/f ∆t. = λ(vs − ve )/vs = λ(1 − ve /vs ). Hence the frequency heard is higher f = f /(1 − ve /vs ).
(1.38)
The case that the observer is moving with velocity vo and the source is at rest is shown in the lower panel of ﬁgure 1.20. This observer receives
Doppler shift
27
28 Introduction
Observer at rest
Source moving
ve ∆ t vs ∆ t
Observer moving
Source at rest
vo ∆ t Fig. 1.20 Doppler shift of frequency when source or observer are moving.
more waves per unit time than one at rest. In a time ∆t the total number of waves passing the observer is (vs + vo )∆t compared to vs ∆t for an observer at rest. Thus the frequency heard is f = f (vs + vo )∆t/vs ∆t = f (1 + vo /vs ).
(1.39)
In the above two equations the velocity of the source or observer needs to be replaced by the component of the velocity toward the observer or source respectively when the motion is not along the line joining them. So the ﬁrst result becomes f = f /(1 − vr /vs ),
(1.40)
where vr is the radial component of the source’s velocity. For electromagnetic radiation from a moving source in free space this analysis would give f = f /(1 − vr /c), (1.41) where c is the velocity of light. However there is an additional relativistic eﬀect that must be taken into account known as time dilation. The the time intervals between events occuring in the source’s rest frame are longer by a factor γ = 1/ [1 − (v/c)2 ] when they are timed by a stationary observer. Although macroscopic objects on Earth do not travel at velocities close to that of light, some elementary particles do so and the result can be astonishing. µleptons created by primary cosmic rays in interactions with atmospheric atoms at heights of 200 km or so mostly survive to reach the Earth’s surface. They are travelling close to the speed of light (relativistically) so the duration of this journey is (200 km)/c ≈ 0.67 ms or somewhat longer. The lifetime of a µlepton at rest measured in the laboratory is only 2 µs and hence, viewed classically, few
1.9
Doppler shift
29
µleptons should survive to reach the Earth! However the time dilation factor increases the lifetime measured in the Earth’s frame of reference by a factor of around 1000, and so most do reach us. Thus the frequency detected by the observer is correspondingly lower than that seen in the rest frame of the source, f0 , by a factor [1 − (v/c)2 ]: (1.42) f = f0 [1 − (v/c)2 ]. Note that this depends on v rather than its radial component. Combining eqns. 1.41 and 1.42 gives the frequency detected by the observer f = f0 [1 − (v/c)2 ]/[1 − vr /c] (1.43) where f0 is the frequency of the source measured at rest with respect to the source. When the source moves directly toward the observer f = f0 (1 + v/c)/(1 − v/c). (1.44) Turning to the situation that the observer is in motion it is found that the same eqns. 1.43 and 1.44 are obtained for the frequency of the electromagnetic radiation determined by the observer. This is but one example of the basic feature of the special theory of relativity that the relative velocity is the relevant parameter. In equations 1.43 and 1.44 we should therefore identify v as the relative velocity of approach of source and observer, and vr as its radial component. When v is small compared to c eqn. 1.44 can be approximated by: f = f0 (1 + v/c),
(1.45)
which, not surprisingly, is also the classical prediction.
Exercises (1.1) The solar ﬂux impinging on the Earth’s atmosphere at the equator is 1.5 kW m−2 . Calculate the magnitude of electric ﬁeld there. What is the Sun’s total energy output? (1.2) Light from a laser of wavelength 633nm in space is reﬂected from a comet and the returning light is found to be redshifted by 10−1 nm. What is the relative velocity of the comet with respect to the observer? (1.3) The Lyman α line in the atomic spectrum of hydrogen has wavelength 121.6 nm in the UV. The wavelength of this same line in the spectrum received from a quasar 01422+2309 is 561.79 nm. Is
this red shift conceivably due to the recession velocity of the quasar away from the Earth? What else could have stretched the wavelength by this factor? (1.4) What are the periods, wavelengths, velocies and directions of these waves: (a) A(x, t) = cos [2π(3t + 15x)]; (b) B(x, t) = exp i[2π(5t − 15x − 20y)]? (1.5) An electromagnetic wave E(x, t) = 15.0 sin [2π(f t − x/λ + φ)] has values 0 Vm−1 and −9.95 Vm−1 at locations (t, x) = (0, 0) and (0, 1500 m) respectively. Calculate f , λ and φ, taking the longest wavelength solution.
30 Introduction (1.6) A laser emits a beam with 1 kW power in free space. What is the energy in a length of 1 m of the beam?
safe level.
(1.9) Microwave ranges operate at 2.45 GHz. What is the corresponding wavelength? (1.7) E0 cos (ωt − kx) and E0 cos (ω t − k x) are two electromagnetic waves in free space and ω = ω + (1.10) A beam of light of wavelength 500 nm falls perpen∆ω where ∆ω is small compared to ω. Show that dicularly on a screen with two holes. One beam k = k − ∆k where ∆k/k = ∆ω/ω. travels in glass, the other in air. How long will the glass need to be to cause a delay of 1 ns between (1.8) Mobile phones operate at a frequency of 1 GHz. the beam in glass relative to that in air? Take the What is the wavelength? The power radiated by refractive indices of air and glass to be 1.0 and 1.5 a digital low power mobile phone is 125 mW. If the respectively. power is assumed to radiate isotropically what is the electric ﬁeld at 2 cm from the antenna? The (1.11) Use ﬁgure 1.17 to calculate the fraction of radiation maximum limit recommended by the IEEE for RF at 10 cm wavelength from the Sun that penetrates power in head tissue is 1.6 mW gm−1 over any 1 gm the atmosphere, assuming there is 10 kg m−2 of waof tissue, and is designed to limit any heating to a ter in the atmosphere.
Reﬂection and refraction at plane surfaces 2.1
2
Light rays and Huygens’ principle
The crisp edges of shadows on a sunny day remind us that on the scale of everyday objects light travels in straight lines. The idea of a light ray indicating the path of light appears in depictions of the Sun during the reign of the monotheistic pharaoh Akhenaton (c. 1370 BC), as sketched in ﬁgure 2.1. Around 1000 AD Al Hazen made a simple and elegant experiment that validates the idea. He placed ﬁve lamps in one room and made a small hole in the partition separating it from an adjacent darkened room. Al Hazen saw distinct images of each ﬂame on the wall of the darkened room, and he noted that he could remove each image simply by putting his hand in the appropriate ray path in the darkened room. In this and the following two chapters the properties of mirrors, lenses and optical instruments will be studied using ray optics. The ﬁrst requirement is to understand how rays and waves are related. In the case of the plane wave, eqn. 1.12, E = E0 cos (ωt − k · r + φ), the rays follow the direction of the wave vector k. Rays therefore point perpendicularly to the wavefronts. In the case of spherical waves given by eqn. 1.13, E = E0 cos (ωt − kr), the rays are outwardly directed radial lines perpendicular to the spherical wavefronts. At a boundary between two media, for example air and water, rays other than those exactly perpendicular to the surface change direction there: they undergo refraction. Refraction occurs because the velocity of light in matter varies from one material to another. The velocity also depends on the state of the material: the velocity being higher, for example, the less dense a gas is. Thus on a clear summer’s day, when the air is hotter the closer it is to the tarmac of the road surface, light from the sky is refracted in this region so that it turns upward. A mirage is seen, an apparent pool of water reﬂecting the sky. The laws of reﬂection and refraction at surfaces can be deduced using Huygens’ principle (1678), an early step in the development of the wave theory of light which is still useful. Huygens proposed a simple picture of how the ‘disturbance’ at one wavefront produces a disturbance at a later time. His principle states that all points on a wavefront can be treated as point sources of secondary spherical waves. Then at a later time the new position of the wavefront is the surface tangential to the forward going secondary
Fig. 2.1 Typical depiction of the Sun, Aten, during the reign of Akhenaten. Each ray ends in a pointing hand.
32 Reﬂection and refraction at plane surfaces
waves. Figure 2.2 shows what happens in the case of a plane wave in free space. After a time t the spherical secondary wavefronts have radius ct. The new wavefront tangential to them is planar and a distance ct ahead; consistent with expectation.
ct
Old wavefront
New wavefront
Fig. 2.2 The construction of a future wavefront, at a time t later. Three of the spherical waves used in Huygens’ construction are shown.
Huygens’ construction is adequate away from any obstruction in the wave path. However at the edge of an aperture the construction predicts that the wave spills round this edge. A more complete version of wave theory is needed in order to explain what happens at apertures. For the present it is suﬃcient to note that light of wavelength λ passing through an aperture of width a shows departures in angle of order λ/a from straight line propagation. When green light passes through an optical component of aperture 1 cm this amounts to 5 10−5 rad, which we can safely neglect in this and the following chapter. In the following two sections the laws of reﬂection and refraction will be deduced from Huygens’ principle. These laws will then be used to study the imaging produced by plane mirrors, plane sheets of transparent material and simple triangular prisms. After that, total internal reﬂection (TIR) at a surface between two media is described: this is how light is guided along optical ﬁbres and the low losses required for telecom transmissions are achieved. For simplicity the refractive index of air will be taken to be exactly unity in discussions of optical elements in air.
2.1.1
The laws of reﬂection
In ﬁgure 2.3 ABC is a plane wavefront which has just reached the mirror surface at A at a given moment. Huygens’ construction of a wave at a later time begins with drawing spherical secondary waves from all points on this wavefront. Representative examples, originating from points A, B and C, are shown at a time t later when they have travelled out a distance ct. In Huygens’ construction the wavefront at time t later
C B
A’ A
θ
C’
θ’ Mirror
B’
Fig. 2.3 Incident wavefront ABC and partially reﬂected wavefront A B C at a plane mirror. The arrowheaded broken line indicate ray paths.
is the surface tangential to these secondary waves: A B C . The part
2.1
Light rays and Huygens’ principle 33
B C is still an incoming wave while part A B is a reﬂected; and these movements are indicated by the arrows on the rays. In the triangles AA B and B BA: (2.1) AA = BB = ct; AB is common;
(2.2) ◦
AA B = ABB = 90 .
Normal Ray
W
(2.3)
Thus the triangles are similar and we have the law of reﬂection:
θ θ
θ = θ.
(2.4)
Now the angle between the ray and the normal to the mirror is identical to the angle between the wavefront and the mirror surface, as shown in ﬁgure 2.4. Thus the incident and reﬂected rays make equal angles with that normal. The plane formed by the incident and reﬂected rays contains the normal to the surface at the point of reﬂection; it is called the plane of incidence.
2.1.2
Snell’s law of refraction
The refraction of light can be handled similarly using Huygens’ principle. In ﬁgure 2.5 light is incident on a plane interface separating a medium of refractive index n1 from one of refractive index n2 ; the velocity of light in the two media is thus v1 = c/n1 and v2 = c/n2 . PP is perpendicular to the interface. The choice is made that n1 > n2 : the ﬁrst medium is said to be optically denser so that light travels more slowly in the ﬁrst medium than in the second. ABC is an incoming wavefront that has
n1
C
P B θ1 A
C’
θ1
P’
B’
θ2
θ2
Surface
n2
A’
t
on
efr
av
n1 > n 2 Fig. 2.5 Refraction at a plane surface separating media of refractive indices n 1 and n2 , with n1 > n2 .
Mirror
Fig. 2.4 The angles between the ray and the surface normal and between the wavefront and the mirror surface are equal.
34 Reﬂection and refraction at plane surfaces
just reached the interface at A. Spherical waves originating from A, B and C are shown at a time t later. Then:
Now
B B = ct/n2 and AA = ct/n1 .
(2.5)
AB = AA / sin θ2 = BB / sin θ1 .
(2.6)
ct/n2 sin θ2 = ct/n1 sin θ1 ,
(2.7)
Thus so we obtain Snell’s law n1 sin θ1 = n2 sin θ2 .
(2.8)
Here the plane of incidence contains the incident ray, the refracted ray, the surface normal and the reﬂected ray.
A θ1 x1
h1 n1
B Surface
n2
dx
h2
θ2
x2 C
Fig. 2.6 ABC is the actual optical path of light rays between points A and C. The broken line is a nearby path.
A diﬀerent way of presenting Snell’s law is informative and will also be useful later. The starting point is to note that the waves match at the interface: exactly at the boundary the wave peaks in the second medium are in precisely the same places as the wave peaks in the ﬁrst medium. The wavefront A B C in ﬁgure 2.5 illustrates this. The wavelength in the ﬁrst medium, λ1 , can be expressed in terms of the free space wavelength, λ, thus λ1 = v1 /f = c/(f n1 ) = λ/n1 , (2.9) with a similar expression for the second medium. Note that the frequency of em radiation remains the same in going from one material to another because an electric ﬁeld at frequency f arriving at the interface produces eﬀects at the same frequency f in the medium on the far side of the surface. Now using the fact that the separation between peaks in the two media are equal along the interface, we have: λ1 / sin θ1 = λ2 / sin θ2 .
(2.10)
In terms of the wave number (k = 2π/λ), this becomes: k1 sin θ1 = k2 sin θ2 .
2.1.3
(2.11)
Fermat’s principle
Yet another way can be used to prove Snell’s law which illustrates a principle that was ﬁrst enunciated in a clear form by Fermat. He proposed that the optical path taken between two points by light is the path which minimizes the travel time of light. The travel time is T = ni i /c, (2.12) i
where the sum is taken over all path elements of length i and refractive index ni . An optical path length is deﬁned as L= ni i . (2.13) i
2.1
Light rays and Huygens’ principle 35
Fermat’s principle correctly predicts that rays in a uniform medium follow straight lines. Its success in the case of refraction is easily proved. The actual path of light, ABC, shown in ﬁgure 2.6 has an optical length L = n1 h1 sec θ1 + n2 h2 sec θ2 ,
(2.14)
and the travel time is L/c. A small displacement, dx, of B to the left along the interface the results in a change in path length dL = n1 h1 sec θ1 tan θ1 dθ1 + n2 h2 sec θ2 tan θ2 dθ2 ,
(2.15)
where dθ1 and dθ2 are the corresponding changes in the angles θ1 and θ2 respectively. From the diagram we have x1 = h1 tan θ1 x2 = h2 tan θ2 . Now dx = −dx1 = dx2 , so that dθ1 = − cos2 θ1 dx/h1 , dθ2 = + cos2 θ2 dx/h2 .
A’’
Substituting for dθ1 and dθ2 in eqn. 2.15 gives
Mirror
Eye
dL = −n1 sin θ1 dx + n2 sin θ2 dx.
P
For a minimum of the path length we require dL/dx = 0, which immediately gives Snell’s law. The reader may like to test Fermat’s principle in the case of reﬂection. It is signiﬁcant that from the simple idea that optical paths should be of extremal length it has been possible to reproduce ray optics. From the standpoint of classical physics Fermat’s principle is simply the principle of least action applied to optics.
2.1.4
Simple imaging
Figure 2.7 shows how an image is formed by a plane mirror. A cone of rays diverging from the object A are reﬂected from the mirror between P and P and focused by the eye onto the retina at A . The rays appear to diverge from the mirror image A : AOA is perpendicular to the mirror and AO = OA . Figure 2.8 shows another simple situation where the object viewed lies within a layer of a material of refractive index n. With the notation of the diagram: n sin r = sin i.
(2.16)
For the case that the object is viewed close to the surface normal so that the angles i and r are small: a = s/ tan i ≈ s/ sin i; d = s/ tan r ≈ s/ sin r.
(2.17)
d/a = sin i/ sin r = n.
(2.18)
Then:
P’ Object A
Image A’ a
a
Fig. 2.7 Virtual image A of point A, distance a from mirror, and focused on the retina at A .
Air: n=1.00
a
Water: n=1.33
s
i
d
r
Image Object
Fig. 2.8 Viewing an object in an optically denser medium.
36 Reﬂection and refraction at plane surfaces
Hence the depth appears shallower by the ratio of the refractive indices. When an object is viewed through a block of glass the image is displaced sideways as in ﬁgure 2.9; and the displacement, d, changes as the block is rotated. Such a rotatable, parallelsided glass plate provides simple lateral alignment in some optical instruments.
2.1.5 Air
i
r Glass
r
Air
d
i
Fig. 2.9 Displacement of an image by a thick parallel sided glass plate.
Deviation of light by a triangular prism
Simple prisms like that shown in ﬁgure 2.10 are used to disperse white light into the diﬀerent colours. Light from a source is ﬁrst collimated into a parallel beam whose crosssection would be a thin line perpendicular to the diagram. The incoming arrow indicates this beam incident on the prism. If the light is of a single wavelength it will follow the arrowed line and produce a line image perpendicular to the diagram on a screen at the right hand side of the prism. Figure 2.11 shows typical spectra produced on the screen when the source slit is illuminated by discharge laboratory lamps each containing gas of one element. Each line in the spectrum is produced by light of a diﬀerent wavelength emitted by the gas in the lamp. They are separated, dispersed, because the refractive index of glass varies across the visible spectrum. The angular deviation of the ray drawn in ﬁgure 2.10 is
α
δ = i − r + e − s. δ
i
β r
s
γ
e
(2.19)
Summing the angles of the triangle bounded by the ray and the prism edges, gives (90◦ − r) + (90◦ − s) + α = 180◦ , so that, r + s = α.
(2.20)
δ = i + e − α.
(2.21)
Then eqn. 2.19 becomes Fig. 2.10 Ray path through prism. The deviation δ is minimum in the symmetrical conﬁguration where i = e.
Applying Snell’s law to the two refractions e = sin−1 (n sin s) = sin−1 [ n sin (α − r) ] = sin−1 (n sin α cos r − n cos α sin r). Now sin r = sin i/n, and cos r =
(2.22)
1 − sin2 i/n2 .
Making these substitutions in eqn. 2.22 gives e = sin−1 sin α (n2 − sin2 i) − cos α sin i .
(2.23)
Then replacing e in 2.21, δ = i − α + sin−1 sin α (n2 − sin2 i) − cos α sin i .
(2.24)
2.2
Total internal reﬂection 37
The dependence of the deviation on the angle of incidence is plotted in ﬁgure 2.12 for a prism with an apex angle α equal to 30◦ and a refractive index of 1.5. There is a quite shallow minimum in the distribution of deviation against the angle of incidence. The minimum occurs in the symmetric arrangement where e = i, and s = r. At minimum deviation eqns. 2.20 and 2.21 become r = α/2, and δmin = 2i − α.
(2.25)
Applying Snell’s law at minimum deviation at either surface, and using 2.25: n = sin i/ sin r = sin [ (δmin + α)/2 ]/ sin (α/2). (2.26) Neon
Now the refractive index of glass changes with wavelength. Thus minimum deviation occurs at diﬀerent angles for the diﬀerent colours, with the results shown in ﬁgure 2.11. In the case of borosilicate crown glass the refractive index changes from 1.51 to 1.54 between the red and blue ends of the spectrum. Of equal importance for spectroscopy, the refractive index changes monotonically with wavelength for any common type of glass so that the dispersion does not superpose colours. Consequently a prism gives a spectrum in which the wavelength increases smoothly from one end to the other. A measurement of the angle of minimum deviation easily determines the refractive index to one part in ten thousand. When the prism angle is small such that sin α ≈ α eqn. 2.26 becomes
Helium
Mercury
Hydrogen Wavelength in nm. 400
550
700
Fig. 2.11 Spectral lines observed with lamps containing gases of various elements at low presure.
n = (δmin + α)/α,
(2.27)
δmin = (n − 1)α.
(2.28)
40
The deviation minimum is ﬂatter for a narrow angle prism, so that this equation for the minimum deviation is a good approximation to the actual deviation over a wide range of angles of incidence around the symmetric arrangement.
35
2.2
Total internal reﬂection
When light is incident on a surface between two materials from the more optically dense material then at suﬃciently large angles of incidence the refracted ray is suppressed. Rewriting Snell’s law: sin θ2 = (n1 /n2 ) sin θ1 ,
(2.29)
θ2 > θ 1 .
(2.30)
so that There is therefore an angle of incidence called the critical angle θc , at which the angle of refraction will reach 90◦ : sin θc = n2 /n1 .
(2.31)
Deviation in degrees
whence
30 25 20 15 10
Prism parameters α = 30 o and n = 1.5
5 0 0
10 20 30 40 50 60 70 80 90 Incident angle in degrees
Fig. 2.12 The deviation of a ray passing through a prism as a function of the angle of incidence.
38 Reﬂection and refraction at plane surfaces
n1
n2 n2 < n1
θ1 θ1
θc θc
θ2
θ 2 = 90
(a)
(b)
θ1
θ1
o
(c)
Fig. 2.13 Refraction at an optically denser medium. Angle of incidence (a) less than the critical angle, (b) equal to the critical angle θc , and (c) greater than the critical angle.
At larger angles of incidence the light is totally reﬂected and there is no refracted ray travelling into the less optically dense material. Figure 2.13 shows the situation for increasing angles of incidence: for the angle of incidence less than the critical angle; equal to the critical angle of incidence when the refracted ray is parallel to the surface; and for a larger angle of incidence when the reﬂection is total. The property of total internal reﬂection is used widely to guide light with low loss. These applications include the use of prisms in optical instruments to manipulate light beams, and transmission of electromagnetic radiation along optical ﬁbres.
2.2.1
300
450 300
Fig. 2.14 Constant deviation prism. The dotted lines mark the boundaries of the hypothetical component prisms. The ray shown undergoes minimum deviation.
Constant deviation prism
Figure 2.14 shows a single compound prism which converts minimum deviation into a ﬁxed deviation of 90◦ . One such minimum deviated ray is indicated on the diagram. The minimum deviation is produced not by a prism of apex angle 60◦ but eﬀectively by two prisms, each of apex angle 30◦ . Between these a 45◦ prism is inserted to give a total internal reﬂection. The notional outlines and angles of these components are shown in the diagram. If the incident ray encounters the ﬁrst half prism at minimum deviation, then it will leave this half prism perpendicular to the mid face. The TIR then deviates the ray exactly 90◦ so that it enters the second half prism perpendicular to its mid face also, and then completes the minimum deviation. The total deviation of the minimum deviated ray is just the 90◦ deviation produced by TIR: the deviations at the external surfaces of the half prisms are equal, but now in opposite senses, so they cancel. This prism is used in constant deviation spectrometers. The light is directed in a pencil beam at the prism by a collimator and observed by a telescope at right angles, all mounted in a ﬁxed frame. Only the prism needs to be rotated to bring each part of the spectrum into view at minimum deviation.
2.2
2.2.2
Total internal reﬂection 39
Porro prisms
Porro prisms have angles of 45◦ , 45◦ and 90◦ . A pair of them arranged as shown in ﬁgure 2.15 are used in binoculars to correct the inversion of the image produced by the lenses, giving an upright ﬁnal image. There are four total internal reﬂections on the light path shown, and each reﬂection turns a right handed object into a left handed image, and vice versa. The overall eﬀect is to preserve the original orientation, a right handed object remains right handed. Porro prisms used in binoculars have a narrow
ey
ez
Fig. 2.15 This arrangement of two Porro prisms is used in binoculars to correct the inversion of the image produced by the lenses, giving a ﬁnal upright image. The light path is shown for an object and its inverted image. Where the optical path lies inside the prisms the ray is indicated with a broken line.
groove ground across the centre of the front face, whose purpose is to absorb scattered light; light entering from outside the ﬁeld of view is scattered at glancing angles from the front faces into the ﬁeld of view and would otherwise give a background haze. The groove runs vertically (horizontally) on the lefthand (righthand) prism. Another advantage accruing from the inclusion of a pair of Porro prisms in binocular designs is that the front lenses can be displaced to be further apart than the eyes are; which gives improved depth discrimination. Finally the use of the prisms allows the designer to fold the optical path, so making the instrument compact and easy to manipulate.
2.2.3
Corner cube reﬂector
A corner cube reﬂector is, as the name implies, a corner cut from a glass cube. An example is shown in the upper panel of ﬁgure 2.16. The sloping face makes equal angles with the cube faces. When a ray enters through the sloping face it will undergo internal reﬂections in some order from each of the three perpendicular faces. It ﬁnally emerges accurately parallel to its original path but travelling in the opposite direction. To prove this assertion we consider an incident ray whose
ae
x+
ex
be
y+
ce
ae
z
+
x
+ be y
ce z
ex Fig. 2.16 The upper panel shows a corner cube and unit vectors perpendicular to the faces. The lower panel shows TIR from the xface.
40 Reﬂection and refraction at plane surfaces
direction on entering the corner cube is along the direction given by the unit vector aex + bey + cez , where the normals to the three mutually orthogonal surfaces of the corner cube are the unit vectors ex , ey and ez . After reﬂection from the ex surface the ray direction becomes −aex + bey + cez .
(2.32)
This is shown in the lower panel of ﬁgure 2.16. After reﬂections from all three orthogonal faces the exiting ray points along the direction −aex − bey − cez
which has reversed the incident ray. This property guarantees that whatever tilt the incoming beam has, it will always be reﬂected back whence it came. A square of 100 such corner cubes was left on the Moon during the Apollo XI mission and they provide a convenient mirror target for measurements of the Moon–Earth separation. Bicycle reﬂectors are arrays consisting of large numbers of small corner cubes moulded in clear plastic. They operate on the same principle and because the reﬂection at each facet is by TIR there is no need to metallize the back surface.
Sample θc θe
Fig. 2.17 The Pulfrich refractometer is illuminated by a focused beam. The critical ray’s path carries the solid arrows.
i a t b
(2.33)
c
2.2.4
Pulfrich refractometer
Measurement of the critical angle of reﬂection can be used to obtain an accurate determination of refractive index for solid and liquid samples. Figure 2.17 shows a Pulfrich refractometer with the test material, refractive index n, sitting on a 45◦ /45◦ /90◦ reference prism of accurately known refractive index, nr > n. Light of the selected wavelength coming from the left is focused onto the boundary so that the beam covers an angular range down to glancing incidence at the interface. The angular range of the beam entering the prism has a sharp cutoﬀ at the critical angle, θc , where sin θc = n/nr . Then applying Snell’s law to the critical ray at its exit from the prism sin θe = nr sin (90◦ − θc )
Fig. 2.18 An anamorphic prism pair.
= nr cos θc = nr 1 − sin2 θc = n2r − n2 . Thus the specimen’s refractive index is: n = n2r − sin2 θe .
(2.34)
(2.35)
Thus all that is required to determine the sample’s refractive index is the measurement of the angle θe .
2.3
Example 2.1 Figure 2.18 shows a beam incident on two identical prisms. These are oriented so that the beam enters each at the same angle of incidence, and in the case shown leave perpendicular to the exit surfaces. If the refractive index of the glass is n show that the beam width in the plane of the diagram is expanded by a factor c/a = [ n2 − sin2 i ] / [ n cos i ]2 . If d is the width of the beam measured along the surface of the ﬁrst prism a = d cos i; b = d cos t. Thus b/a = cos t/ cos i. Similarly c/b = cos t/ cos i. Hence c/a = cos2 t/ cos2 i = [1 − sin2 t ] / cos2 i = [1 − sin2 i/n2 ] / cos2 i = [n2 − sin2 i ] / [ n cos i ]2 .
(2.36)
The prisms do not aﬀect the width of the beam in the perpendicular dimension. This arrangement of prisms is called an anamorphic pair and it is used to render circular the elliptically shaped proﬁle of a laser diode beam. Anamorphic pairs are preferred over cylindrical lenses for circularizing laser diode beams because the prisms occupy less space and are cheaper. The degree of shaping can be altered as required by rotating the prisms.
2.3
Optical ﬁbre
An optical ﬁbre consists of a cylindrical glass core surrounded by concentric glass cladding of slightly lower refractive index. Optical ﬁbre communication relies on total internal reﬂection to guide em radiation along the core over paths extending to hundreds of kilometres. Near infrared radiation is used, this being the wavelength range for which absorption in the glass is at a minimum.
Optical ﬁbre
41
42 Reﬂection and refraction at plane surfaces
Cladding n 2 Air n
θ
θc
θc
Wavefronts Core n 1 (>n 2) Source
θacc
Fibre axis
Fig. 2.19 Longitudinal crosssection through an optical ﬁbre illuminated by a point source. The ray drawn is the one at the widest angle oﬀ axis to be totally internally reﬂected at the core–cladding interface.
Figure 2.19 shows a crosssection taken along the ﬁbre length. Some radiation travels inside the core and is totally reﬂected at the core– cladding interface. For clarity the diameter of the core in the diagram is made overlarge. On the left is a point source and a ray from the source which meets the core–cladding interface at the critical angle. After the ﬁrst reﬂection at this interface the ray will travel to the opposite side of the core where it will meet the core–cladding surface at the critical angle of incidence, just as in the ﬁrst reﬂection. Evidently, once a ray is trapped in the core it generally remains in the core. The rays from the source which are retained inside the core lie inside the cone of semiangle θacc . This acceptance angle will be a function of the refractive indices of the external medium, n, the core, nc and the cladding, ne . Snell’s law can be applied to the ray where it enters the core: n sin θacc = n1 sin θ.
(2.37)
At the core–cladding interface for the critical ray: n1 sin θc = n2 sin 90◦ = n2 ,
(2.38)
but θ = 90◦ − θc , so that: n1 cos θ = n2 . Hence sin θ =
1 − n22 /n21 .
(2.39)
(2.40)
2.3
Optical ﬁbre
43
Substituting for sin θ in 2.37: n sin θacc =
(n21 − n22 ).
(2.41)
The quantity n sin θacc is called the numerical aperture (NA) of the ﬁbre. The square of the NA, deﬁned in this way, is a standard measure of the light gathering power of an optical instrument, not simply of optical ﬁbres. A fuller account of the properties and use of optical ﬁbres is given in Chapter 16.
Exercises (2.1) A thin sheet of plastic is placed on a microscope stage. The distance between the top and bottom surfaces as measured by the microscope is 300 µm. If the plastic has refractive index 1.55 what is its actual thickness? (2.2) The angle of minimum deviation of a monochromatic beam of light is 20◦ as measured with a triangular prism with apex angle 25◦ . What is the refractive index of the prism for this wavelength of light? (2.3) What is the critical angle for a diamond/air interface, and for a diamond/water interface? The refractive indices of water and diamond are 1.33 and 2.47 respectively. (2.4) What is the NA of an optical ﬁbre with core and cladding of refractive indices 1.505 and 1.50 respec
tively at wavelength 1.3 µm? (2.5) A glass thread in air would guide light in the same way that an optical ﬁbre does. Why is this not a competitive solution for communications? (2.6) Use Fermat’s principle to prove the law of reﬂection. (2.7) A light ray travels through a pile of clear accurately parallel faced glass sheets, each in contact with its neighbours above and below. The glass sheets can have diﬀerent refractive indices. If the angle of incidence of the ray is 45◦ and it emerges from the ﬁnal sheet what can you say about the direction of the emergent ray? Will it ever be the case that total internal reﬂection will prevent the ray traversing all the sheets?
This page intentionally left blank
Spherical mirrors and lenses 3.1
Introduction
A good proportion of the adult population need lenses, in the form of spectacles or contact lenses, to carry out daytoday activities such as driving. Mirrors also abound: in the home concave mirrors provide a magniﬁed closeup view of the face, and convex mirrors on the car driver’s door give the driver a wider ﬁeld of view, with the catch that the car following appears further oﬀ than it actually is! Everything seen live on TV comes through a lens system, often of sophisticated design. These lenses contain ten or more elements: they can zoom in without loss of focus, compensate for the shaking hand that holds them, and provide an image that faithfully reproduces the colour and proportions of the scene. Other lens systems are used by scientists and engineers to study the very small (microscopes) and the very distant (telescopes). On many production lines there are monitoring systems relying on lenses to give noncontact measurements of size, location and orientation of rapidly moving items. Lenses and mirrors can also produce an intense concentration of energy at the focal point: this ranges from a lens concentrating sunlight so that a piece of paper catches ﬁre to the proposed ignition of nuclear fusion in deuterium/tritium pellets using laser beams. The properties of single lenses, single mirrors and complex combinations can be predicted using the laws of reﬂection and refraction derived in Chapter 2. Most of the lenses and mirrors met in instruments have spherical surfaces. When the rays from an object are paraxial, that is to say they lie close to the axis of an optical system and make small angles with that axis, then the equations for calculating the image position for optical systems with spherical surface are relatively simple. This is the paraxial approximation, and the paraxial formulae for determining the image position in the cases of reﬂection and refraction will be derived ﬁrst. Then the paraxial formulae for lenses and systems of lenses are derived and reformulated in the more convenient matrix form. Rays from a point object that are either at wide angles to the optical axis of the lens/mirror, or travel far from the optical axis will not converge precisely at the paraxial image point. The deviations of these rays from the paraxial image are called aberrations. Because the refractive in
3
46 Spherical mirrors and lenses +
Light
REFLECTION
in + Pole
Light
out +
Measure from pole: positive if in direction indicated +
Pole
Light
in/out + REFRACTION
Fig. 3.1 Cartesian sign conventions.
dex changes with wavelength the position of the image formed by a lens will depend on wavelength, and this will be the case even for paraxial rays. The resulting deviations from perfect imaging are called chromatic aberrations. Mirrors and instruments that only use mirrors are evidently free from chromatic aberrations. In the latter half of the chapter aberrations and the techniques used to reduce them are described.
3.1.1
These conventions lead to a relatively simple formalism: the paraxial lens and mirror formulae for imaging and magniﬁcation are identical; surfaces and lenses that converge (diverge) a parallel beam all have positive (negative) power; the transition to matrix manipulation is simple.
Cartesian sign convention
Figure 3.1 introduces the Cartesian sign convention that will be used systematically when analysing the paths of rays incident on mirrors, refracting surfaces and lenses. The point on the mirror or refracting surface lying on the optical axis is called the pole or vertex. In the case of that useful ideal, the thin lens having zero thickness, the pole is at the centre of the lens. This point is taken as the origin of Cartesian coordinates, and the object distance, image distance and radius of curvature are all measured from the pole. The positive direction along the optical axis is always the direction light is travelling. Thus the object distance is positive if it is along the direction of the incident light, while the image distance and radius of curvature are positive if they are in the direction of the outgoing light. Distances upward from the optical axis are positive. Most optical systems consist of several mirrors, refracting surfaces or thin lenses. When calculating the reﬂection or refraction at each such element the coordinates for that element must be used.
3.2
Sometimes it is useful to distinguish real from virtual images or objects. In ﬁgure 2.7 the object is real but the image is virtual. The actual rays start from real objects and pass through real images. Objects and images are virtual when only the extensions of the actual rays pass through them.
3.2
Spherical mirrors
Ray paths after reﬂection from spherical mirrors are deduced from the reﬂection laws of Chapter 2. The reﬂected ray lies in the plane deﬁned by the incident ray and the normal to the surface at the point of contact, and it makes an angle to the normal equal to that made by the incident ray. Figure 3.2 shows the ray construction used for a concave mirror
X
θ
O
α
C
β
I
h
θ
γ
P u r v
Fig. 3.2 Reﬂection at a spherical mirror with centre of curvature at C. O is the object and I the image. A representative ray from O is reﬂected from the mirror at X toward I.
forming the image of an object at O a distance u from the mirror. The mirror has radius of curvature r and its centre of curvature is at C. A representative ray is reﬂected from the mirror at X a height h above the optical axis and travels to cut the axis at I a distance v from the mirror. The angles are drawn large here for clarity: the rays considered are paraxial so that the angles are in fact very small. We will show that all paraxial rays reﬂected from the mirror converge on this same point, proving that I is the image of O. Applying the law of reﬂection at X the angles marked θ are equal. Note that with the Cartesian convention u is measured opposite in direction to the incoming light so it is negative; r and v are measured in the direction of the outgoing light so both are positive. The labelled angles
Spherical mirrors 47
48 Spherical mirrors and lenses
are given in the paraxial approximation by α = −h/u; β = h/r; γ = h/v.
(3.1)
From triangles OCX and OXI respectively β = α + θ, and γ = α + 2θ.
(3.2)
Eliminating θ we obtain 2 β = γ + α.
(3.3)
Substituting for the angles and cancelling h gives an equation independent of the choice of ray 1/v = 1/u + 2/r,
which is the paraxial imaging equation for a spherical mirror. Now consider light coming from a distant object lying on the optical axis. It arrives as beam parallel to the axis. After reﬂection this beam converges at the focus or focal point, F, as shown by the solid arrowed lines in ﬁgure 3.3. The plane perpendicular to the optic axis through the focus is called the focal plane. Substituting in the above equation shows that the distance from the mirror to the focal plane is r/2, which is called the focal length, f , of the mirror. It locates the focus half way between the mirror and its centre of curvature. The focal length is thus positive for a converging mirror, that is one concave toward the incoming light: it is negative for a diverging, convex mirror. Then rewriting the above equation 1/v = 1/u + 1/f. (3.5)
α
α
F
α
I2
f
Fig. 3.3 F is the image of a distant point object lying on the optical axis. I2 is the image of a similar distant point lying oﬀaxis. Mirror
Object
Image
(3.4)
θ θ v
u
Fig. 3.4 Transverse or lateral magniﬁcation of a ﬁnite width object.
The power of the mirror is deﬁned as P ≡ 1/f . The mirror equation, eqn. 3.5, applies equally to concave as well as convex mirrors and for all locations of the object; whether it is in front of the mirror, or virtual and behind the mirror. Figure 3.3 shows images formed by light from two distant ojects, such as stars: one on the optical axis, the other oﬀ axis at an angle α. The incoming parallel rays from the onaxis star are shown as solid lines in the ﬁgure; while the incoming parallel rays from the second star are drawn as broken lines. Each set converges to an image point in the focal plane: for the onaxis star at F, and for the star oﬀaxis at I2 . The angular separation α of the stars determines the separation of their images in the focal plane, and for paraxial rays this separation is simply, s = f α.
(3.6)
An extended plane object at a distance u from a mirror will produce a plane image at a distance v provided that the rays from all parts of the object are paraxial. This object will in general be of a diﬀerent size to
3.2
Spherical mirrors 49
the object, as illustrated in ﬁgure 3.4 for the case of a concave mirror. The transverse or lateral magniﬁcation is mt = v/u.
(3.7) Concave mirror
Not only will the image diﬀer from the object in its lateral size, but also in its length along the axial direction. Consider therefore an object in the form of a rod of length ∆u placed along the optical axis at a distance u from the mirror. Diﬀerentiating the mirror equation, eqn. 3.5 relates the image length along the optical axis ∆v to ∆u, C
(1/v 2 )∆v = (1/u2 )∆u. The longitudinal magniﬁcation is deﬁned as ml = ∆v/∆u = (v/u)2 .
(3.8)
F O
I
Fig. 3.5 Ray tracing for an object O in front of a concave mirror with centre of curvature C and focal point F. The image is formed at I. Rays are solid lines, construction lines are broken.
If the object in ﬁgure 3.4 moves toward the mirror ∆u is positive. Then ∆v is also positive and the image moves away from the mirror.
3.2.1
Ray tracing for mirrors
Ray diagrams are helpful in visualizing the formation of images by mirrors and lenses. The three most useful rays are shown in ﬁgures 3.5 and 3.6 for concave and convex mirrors respectively. Actual ray paths are drawn in solid lines, while the construction lines described below are broken. • A ray from the object towards the focus in the case of ﬁgure 3.6 and away from the focus in ﬁgure 3.5. After reﬂection this ray travels parallel to the optical axis. • A ray from the object going parallel to the axis. After reﬂection from the mirror it travels through the focal point, as in ﬁgure 3.5, or away from the focal point as in ﬁgure 3.6. • A ray from the object pointing toward the centre of curvature as in ﬁgure 3.6 or directly away from the centre of curvature as in ﬁgure 3.5. After reﬂection this ray will travel back along its own path because it is travelling radially with respect to the mirror. Having three (or more) rays permits a simple crosscheck on the accuracy of the ray tracing: they must intersect at a single image point. All the rays from the object will, after reﬂection, either converge towards a real image; or in the case of a virtual image, as in ﬁgure 3.6, the reﬂected rays diverge from the image. The concave mirror example shows how an enlarged image is formed of a person’s face placed nearer than the focal point. The convex mirror gives a wide ﬁeld of view and a demagniﬁed image, which are features useful for surveillance.
Convex mirror
O
I
F
C
Fig. 3.6 Ray tracing for an object O in front of a convex mirror centre of curvature C and focal point F. The image is formed at I. Rays are solid lines, construction lines are broken.
50 Spherical mirrors and lenses
n1
n 2 ( > n1 ) X θ1
h
θ2 γ
β
α
O
P
C
I
v
u r
Fig. 3.7 Refraction at a spherical surface. A ray from the object O on the optical axis is refracted at X towards the image I. The radius from the centre of curvature, C, through X is shown as a broken line.
3.3
Refraction at a spherical interface
In this section an equation is found for the position of the image when an object is viewed through a spherical refracting surface. Lenses rely for their focusing power on the refraction at a pair of spherical interfaces so the analysis presented here will prepare the ground for deriving the thin lens equation. As noted earlier we make the paraxial approximation: all rays are at small angles to the optical axis and any distances oﬀaxis are also small. Figure 3.7 shows the path of a paraxial ray from an object, O, on axis in a medium of refractive index n1 , which is refracted at a spherical interface into a second medium of refractive index n2 at X. Then the ray cuts the optical axis at I. The surface has radius r and we choose n2 > n1 . Applying Snell’s law to the refraction at X, and remembering that all the angles are small, we have n 1 θ1 = n 2 θ2 .
(3.9)
For the angles shown, and recalling that according to the Cartesian sign convention u will have a negative value α = −h/u, β = h/r, γ = h/v.
(3.10)
From triangles OXC and CXI respectively, θ1 = α + β, and θ2 = β − γ. Substituting for θ1 and θ2 in eqn. 3.9 n1 (α + β) = n2 (β − γ). Rearranging this gives n1 α + n2 γ = (n2 − n1 )β.
(3.11)
3.4
Now substituting for the angles using eqn. 3.10 gives n2 h/v − n1 h/u = (n2 − n1 )h/r. Cancelling h gives the imaging equation for a spherical refracting surface in the paraxial approximation n2 /v − n1 /u = (n2 − n1 )/r.
(3.12)
P = (n2 − n1 )/r
(3.13)
The quantity is known as the bending power or simply the power of the refracting surface.
3.4
Thin lens equation
Figure 3.8 shows a standard thin lens with spherical refracting surfaces. The lens is assumed to be immersed in one medium of refractive index n1 while the lens itself has refractive index n2 . Applying eqn. 3.12 to
n1
n1
n2 I Object
I’ Final image v’
u
v = u’
Fig. 3.8 Formation of an image of an object on the optical axis of a thin lens. The intermediate image produced by refraction at the ﬁrst lens’ surface is I. This is imaged by refraction at the second surface at I .
the refraction at the ﬁrst surface n2 /v − n1 /u = (n2 − n1 )/r1 ,
(3.14)
where r1 is the radius of curvature of the ﬁrst surface. Although the rays do travel a ﬁnite distance inside the lens this is neglected in what is called the thin lens approximation. Then the object distance of I from the second surface is just the image distance of I from the ﬁrst surface u = v.
Thin lens equation 51
52 Spherical mirrors and lenses
Applying eqn. 3.12 now to the refraction at the second surface n1 /v − n2 /u = (n1 − n2 )/r2 ,
(3.15)
where r2 is the radius of the second lens surface, and v is the distance of the ﬁnal image, I , from the lens. Adding eqns. 3.14 and 3.15 gives F2 f
Fig. 3.9 Image formed by a positive lens with light incident from an object at inﬁnity.
n1 /v − n1 /u = (n2 − n1 )/(1/r1 − 1/r2 ). Here we drop the prime on the ﬁnal image distance to obtain the paraxial thin lens equation n1 /v − n1 /u = (n2 − n1 )/(1/r1 − 1/r2 ) ≡ 1/f.
(3.16)
If the lens is in air the thin lens equation becomes 1/v − 1/u = (n − 1)/(1/r1 − 1/r2 ) ≡ 1/f, F2 f
Fig. 3.10 Image formed by a negative lens with light incident from an object at inﬁnity.
(3.17)
where u is the object distance, v is the ﬁnal image distance and n is the refractive index of the lens. Recalling the analysis of the mirror we can recognize that the quantity f deﬁned above is the lens focal length. The bending power of a thin lens is the sum of the powers of the two surfaces F = P1 + P2 = (n − 1)(1/r1 − 1/r2 ) ≡ 1/f.
(3.18) (3.19)
If an object is located at inﬁnity then its image is at the focus of the lens and vice versa. The situation for the lens is symmetric: eqn. 3.16 gives the same focal length f whichever lens surface is facing the incoming light. Reversing the lens exchanges the radii, r1 ⇒ −r2 and r2 ⇒ −r1 , and then 1/r1 −1/r2 is unchanged. Figure 3.9 shows how light from a distant point on axis gives an image at a distance f from the lens. Biconvex Planoconvex
Negative meniscus
Positive meniscus
Planoconcave Biconcave
Fig. 3.11 The range of lens’ shapes. Lenses in the top row are all converging lenses and those in the bottom row are all diverging lenses.
All the lenses discussed so far have positive focal length f and bring a parallel beam to a real focus, in the language of Section 3.1.1. Having two convex faces they are called biconvex lenses. Both surfaces are positive because each would alone converge a parallel beam: referring to eqn. 3.13 P is positive for both surfaces. By contrast in ﬁgure 3.10 a biconcave lens is drawn, each of whose surfaces would diverge a parallel beam and have negative power. Then eqn. 3.17 shows that the focal length f is negative. After passing through the lens an incident parallel beam diverges so that, as indicated by the dotted construction lines in ﬁgure 3.10, it appears to come from a focal point in the incident medium. The focal point is virtual. In the case that one surface is convex and the other concave the net focusing depends on the relative curvature of the surfaces. If the positive (convex) surface is more strongly curved then the lens is convergent and has positive focal length. It is called a positive meniscus lens. Conversely if the concave surface is more strongly curved then the lens is overall divergent: it is a negative meniscus lens. Some
3.4
Thin lens equation 53
examples of lens types are drawn in ﬁgure 3.11. Proceeding from left to right along the upper row, and then along the lower row the lens power steadily decreases. A plane sided sheet of glass marks the boundary between positive and negative lenses: it has inﬁnite focal length and zero power.
Positive lens F2
α
3.4.1
O
Ray tracing for lenses
Three rays useful in visualizing the formation of images by a thin lens are shown in ﬁgure 3.12 for a positive lens and in ﬁgure 3.13 for a negative lens. The rays are: • A ray from the object through the pole (centre) of the lens. This ray emerges undeviated because at the centre of the lens its faces are parallel. The displacement sideways is negligible because the lens is thin. • A ray from the object travelling parallel to the optical axis. On leaving the lens it travels towards the second focal point F2 in the case of a positive lens, and away from F2 in the case of a negative lens. • A ray through the ﬁrst focal point F1 for a positive lens or towards F1 for a negative lens. After the lens it travels parallel to the optical axis. Notice that for the negative lens the focal points (F1 and F2 ) have switched sides because the focal length is negative. The rays from the object in ﬁgure 3.12 converge after passing through the lens to a real image. In the case of the negative lens in ﬁgure 3.13 the rays diverge after passing through the lens so that they appear to come from a virtual image behind the lens. The distances from the lenses have been chosen to be u = −2f  in each case. Applying the thin lens equation 3.17 gives v = 2f for the positive lens and v = 2f /3 for the negative lens example. The reader may like to construct examples where a positive lens produces a virtual image or a negative lens produces a real image. The lateral magniﬁcation produced by a thin lens can be calculated by considering the ray through the lens’ pole in either ﬁgure 3.12 or 3.13. The size of the object is −u tan θ and of the image −v tan θ, where θ is the angle the ray makes with the optical axis. Hence the transverse magniﬁcation is mt = v/u. (3.20) When a short object is put along the optical axis the magniﬁcation can be obtained by diﬀerentiating eqn. 3.17: −(1/v 2 )∆v + (1/u2 )∆u = 0, so that longitudinal magniﬁcation is ml = (v/u)2 . If the object moves to the right, then the image moves right too.
(3.21)
I
α
F1
u
v f
f
Fig. 3.12 Ray tracing through a positive lens. The object at O produces a real image at I.
Negative lens
O
F2
I
F1 v
u
f f
Fig. 3.13 Ray tracing through a negative lens. The object at O produces a virtual image at I.
54 Spherical mirrors and lenses
Example 3.1 Two lenses are separated by 10 cm, the ﬁrst of focal length +20 cm, the second of focal length −20 cm. An object is placed 40 cm in front of the positive lens. Where is the ﬁnal image? What is its size? Is it upright or inverted? Use ray tracing to check the answer.
Image F2 Object
F1
First step
F1 Image
F2 Object Second step
Fig. 3.14 Steps in dealing with image formation by a sequence of lenses. In the ﬁrst step the second lens is ignored, and in the second step the ﬁrst lens is ignored. The ﬁrst and second focal points are labelled for each step.
The action of each lens is treated in its turn, ignoring the other one. These two steps are illustrated in ﬁgure 3.14, where the lens with the broken outline is the inactive one. For the ﬁrst lens, in eqn. 3.17 u = −40, f = +20, so that 1/v = −1/40 + 1/20 = 1/40. Hence the image distance is +40 cm. Now this image lies 30 cm to the right of the negative lens. The rays are refracted before they get to this image point by the negative lens, so this is a virtual object for the second lens. The values to be inserted in eqn. 3.17 when calculating the ﬁnal image produced by the negative lens are u = +30, f = −20, so that 1/v = +1/30 − 1/20 = −1/60.
3.5
Magniﬁers 55
Hence the ﬁnal image distance is located 60 cm to the left of the negative lens, and so it is virtual. Rays emerge from the negative lens pointing back to this location. The image magniﬁcation is the product of the individual lens magniﬁcations. Thus m = (v/u)+lens (v/u)−lens = [+40/(−40)][−60/(+30)] = +2, hence the image is twice as big as the original object and upright. Figure 3.14 shows the ray constructions required following the method given in Section 3.4.1. The upper diagram shows the construction for the positive lens; the lower for the negative lens. In the lower diagram the actual ray paths are drawn with full lines and their extensions to the object and image are drawn in broken lines. Notice that the rays used for the construction are generally diﬀerent for the two lenses.
3.5
Magniﬁers
A single positive lens or loup is used by electronic technicians to improve on the eye’s ability to resolve detail when wiring circuit boards. In the lower half of ﬁgure 3.15 an object is placed at the near point which is the point closest to the eye at which one can focus comfortably. For young people the near point dnear is about 25 cm from the eye. In the upper diagram a positive lens is placed in front of the eye, and now the eye is relaxed, that is to say it is focused at inﬁnity. The factor increase in the image size produced by using the lens is Eye
m = tan α/ tan β = dnear /fe ,
(3.22)
where dnear is the distance to the near point. The magniﬁcation m cannot be increased indeﬁnitely. A limit is reached when the lens’ surfaces become suﬃciently curved that departures from paraxial imaging become substantial. Magniﬁcation beyond a factor of about 10 with good quality imaging requires the use of a compound microscope containing two or more lenses.
3.6
Matrix methods for paraxial optics
The analysis of optical systems with several optical components is best handled using matrix methods. All refractions on the path of a ray are similar to that occuring at X in ﬁgure 3.16 for a surface of radius of curvature r. The paraxial approximation applies so that: φ = y/r,
(3.23)
α Focus
α Retina
fe
Eye β
Retina
Near point distance
Fig. 3.15 In the lower diagram the eye views an object located at the near point. In the upper diagram the same object is viewed through a positive lens, of focal length fe , with the eye relaxed.
56 Spherical mirrors and lenses
n1
n 2 ( > n 1)
X θ1 α 1
α2 φ
θ2
y
φ C
r Fig. 3.16 Incident and refracted ray at an interface at a typical location in a compound lens system.
and Snell’s law gives n 2 θ2 = n 1 θ1 .
(3.24)
The θ angles are related to the tilts of the incoming and outgoing rays respectively θ2 = α2 + y/r, θ1 = α1 + y/r. (3.25) Substituting for θ1 and θ2 in eqn. 3.24 yields n2 α2 + n2 y/r = n1 α1 + n1 y/r, so or
n2 α2 = n1 α1 − (n2 − n1 )y/r, n2 α2 = n1 α1 − P y,
(3.26)
where P is the power of the surface (n2 − n1 )/r. The lateral position after refraction is unchanged y2 = y 1 .
(3.27)
Equations 3.26 and 3.27 tell us what happens to position and angle at an interface. The analoguous equations for a ray travelling in one and the same material between two interfaces a distance l apart n2 = n1 and α2 = α2 so that n2 α2 = n1 α1 , and y2 = y1 + n1 α1 t,
(3.28) (3.29)
where t = l/n1 . This looks unnecessarily complicated, but brings the advantage that the eﬀect of free propagation in a medium and refraction at a surface can each be expressed by a matrix operating on the same column vector n2 α2 n1 α1 =M . (3.30) y2 y1
3.6
Matrix methods for paraxial optics
57
In the case of refraction, Mr =
1 0
−P 1
Actual path
,
(3.31)
and for travel in one medium, Mt =
1 0 t 1
.
(3.32)
In order that reﬂections can be included simply in the matrix analysis the light path must be unfolded at each reﬂection as shown in ﬁgure 3.17. This shows the actual path of a ray in the upper panel and the corresponding unfolded path in the lower panel. In the lower panel we see that each refractive surface crossed after the reﬂection is placed at its its mirror image in a plane through the mirror’s pole. Then the matrix for a reﬂection is 1 −P Mm = . (3.33) 0 1 None of these three matrices depend on n1 or y1 . The matrix that describes an optical system will be the product of a sequence of such matrices Mt1 , Mr1 , Mt2 .... Mtn S = Mtn ....Mt2 Mr1 Mt1 .
(3.34)
Mr , Mm and Mt have unit determinant, so any overall product matrix S will also have unit determinant: a b , (3.35) S= c d with ad − bc = 1. In the case of a thin lens with surfaces of power P1 and P2 1 −P1 1 −P2 STL = 0 1 0 1 1 −F , = 0 1
(3.36)
where F = P1 + P2 ≡ 1/f is the power of the lens. The techniques developed above are useful in calculations of the properties of lens systems. As an example take the case of a pair of thin lenses, of power F1 and F2 which are separated by a distance t in air. The overall matrix is 1 −F1 1 0 1 −F2 S= t 1 0 1 0 1 1 − tF2 −F1 − F2 + tF1 F2 = . (3.37) t 1 − tF1
Lens A
Mirror
Lens A
Unfolded path
Fig. 3.17 Ray paths at reﬂection: in the upper panel the actual path, in the lower panel the unfolded path.
58 Spherical mirrors and lenses
The total power of the lenses is thus −F = −F1 − F2 + tF1 F2 .
(3.38)
Rewriting this in terms of focal lengths, and reversing the signs, gives 1/f = 1/f1 + 1/f2 − t/f1 f2 . F1
P1
P2
F2
3.6.1 α O
f u
(3.39)
I
α
f v
Fig. 3.18 Equivalent thin lens for a compound lens system. F1 and F2 mark the focal planes. P1 and P2 mark the principal planes, and also the nodal planes when the exit and entrance media are identical.
The equivalent thin lens
In the paraxial approximation any system of lenses can be replaced by an equivalent thin lens. The result is proved formally in Appendix B using matrix methods. Figure 3.18 shows the location of the cardinal planes of the equivalent thin lens: the cardinal planes are the focal planes, the principal planes and the nodal planes. The corresponding cardinal points are located where the optical axis cuts each such cardinal plane. P1 and P2 mark the two principal planes. In the case of a single thin lens the principal planes would simply coincide. The focal length, f , of the equivalent lens is measured from the principal plane to the focal plane. Ray tracing in ﬁgure 3.18 resembles that for a single thin lens. A ray striking the ﬁrst principal plane at height h will emerge from the second principal plane at precisely the same height h. The two principal planes are therefore planes of unit magniﬁcation. When, as here, the same medium ﬁlls the object and image spaces the principal points are also points of unit angular magniﬁcation, that is they are nodal points. A ray passing through the ﬁrst nodal point at an angle α to the optical axis will emerge from the second nodal point at exactly the same angle α. If on the other hand the image and object media are diﬀerent the principal and nodal planes do not generally coincide. It is at ﬁrst sight surprising that the cardinal planes may lie inside or outside the region occupied by the lenses; the principal planes may even cross over, that is to say the ﬁrst principal plane can lie to the right of the second principal plane. In this case the right hand half of the diagram in ﬁgure 3.18 would be moved bodily to the left so that it overlapped the left half. The prescriptions given above for ray tracing remain valid in all such cases.
3.7
Aberrations
Monochromatic aberrations are the departures from paraxial imaging that appear in practical optical instruments: point objects are no longer imaged as point images. In addition there are chromatic aberrations which arise because the refractive index of glasses varies with the wavelength of the radiation, so that in turn the power of a surface given by eqn. 3.13 varies with wavelength. Thus a point object emitting, or illuminated by white light is imaged in the diﬀerent colours at different points. Chromatic aberration eﬀects occur for paraxial as well as nonparaxial rays, but they do not aﬀect pure mirror systems. The
3.7
Object plane
O
Lens/pupil
θ ρ Optic axis x
δx h y
δy Image plane
Fig. 3.19 Ray path from an object through the lens and continuing to the image plane.
chromatic and monochromatic aberrations for a typical 2.5 cm diameter crown glass lens of focal length 10 cm are similar in magnitude, ∼ 1 mm at the focal plane, so that there is equal interest in reducing both types of aberration.
3.7.1
Monochromatic aberrations
Paraxial imaging is based on the approximation that sines of ray angles can be approximated by the angles. If the angle is large then more terms need to be included in the expansion of sin θ, sin θ = θ − θ3 /3! + θ5 /5! ....
(3.40)
Aberrations that arise from neglecting the second (third) term in the expansion are called thirdorder (ﬁfthorder) aberrations. The thirdorder aberrations are dominant and are of ﬁve distinctive types. These thirdorder aberration and their reduction by suitable lens combinations will be considered in the following sections. Figure 3.19 shows a schematic ray path from the object O which emerges from the lens (or exit pupil of a lens system) and intersects the image plane some distance from the paraxial image point (starred). Let the point on the lens have polar coordinates (ρ, θ), and the image point the Cartesian coordinates (h + δx, δy), while the paraxial image point is (h, 0). In the paraxial approximation the wavefront leaving the lens would be spherical with its centre at the image point. Actual wavefronts deviate from this shape and this is the origin of aberrations. The deviation of the actual from
Aberrations 59
60 Spherical mirrors and lenses
the ideal wavefront, which is for simplicity measured radially in the direction of the image, can only depend on two things: where the ray hits the lens relative to the optical axis, ρ; and where the image lies in the image plane relative to the optical axis, h. Hence the deviation is some function of these vectors ∆R = ∆R(ρ, h). In the case of system of lenses the lens aperture indicated in ﬁgure 3.19 is replaced by the exit pupil of the system, deﬁned as follows. Each lens system has one aperture which is the most restrictive in limiting the angular size of the cone of rays from an object point which actually reach its image point. This limiting aperture is known as the aperture stop. Its image seen from the image space is called the exit pupil, and its image seen from the object space is called the entrance pupil. The ray passing through the centre of the aperture stop is called the chief or principal ray. This ray also passes through the centre of the entrance and exit pupils. Rays that pass through the edge of the aperture stop, and the entrance and exit pupils, are known as marginal rays. If the system shown in ﬁgure 3.19 is rotated about the optical axis, then because the system is axial symmetric the aberration should not change. It follows that ∆R can only depend on rotational invariant quantities made up from ρ and h. There are just three of these: ρ2 , h2 and ρ · h = ρ h cos θ. Then rewriting ∆R to include all possible quadratic and quartic terms made up from these quantities gives ∆R = a1 ρ2 + a2 ρ h cos θ + a3 h2 + b1 ρ4 + b2 ρ3 h cos θ +b3 ρ2 h2 + b4 ρ2 h2 cos2 θ + b5 ρ h3 cos θ + ....
(3.41)
The quadratic terms with the a coeﬃcients can be dropped, because they give overall movements of the image for paraxial and nonparaxial rays alike. Thus the term in ρ2 gives focusing and moves the image plane for all rays; the term in ρ h cos θ tilts all rays to make the image larger/smaller; while the term in h2 is a constant over the whole aperture. The remaining quartic terms with b coeﬃcients contain the thirdorder aberrations. It is straightforward to extract the image displacements from ∆R, the displacement of the wavefront from a sphere centred on the image point. The Cartesian coordinates at the exit pupil are x = ρ cos θ and y = ρ sin θ. The tilt of the wavefront in the x and ydirections are (∂∆R/∂x) and (∂∆R/∂y). Multiplying by the distance, v, from the lens (or exit pupil), converts these tilts into displacements at the image plane δx = v (∂∆R/∂x) and δy = v (∂∆R/∂y).
(3.42)
3.7
Then using ∂ρ/∂x = cos θ, ∂ρ/∂y = sin θ, ∂θ/∂x = − sin θ/ρ and ∂θ/∂y = cos θ/ρ these equations become
Aberrations 61
Lens
FM
FP
δx = v cos θ (∂∆R/∂ρ) − v (sin θ/ρ) (∂∆R/∂θ), and δy = v sin θ (∂∆R/∂ρ) + v (cos θ/ρ) (∂∆R/∂θ). After some manipulation this yields δx = c1 ρ3 cos θ + c2 (2 + cos 2θ)ρ2 h + (3c3 + c4 )ρ h2 cos θ + c5 h3 , (3.43) δy = c1 ρ3 sin θ + c2 ρ2 h sin 2θ + (c3 + c4 ) ρ h2 sin θ.
(3.44)
f
The ﬁve aberrations separated here with coeﬃcients cn are physically quite distinct and are called the Seidel aberrations: • • • • •
spherical aberration (c1 = 4vb1 ), coma (c2 = vb2 ), astigmatism (c3 = vb4 ), ﬁeld curvature (c4 = v(2b3 − b4 )), distortion (c5 = vb5 ).
Fig. 3.20 Spherical aberration for a positive lens. The paraxial focus is FP and the marginal ray focus is FM . The broken line indicates the plane of the circle of least confusion.
von Seidel developed the analytic study of aberrations in the 19th century after being engaged by the optical entrepreneur Zeiss to examine ways of improving lens performance.
3.7.2
Spherical aberration
The ﬁrst terms in eqns. 3.43 and 3.44 are the only ones that contribute to aberrations of objects on axis, δx = c1 ρ3 cos θ, δy = c1 ρ3 sin θ. These relations show that the intersections with the paraxial image plane of the rays passing through an annular section at radius ρ of the exit pupil lie on a circle. The radius of this circle increases with ρ so that the image becomes a circular blob. This spherical aberration is illustrated in ﬁgure 3.20 for an axial object at inﬁnity. Positive lenses focus marginal rays more strongly than paraxial rays. The paraxial rays focus at FP , while the marginal rays through the periphery of the exit pupil come to a focus at FM . At the surface indicated by the dotted line lies the circle of least confusion where the spread of the rays is minimum. The distance FM FP is the longitudinal spherical aberration, and the radius of the ray cone at the paraxial focus is the transverse spherical aberration. Figure 3.21 shows how a lens’ surfaces can be bent while keeping the focal length ﬁxed. The spherical aberration depends strongly on the lens shape and is smallest when the shape is close to being planoconvex, with the curved surface facing the light. In this conﬁguration, with a parallel beam incident, the deviation of rays is shared equally between the two lens surfaces.
Fig. 3.21 Bending of the surface curvature of positive lenses, while keeping the focal length constant.
62 Spherical mirrors and lenses
For certain combinations of object and image points spherical aberration is absent and these conjugate points are used in microscope design. In ﬁgure 3.22 P is a point on the surface of a planoconvex lens of refractive index n, such that its distance from the centre of curvature, C, of the convex face, radius r, is exactly r/n. Then we can show that rays from P passing through the lens at any angle converge at a point image P where P C is n · r. P CT and TCP are similar triangles because P C ˆ is common. Hence = n· TC, CT = n· CP and the angle PCT T Pˆ C = α. Also
P TˆC = α .
α’ R T
α P’
P
Applying the sine rule to ∆TP C sin (T Pˆ C)/r = sin (P TˆC)/n · r.
C r
Multiplying by n r we get n sin α = sin α .
r/n n.r
Fig. 3.22 Conjugate points (P and P ) for a spherical refracting surface whose centre lies at C. CP = r/n, and CP = n × r, where r is the radius and n the refractive index.
This is exactly Snell’s law so that TR is precisely the path of ray PT after refraction at T. This result holds for any ray from P which means that the image P is free of spherical aberration. P and P are known as conjugate points for the curved refracting surface. In high magniﬁcation microscopes the object is located at one conjugate point of the ﬁrst lens so that spherical aberration is avoided while maximizing the numerical aperture.
3.7.3
Coma
Coma is familiar to anyone who has used a positive lens to focus the Sun’s image on paper. When the lens is tilted the Sun’s round image changes to a comet shaped ﬂare. It is the following terms in eqns. 3.43 and 3.44 that cause coma δx = c2 (2 + cos 2θ)ρ2 h δy = c2 ρ2 h sin 2θ. One term, 2c2 ρ2 h, gives a simple radial displacement of the image point, whilst the sinusoidal terms cause the intersection of the rays with the paraxial image plane to travel round a circle as θ varies from 0 to π and again as θ increases from π to 2π. This is illustrated in ﬁgure 3.23 on which the chief ray follows the arrowed line. Rays travelling through an annular section of the exit pupil (ﬁxed ρ) intersect the paraxial image plane in a ring as indicated by the letters a, b, c, d. The sizes of the rings and their radial oﬀsets grow with the radius squared (ρ2 ) of the annular section of the lens through which the rays pass. The circles overlap to give a comet shaped ﬂare which may extend, as in ﬁgure 3.23, outward
3.7
Aberrations 63
a d
b
c
Image
c
plane a b
b
d Exit pupil
d c
a Optic axis
Fig. 3.23 Coma for an oﬀaxis point object. Rays passing through a ring on the lens, like the ones shown, intersect the image plane on a ring. The total image is a comet shaped ﬂare that starts at the paraxial image point and has a spread of 60 ◦ .
from the paraxial image, when it is called positive coma; or inward, which is known as negative coma. Coma also varies with lens’ shape. It is smallest for a positive lens close to planoconvex in shape, with the object far distant and the light incident on the curved face. This is very similar to the lens shape that also minimizes spherical aberration. A neat arrangement of two planoconvex lenses which eliminates coma when the object is not an inﬁnity is shown in ﬁgure 3.24. The ﬁrst lens produces an image at inﬁnity, and the second refocusses the light and then the condition for reduced coma and spherical aberration holds for both lenses. This arrangement of biconvex lenses is often used in condensers which are optical systems for projecting light from a source onto the object viewed.
3.7.4
Astigmatism
Astigmatism is illustrated in ﬁgure 3.25, where the image of an oﬀaxis point object consists of two line segments at right angles. The image nearer the exit pupil is a tangential line, the other is a radial line, while between them lies a circle of least confusion. The tangential line image is formed by the rays which lie in the plane containing the chief ray and the optical axis. These rays in it are called tangential or meridional rays. The radial line image is formed by the rays lying in the perpendicular sagittal plane which contains the chief ray. In ﬁgure 3.26 the object is shaped like a spoked wheel. The tangential image has a sharply deﬁned
I O
f PC
f PC
Fig. 3.24 An arrangement of two planoconvex lenses which minimizes coma and spherical aberration.
64 Spherical mirrors and lenses
Lens/pupil Object
Tangential line image Radial line image Fig. 3.25 Image formation with astigmatism. The solid lines are rays in the meridional plane containing the optical axis. The broken lines are sagittal rays forming a perpendicular plane. These give tangential and radial line images respectively.
rim while the spokes are fuzzy. Conversely the sagittal image has sharply deﬁned spokes and a fuzzy rim.
Object Exit pupil
Referring to eqns. 3.43 and 3.44, the components which cause astigmatism are δx = 3c3 ρ h2 cos θ, δy = c3 ρ h2 sin θ. Tangential image Sagittal image
Fig. 3.26 Astigmatic image of a wheelshaped object centred on the optical axis. At the tangential focus radial lines blurred and at the sagittal focus the tangential lines are blurred.
The tangential rays are those having θ = 0 or π and have δy = 0. They form a point image in a tangential focal plane before reaching the paraxial image plane. On the other hand the sagittal rays with θ = π/2 or 3π/2 have δx = 0 and form a point image in a sagittal image plane one thord as far from the paraxial image plane. The tangential rays when reaching the sagittal image plane still have δy = 0, and so form a sagittal or radial line image. Similarly the sagittal rays form a tangential line image in the tangential image plane. An extreme degree of astigmatism is produced by cylindrical lenses, an example of which is shown in ﬁgure 3.27. There is focusing in the vertical plane only and the image of a distant point object is a horizontal line. Intermediate between cylindrical and spherical lenses are the toric lenses, which have diﬀerent curvatures in perpendicular planes through the optical axis.
3.7.5
Field curvature
In eqns. 3.43 and 3.44 there are a second set of terms in ρ h2 cos θ and ρ h2 sin θ which resemble the terms responsible for astigmatism: namely
3.7
Aberrations 65
those with the coeﬃcient c4 . δx = c4 ρ h2 cos θ, δy = c4 ρ h2 sin θ. Here however the coeﬃcients, c4 , are identical which makes all the difference. Following the argument made in the case of astigmatism, the axial displacement of the tangential and sagittal images from the paraxial image plane due to these terms are equal, hence there is simply a displacement of the image along the axial direction which increases quadratically with h, the distance oﬀ axis. This aberration is called ﬁeld curvature and the curved image surface is called the Petzval surface ΣP . Any astigmatism gives further displacements, with the tangential image surface (ΣT ) moving three times further from ΣP than the sagittal image surface (ΣS ). In ﬁgure 3.28 the displacements of the image planes are shown as a function of the radial distance oﬀ axis for a biconvex lens.
ΣT
ΣS ΣP
Distance off axis
Axial image displacement Fig. 3.28 Image surface curvature for a positive biconvex lens. See text for details.
In the case of a negative lens the ﬁeld curvature is away from the lens so that by a judicious combination of lenses the ﬁeld curvature can be reduced. The total displacement of the Petzval surface in an image formed by a series of thin lenses is ∆z ∝ (1/[nj · fj ]), (3.45) j
where nj is the refractive index, and fj the focal length of the jth lens. With a single negative and positive lens made of the same glass the
Line focus
Fig. 3.27 Cylindrical lens.
66 Spherical mirrors and lenses
ﬁeld can be made ﬂat by choosing f− = −f+ . The overall focal length can be adjusted by the choice of f+ and the lens separation: see eqn. 3.39. One widely used technique employed in lens systems to remove an unacceptable level of ﬁeld curvature is to add a ﬁeld ﬂattener, which is a negative lens placed a small distance u from the ﬁnal image plane of the system. The ﬁnal image position is given by v = uf /(f − u) ≈ u, so the image is otherwise unchanged in size.
3.7.6
Distortion
The ﬁfth and ﬁnal terms in eqns. 3.43 and 3.44 are δx = c5 h3 , δy = 0.
Fig. 3.29 Image distortion. The lower diagram shows a square object. The upper diagrams show images with barrel and pincushion distortion.
They give a purely radial displacement of the image point that increases with the cube of the distance oﬀaxis, h. If c5 is negative (positive) then points on the image are displaced radially inward (outward), the displacement growing with the distance of the object oﬀ axis. Both types of distortion are shown in ﬁgure 3.29 for the object in the lower panel. The images are said to show barrel and pincushion distortion. The distortion produced by any lens is reversed if the lens is turned around, so a pair of lens arranged symmetrically will cancel each others distortion. Aperture stop eﬀects
Fig. 3.30 The eﬀect of aperture stop placement.
The positioning of the aperture stop in a lens system signiﬁcantly affects the aberrations. Thin positive lenses do not show distortion if the stop is placed against the lens. However if, as in ﬁgure 3.30, the stop is placed elsewhere distortion results: barrel distortion if the stop is in front of the lens and pincushion distortion if the stop is behind the lens. The essence of the aperture stop’s inﬂuence, both in this example and in cases where its eﬀect is favourable, is that it limits the range of rays that form the image. An intuitive way to view this eﬀect is to note that in the upper diagram the distance rays travel to the lens is longer than that travelled by the ray through the lens’ pole, so that the object distance is larger and the magniﬁcation smaller than for the paraxial rays. Consequently barrel distortion results. In the lower diagram the object distance is shorter for rays at wide angles, the magniﬁcation is correspondingly larger and pincushion distortion results. In general the introduction of an aperture stop will only aﬀect an aberration if one of the earlier aberrations in the sequence (spherical aberration, coma, astigmatism, ﬁeld curvature, distortion) is present. The introduction of an aperture stop will aﬀect neither the spherical
3.7
Aberrations 67
Distance ΣS
off axis
ΣP
ΣT
Axial image displacement Fig. 3.31 Positive meniscus lens with an aperture stop and the resulting image surfaces. ΣP , ΣS and ΣT are the Petzval, sagittal and tangential image surfaces respectively. The surface of least confusion is close to the ordinate axis.
aberration nor the Petzval ﬁeld curvature. There is an optimal location of the stop for any given lens bending which minimizes the coma, and this is called its natural location. Bending the lens surfaces moves the image ﬁeld surface. The meniscus lens with a front aperture stop, shown in ﬁgure 3.31, gives astigmatic images, but the surface of least confusion is now ﬂat. In addition the stop also reduces coma. This lens arrangement was often used in cheap cameras. It is not possible to choose a lens shape that will minimize simultaneously all the aberrations even if the aperture stop is located at the natural location. Combining a shape close to planoconvex and having the aperture stop at the natural location minimizes spherical aberration and distortion, but the level of astigmatism is then unacceptable. With a meniscus shape the aberrations are all similar but not particularly small. Combinations of lenses are needed to give images of overall high quality.
3.7.7
Chromatic aberration
Prisms made of glass disperse white light into a spectrum; and because any segment of a lens is a prismshaped it follows that the red and blue images produced by a single lens will not coincide. This chromatic aberration in the image of a point object onaxis is shown in ﬁgure 3.32. Blue light is focused more strongly than the red because the refractive index of glasses and most other transparent substances falls oﬀ with increasing wavelength. This is known as normal dispersion. The distance between
Blue
Red
Red
Blue
Fig. 3.32 Chromatic aberration produced by a positive and a negative lens of a distant point object on axis. The circle of least confusion lies in the plane indicated by the dotted line.
68 Spherical mirrors and lenses
the red and blue images is the longitudinal chromatic aberration. At the waist in the ray envelope lies the circle of least confusion where the image spot is smallest. A commonly used measure of the dispersion is dispersive power ν = (nF − nC )/(nd − 1), (3.46) where the refractive indices nF , nd and nC are those of the glass measured at three wavelengths of Fraunhofer absorption lines in the Sun’s spectrum. Historically these were selected to span the visible spectrum: the Fline is blue (486 nm), the dline is yellow (589 nm) and the Cline is red (656 nm). Another measure of the dispersion is the dispersive in
Fig. 3.33 A selection of Schott glasses: refractive index plotted against the dispersive index (Abbe number). The less dispersive crown glasses lie to the left of the centre line and the more dispersive ﬂint glasses to the right. (With permission from Schott guide to Glass: courtesy Wolfgang R. Wentzel, Schott AG, Hattenbergstr. 10, 55122 Mainz, Germany.)
dex, V , which is the inverse of the dispersive power. This is often the quantity speciﬁed by glass manufacturers: be aware that V is smaller if the dispersion is larger! Figure 3.33 plots the refractive index against the dispersive power for glass types produced by Schott AG. There are two basic layouts of a pair of lenses with overall positive power that remove most of the chromatic aberration: (1) A positive and negative lens made of glass with diﬀerent dispersions which are in contact. (2) A pair of positive lenses spaced apart by the mean of their focal lengths. A lens doublet, that is a pair of lens in contact, has focal length 1/f = 1/f1 + 1/f2 .
3.7
Aberrations 69
Table 3.1 Dispersion of two standard types of crown and ﬂint glass. Crown glass is made of soda, lime and silica; ﬂint glass of alkalis, lead oxide and silica. The code number of a glass is constructed, as indicated by the boldface characters from the Dline refractive index and the dispersive index. Glass
486 nm F (blue)
589 nm D (yellow)
656 nm C (red)
Dispersive index, V
BK7 517642 DF 620364
1.5224 1.6321
1.5168 1.6200
1.5143 1.6150
64.2 36.4
Now for either lens 1/fi = (ni − 1)ρi , where ρi is the geometric factor 1/r1 − 1/r2 formed from the radii of curvature of the lens’ surfaces, and ni is the refractive index of the glass at 589 nm. Thus for each lens the diﬀerence in focal length between 486 nm and 656 nm is given by ∆(1/fi ) = ∆ni ρi = νi (ni − 1)ρi = νi /fi .
(3.47) (3.48)
Using this result the change in the focal length of the system between 486 nm and 656 nm is given by ∆(1/f ) = ν1 /f1 + ν2 /f2 .
(3.49)
This chromatic diﬀerence in focal length vanishes if ν1 f2 = −ν2 f1 .
(3.50)
Dispersive power is positive for glass, so that in order to satisfy this equation one lens needs to be converging, the other diverging. A simple convergent colour corrected system can be made by cementing a double convex crown glass lens to a planoconcave ﬂint glass lens as shown in ﬁgure 3.34. This is called a Fraunhofer achromat. Table 3.1 gives the optical parameters for two types of optical glass that are frequently used in making achromatic doublets. Crown glass has the larger dispersive power, hence from eqn. 3.50 the positive (crown glass) lens will have shorter focal length, making the focal length of the combination positive. In the case of thick doublets there is a further diﬃculty. Although the focal lengths for blue and red light have been made equal, the images do not coincide because the principal planes for the two colours are separated by a distance of order 10% of the lens thickness. What will remain true is that the magniﬁcation of objects is the same for both colours, which removes lateral chromatic aberration. In addition all achromatic
Red ray Blue ray
Crown glass
Flint glass
Fig. 3.34 Achromat doublet lens.
70 Spherical mirrors and lenses
doublets suﬀer residual chromatic aberration at other wavelengths away from 656 nm and 486 nm, a feature that is called secondary colour. The bending of the surfaces of an achromat remains at the disposal of the designer. Thus a lens shape for a given focal length can be chosen to minimize spherical aberration and coma, as well as chromatic aberration. Combinations of thin lenses made of three diﬀerent types of glass give colour correction at three wavelengths, and such lenses are called apochromats. The recent availability of expensive extralow dispersion ED glass has made the correction of chromatic aberration that much easier, but with a cost penalty. An ED lens has lower dispersion across the visible spectrum than either an achromat or apochromat. ED lenses are essential for modern very long lens systems. It is also possible to correct chromatic aberration at two wavelengths with lenses made of the same glass type. From eqn. 3.39 two lenses separated by a distance t have focal length 1/f = 1/f1 + 1/f2 − t/(f1 f2 ). Then the diﬀerence between the focal lengths at 656 nm and 486 nm is given by
Limiting ray
∆(1/f ) = −∆f1 /f12 − ∆f2 /f22 + t∆f1 /(f12 f2 ) + t∆f2 /(f22 f1 ) = −ν/f1 − ν/f2 + 2tν/(f1 f2 ). Eye
Mirror
α
θ
Centre of curvature
Thus in order to remove the chromatic aberration we must have
β=θ+α
0 = [ν/(f1 f2 )] [−f2 − f1 + 2t],
β Limiting ray
that is t = (f1 + f2 )/2.
Fig. 3.35 Angular coverage with convex mirror.
(3.51)
(3.52)
The lens separation must be half the sum of the two focal lengths.
3.8
Further reading
The seventh edition of Principles of Optics by M. Born and E. Wolf, and published by Cambridge University Press (1999) contains a thorough mathematical acccount of aberrations and many other matters. The Handbook of Optics, volume 1, contains a broad discussion of aberrations in the context of fundamentals, techniques and design. The editor in chief is M. Bass and it was published in 1995 by McGrawHill, New York.
3.8
Further reading 71
Exercises (3.1) Two lenses are separated by 25 cm, the ﬁrst of focal length −30 cm, the second of focal length +40 cm. An object is placed 10 cm in front of the negative lens. Where is the ﬁnal image? What is its size? Is it upright or inverted? Use ray tracing to check the answer. (3.2) Apply the results of Appendix B to locate the cardinal points of the system described in the previous question. (3.3) A shopkeeper uses a convex mirror of radius of curvature 1 m and arc length across the diameter of 50 cm to view his shop. When he is 2 m from the mirror what angular coverage does this give him? What angular size would a customer’s hand (10 cm) subtend at the shopkeeper’s eye if the customer is 7 m from the mirror. Figure 3.35 shows the arrangement.
order to mimimize aberrations. What aberrations are reduced by this arrangement. (3.6) Design an achromatic doublet from the glasses given in Table 3.1 having a focal length of 45 cm. If only lenses of the same glass type are available how could you reduce the chromatic aberration? (3.7) Show that the smallest separation of a real image and object is four times the focal length of the (positive) lens and that the lens is then midway between image and object. (3.8) Show that if the object is a distance x1 from the ﬁrst focus and the image is a distance x2 from the second focus of a lens of focal length f , then x1 x2 = −f 2 . This can be done analytically or by using similar triangles in ﬁgure 3.12. (3.9) A biconcave lens made from glass of refractive index 1.65 has surfaces with radii of curvature 25 cm and 45 cm. What is the focal length of the lens?
(3.4) An object is located 25 cm from a concave mirror whose radius of curvature is 40 cm. The object 0.1 cm in length lies along the optical axis. Where (3.10) Apply matrix methods to a lens of thickness T , refractive index n having surfaces with radii of curvais its image located and how long is it? Use ray ture r1 and r2 . What is the focal length and where tracing to check the image location. are the principal planes of the lens? (3.5) What radius of curvature is required to make a planoconvex lens with focal length 30 cm from glass (3.11) A lens of refractive index 1.73 has focal length 20 cm in air. What will its focal length become of refractive index 1.7? How would you arrange a when immersed in water? pair of planoconvex lenses and an aperture stop in
This page intentionally left blank
Optical instruments 4.1
4
Introduction
The refracting telescope and the microscope are the ﬁrst instruments discussed in this chapter; the one used to magnify distant objects, the other to magnify tiny objects. Camera design and construction are examined next. Several special lens systems will be described, including the telephoto, zoom and telecentric lenses. The last is vital in process control in industry. Reﬂecting telscopes will be discussed in Chapter 8. In order to give bright images optical instruments need to have large lenses that collect nonparaxial as well as paraxial rays. Another fundamental reason that aperture size is important is that the smallest detail that can be seen clearly depends critically on the aperture size: even aberrationfree optical systems produce extended images of point objects due to the wave nature of light. The plane beam of light incident on any aperture is diﬀracted, that is it spreads out on emerging from the aperture by an amount that depends on the ratio of the wavelength to the aperture size. Diﬀraction is discussed more fully in Chapter 6, where it will be shown that a circular aperture of diameter D limits the angular resolution with radiation of wavelength λ to at best ∆θ = 1.22λ/D.
(4.1)
The corresponding resolution in an image in the focal plane is then f ∆θ. Aberrations can only make the resolution worse; hence the methods for reducing aberrations need to be applied if the limit on resolution imposed by diﬀraction is to be attained. The ﬁnal sections of the chapter contain descriptions of widely used, nonstandard lenses. Aspheric lenses are slightly nonspherical lenses and simplify the removal of aberrations in modern camera lens design. Graded index lenses rely on axial or radial refractive index variation, rather than surface curvature to give focusing. Lastly Fresnel lenses are discussed, which are ﬂattened versions of normal lenses.
4.2
The refracting telescope
The simple refracting telescope design illustrated in ﬁgure 4.1 using a pair of lenses to give improved magniﬁcation over a single lens was invented early in the seventeenth century. It uses two positive lenses, a
74 Optical instruments
Objective
Eyelens
α
α
I
β
β
Exit fe
pupil
fo Fig. 4.1 Terrestrial telescope design as used in binoculars. The eye ring is located at the exit pupil, where the image of the objective is formed by the eyelens. The angular magniﬁcation is β/α.
large radius objective and a smaller eyelens spaced apart by the sum of their focal lengths. The objective produces an intermediate image I in its focal plane and the eyelens images this at inﬁnity. A telescope produces angular magniﬁcation: parallel rays from a point on an object at inﬁnity exit as parallel rays toward the image point at inﬁnity. Such optical systems are called afocal and are unique in not possessing principal planes. The magniﬁcation achieved is the angular size of the image divided by the angular size of the same object seen without the telescope: Mθ = β/α ≈ tan β/ tan α so that Mθ = fo /fe .
(4.2)
As mentioned in the previous chapter, one element in any optical instrument will restrict the amount of light from the object reaching the image more than the others. This element is known as the aperture stop. In order to determine which element is the aperture stop the image is calculated for each element in the object space. Then the element whose image subtends the smallest angle, as seen from the object, is the aperture stop. Its image is called the entrance pupil. Similarly the image of the aperture stop seen from the image side is called the exit pupil. In the case of a telescope the aperture stop is the objective lens and, because it is the ﬁrst element, it is also the entrance pupil. The exit pupil of the telescope, that is the image of the objective produced by the eyelens is at a distance vp behind the eyelens where 1/vp = −1/(fo + fe ) + 1/fe
4.2
= fo /[fe (fo + fe )]. Then vp = fe (fo + fe )/fo .
(4.3)
Now fo fe so the image is close to the focal plane of the eyelens. The exit pupil diameter is then D = Dvp /(fo + fe ) = D/Mθ .
(4.4)
According to eqn. 4.1 the lower limit imposed by diﬀraction on the telescope’s angular resolution is ∆θtel = ∆θobjective · Mθ = (1.22λ/D) (D/D ) = 1.22λ/D .
(4.5)
When used with the eye the light collected by the instrument should all enter the observer’s eye. The eye pupil is placed at the exit pupil, which is therefore called the eyering for visual observation, and ideally the eye and exit pupil sizes should be the same. Both the angular resolution and light gathering power of the eye–telescope combination are limited by whichever of the eye pupil and the telescope’s exit pupil is the smaller. If the eye pupil is the smaller then, when viewed by eye, some of the light collected by the telescope and some of its potential resolution are wasted. Alternatively if the eye pupil is the larger the lightgathering and resolution of the eye are not fully utilized. Matching the position and size of the exit and entrance pupils of coupled optical systems has to be a general design goal: in that way no component throws away the light transmitted by the others, and equally no component degrades the resolution of the others. The eye pupil diameter depends on the light intensity; from a value of 2 mm in bright light to 8 mm in near darkness. In full daylight the pupil diameter is 2.5 mm so the eye resolution at 550 nm wavelength light (green light) is ∼ 3 10−4 rad, or 1 minute of arc. Thus the eye can comfortably resolve points 0.1 mm apart on an object placed 25 cm away at the near point. In Chapter 2 numerical aperture was introduced to quantify the light collecting power of an optical ﬁbre. Here we generalize the deﬁnition: NA = n sin θ,
(4.6)
where θ is the semiangle subtended by the entrance pupil at the object lying in a medium of refractive index n.
4.2.1
Field of view
Although the eyelens in a telescope does not restrict the illumination of the image, it does restrict the angular range over which objects can be
The refracting telescope 75
76 Optical instruments
seen. This range is called the ﬁeld of view, and the eyelens is called the ﬁeld stop. Its image in the object space is called the entrance window of the system, and the exit window is the eyelens itself. Similar deﬁnitions will apply to other optical instruments, the ﬁeld stop being the aperture whose image in the object space restricts the ﬁeld of view from the entrance pupil. Suppose the eyelens has diameter d, then the apparent ﬁeld of view seen by the eye is α = d/fe .
(4.7)
The actual angular ﬁeld of view is the angle subtended by the eyelens at the objective: α = d/(fo + fe ). (4.8) Combining the last two equations
Then using eqn. 4.2
fo
f e
Fig. 4.2 Design suitable for either a Galilean telescope or a beam expander. The arrows on the rays would be reversed when used as a beam expander.
α = (fo + fe )α/fe .
(4.9)
α = (Mθ + 1)α.
(4.10)
The full ﬁeld is only seen from the centre of the exit pupil; from other points across the exit pupil a smaller angular range is visible, so the outer part of the ﬁeld of view is less well illuminated. This eﬀect is called vignetting. In order to remove the outer poorly illuminated part of the view an aperture can be placed at the intermediate image in a telescope or microscope; and this aperture then becomes the ﬁeld stop. The distance from the eyering to eyelens is called the eye relief and this needs to be at least 1.5 cm for comfortable use. The image produced by the telescope is inverted and to correct this either a further lens is needed or an inverting prism. The Porro prism described in Chapter 2 is one choice. Another type of telescope shown in ﬁgure 4.2 uses a negative power eyelens located so that the lenses are separated by the sum of their focal lengths. This design, the Galilean telescope, produces an upright image so that this design is used in opera glasses. Although not invented by Galileo he improved the performance to ×30 power by experiment and skillful lens grinding. He was then able to observe four of the moons circulating round Jupiter, the mountainous surface structure of the Moon, and found that Venus shows phases like the Moon. These observations dealt the death blow to the old cosmology in which the Sun, Moon, planets and stars were perfect spheres and themselves lay on transparent spheres that rotated around a static Earth. The lens arrangement of the Galilean telescope is often used to expand the diameter of laser beams which are inherently narrow. For this application the Galilean design has the advantage over a design using two positive lenses that there is no intermediate focus (see ﬁgure 4.1). This avoids the local heating that occurs at the focus of an intense laser beam.
4.2
Example 4.1 A binocular has an objective of focal length 16 cm and diameter 48 mm. What eyelens’ properties will give a magniﬁcation of 8 times and a 5 ◦ ﬁeld of view? What will the eye relief and eyering diameter be? What will be the apparent ﬁeld of view? The overall angular magniﬁcation is given by eqn. 4.2 so the focal length of the eyelens is fe = 16/8 = 2 cm. If the ﬁeld of view is 5◦ then using eqn. 4.8 the eyelens diameter is d = 18π(5/180) = 1.57 cm. The eye relief is the distance of the eyering from the eyelens, given by eqn. 4.3 vp = 2 × 18/16 = 2.25 cm. In addition the eyering diameter is given by eqn. 4.4 D = 48/8 = 6 mm. The apparent ﬁeld of view is given by eqn. 4.10 α = 9 × 5 = 45◦ . Aberrations would be large for the system just outlined and the single objective and eyelens are replaced by lens combinations; the eyelens replacement being called an eyepiece.
4.2.2
Etendue
In a complex optical system consisting of several components, the numerical apertures and ﬁelds of view of the components should match from component to component along the chain. That is to say the exit pupil (window) of any component should coincide in position and size with the entrance pupil (window) of the following component. If just one component has a signiﬁcantly smaller pupil (window) than its neighbours the additional light gathering power (ﬁeld of view) of the other components is simply wasted. Numerical aperture and ﬁeld of view are not however independent properties, and this permits more ﬂexibility than the bare statements above imply. It may be possible to trade ﬁeld of view for light gathering power and vice versa by inserting some combination of lenses between
The refracting telescope 77
78 Optical instruments
the unmatched components. A quantity which expresses the overall light gathering power, called the etendue, determines whether components may be matched in this way or not. The etendue is ﬁrst deﬁned and then the way the trading occurs is illustrated with a single lens.
2θ’ α’
h’
α
h 2θ
v
u
The solid angle Ω subtended by the entrance pupil at the object measures the light gathering power, that is the fan of rays received from any point on the object. Equally the unobstructed area A of the object visible in the image deﬁnes the ﬁeld of view. The product of these two quantities and the refractive index squared T = n2 ΩA
Fig. 4.3 Etendue through one optical component.
(4.11)
is variously known as the throughput, luminosity or etendue. Another useful expression for the etendue can be obtained by reexpressing the solid angle in terms of the numerical aperture, NA = n sin θ. Taking the semiangle θ subtended by the entrance pupil to be suﬃciently small, we have to a good approximation, Ω = 2π(1 − cos θ) = πθ 2 ,
(4.12)
T = π(NA)2 A.
(4.13)
and then Taking the product of the etendue and the radiance gives the total radiant ﬂux in the beam, so the etendue gives a measure of the light that an optical system can transmit. Suppose the incident radiance on a optical system over the wavelength interval ∆λ at wavelength λ is I(λ)∆λ, then the ﬂux of radiation through the system in this wavelength interval is F = T I(λ)∆λ.
Field lens
Eye lens Exit pupil
(4.14)
Figure 4.3 shows a single lens placed between media of refractive indices n and n . The object and image heights are h and h , while their areas are h2 and (h )2 respectively. The cone of rays accepted by the entrance pupil from any point on the object has semiangle θ, and these rays converge in a cone of semiangle θ at the image point. In the paraxial approximation the ratios are −h /h = v α /(u α) = v n/(u n ), and θ /θ = u/v. Therefore
Field stop
Eye relief
Fig. 4.4 Huygens’ eyepiece.
(n )2 Ω A = (n )2 π(θ )2 (h )2 = n2 πθ2 h2 = n2 Ω A. Thus at each intermediate image through an optical system the image size and the angular spread of the fan of rays illuminating a point on the image will both change but the product, the etendue, is seen to be invariant. More precisely, the etendue of the light that makes its way
4.3
Telescope objectives and eyepieces 79
through the whole of an optical system is invariant at each step: we cannot include light cut oﬀ by an aperture at some intermediate surface. It is always possible in principle to match two optical components with the same etendue by using intermediate lenses. However this may be impractical if the diﬀerence in pupil sizes is very large. Some related conclusions can be drawn about the image brightness. Here brightness is always used to mean the radiance or the luminance, as deﬁned in Section 1.7. Sticking to energy parameters, we see that the invariance of the etendue means that radiance or brightness is preserved. This result is known as the law of conservation of radiance. There is one special case: that of a point source, for example a distant unresolved star. With such an object there is no area so that the etendue cannot even be deﬁned. Clearly the bigger the telescope aperture stop (the objective) used to collect light from a point source the better.
4.3
Telescope objectives and eyepieces
Generally the ﬁeld of view needed in a telescope is narrow so that an achromatic doublet or apochromatic triplet is adequate as the objective. Among the aberrations spherical aberration and coma can be kept small by appropriate lens shaping. Astigmatism, ﬁeld curvature and distortion are less important over narrow ﬁelds of view. In the case that a wide ﬁeld is needed, one of the camera lenses described below would be used. Eyepieces on the other hand are required to give good image quality over the apparent ﬁeld of view, which is wider in angle than the ﬁeld of view by a factor equal to the angular magniﬁcation of the telescope. Three popular examples of eyepieces are shown in ﬁgures 4.4, 4.5 and 4.6. Each has two components: the ﬁeld lens and the eyelens. As its name implies the ﬁeld lens enlarges the ﬁeld of view accessible by the eyelens. To do this the ﬁeld lens is placed near the focal plane of the objective so that it does not change the image location appreciably but does pull in ray bundles from the edge of the ﬁeld of view. The ﬁeld lens is not placed too close to the image of the objective; if it were, then any imperfections of the ﬁeld lens would be seen superposed on the ﬁnal image. Many optical instruments use a ﬁeld lenses to extend the ﬁeld of view. Huygens’ eyepiece is the oldest, cheapest and least satisfactory. Two planoconvex lenses are spaced apart by half the sum of their focal lengths so that spherical aberration, coma and lateral chromatic aberration are small, but there is considerable longitudinal chromatic aberration and pincushion distortion. One great drawback is the small eye relief of only 4 mm. The Ramsden eyepiece (not shown) has facing planoconvex lenses and an eye relief of over 10 mm. The Kellner eyepiece is a Ramsden eyepiece with an achromatic doublet for the eyelens. This gives a well corrected wide ﬁeld of view. Another signiﬁcant advantage which it has over the Huygens’ eyepiece is that the image formed by the objective lies in front of the eyepiee. A calibration scale can be mounted there and because
Exit pupil Field lens
Eye lens
Field stop
Eye relief
Fig. 4.5 Kellner eyepiece.
Exit pupil Field and eye lenses
Field stop
Eye relief
Fig. 4.6 Ploessl eyepiece.
80 Optical instruments
Eyepiecee
Objective F 1’
h F1
F2
F 2’
h’ Exit pupil
fo
fo
L
fe
Fig. 4.7 Schematic structure of a microscope. The objective forms an image in the focal plane of the eyepiece, so that the ﬁnal image is at inﬁnity.
this is viewed through the complete eyepiece it will be colour corrected. The ﬁeld stop is no longer the edge of the single lens eyepiece, as it was for the schematic telescope of Section 4.2, but is an aperture placed at the location of the image formed by the objective. Finally the Ploessl eyepiece uses a symmetric layout of achromatic doublets which gives superior imaging. Its ﬂat ﬁeld is particularly important for use with media such as ﬁlm and CCDs which cannot adapt, like the eye, to a curved ﬁeld. Typical eyepieces have focal length 10 mm or 25 mm. Kellners give a 40◦ , and Ploessls give a 50◦ ﬁeld of view. The eyepieces described here are used in both microscopes and in telescopes. Eyepieces are matched to the objective. Take for example a 2.5 m focal length objective with a diameter of 25 cm used visually. Suppose a Ploessl eyepiece is used to give a total magniﬁcation of ×100, then it would require a focal length of 25 mm. If this Ploessl eyepiece has a ﬁeld of view of 50◦ , then the actual ﬁeld of view given by eqn. 4.10 would be 0.5◦ .
4.4
The microscope
Microscopes are designed to give high magniﬁcation and suﬃcient resolution to distinguish features as small as a few wavelengths of light. A microscope is depicted schematically in ﬁgure 4.7. The objective gives an image at the focal plane of the eyepiece and the ﬁnal image at inﬁnity is viewed with the eye relaxed. Suppose the objective has focal length
4.4
The microscope
81
fo , the eyepiece has focal length fe , and that the intermediate image lies a distance L beyond the focus of the objective. Then the magniﬁcation produced by the objective shown in the diagram is Mo = h /h = −L/fo.
(4.15)
The eyepiece magniﬁcation is given by eqn. 3.22, m = dnear /fe , so that the overall magniﬁcation of the microscope is M = −(L dnear )/(fe fo ).
(4.16)
More details of the microscope are shown in ﬁgure 4.9. The size of a microscope is dictated by the average person’s reach, and one common choice is to make the distance between the object and intermediate image equal to 195 mm.1 The objective must accept a wide cone of rays from the object in order to give a well resolved and bright image, and it must also be aberration free. A typical high power oil immersion objective might have focal length 2.0 mm, giving Mo = 100 and NA 1.4 (and would be labelled 100× NA 1.4 2 mm). The ﬁrst stages of such an objective are shown in ﬁgure 4.8 and illustrate the use of conjugate object and image points to give perfect images (see Section 3.7.2). Note that the lens is eﬀectively extended to the object (slide) by ﬁlling the gap between them with oil of the same refractive index as the lens. The object P and its image P1/2 are conjugate points for the lens surface labelled 1. P1/2 also lies at the centre of curvature of surface 2 so that the rays enter the second lens undeviated. Finally P1/2 and its image, P3 , are conjugate points for the surface 3. The wide cone of rays captured from P emerge in a much tighter cone from P3 , so that standard lenses can be used to handle the beam thereafter. In this way a NA as large as 1.4 is achievable. In ﬁgure 4.9 two achromats or apochromats complete the objective. The image is corrected for spherical aberration, coma and chromatic aberration; and the ﬁeld is ﬂat. The eyepieces are similar to those mentioned for telescopes, with a Ploessl eyepiece being shown in ﬁgure 4.9. The illumination of the object is required to ﬁll the angular acceptance of the objective in order that this acceptance is not wasted. Figure 4.10 shows a source and condenser lenses below the microscope stage providing what is known as Koehler illumination. This arrangement has a number of simple advantages. Firstly the area illuminated, the numerical aperture and the brightness can be adjusted independently. The left hand diaphragm (ﬁeld stop) is imaged at the object and controls the ﬁeld of view. Then the right hand diaphragm (aperture stop) is used to match the numerical aperture of the condenser to that of the microscope: if it is smaller then the resolution of microscope is degraded, but if larger light that does not enter the objective directly can be scattered in causing a background haze. Finally the intensity can be changed 1 This
is the DIN (Deutsche Industrie Norm) standard.
2
3
1 P3
P1/2
Glass P C1 C3
Fig. 4.8 Oil immersion objective showing conjugate points.
82 Optical instruments
Exit pupil
Eyepiece first P.P. fe Field stop h’
and image
195mm
L
Objective focal plane fO
Objective second P.P.
Object h
Fig. 4.9 Microscope construction showing the ray cone collected from an oﬀaxis object point.
4.5
by altering the source voltage or with neutral density ﬁlters that absorb equally across the spectrum. A further advantage is that light from each point on the source forms a parallel beam at the object. This ensures that the illumination is coherent2 and also that it is is uniform whatever the variations of brightness across the source.
Cameras 83
2
This topic will be treated in detail in Chapter 7.
The high intensity of illumination needed in a microscope leads to considerable scattering of light from the region around the point being viewed into the microscope. This signiﬁcant background can be eliminated by the widely used technique of confocal illumination. The ﬁeld stops in the illumination system and in the microscope are replaced by small apertures that limit the region on the object illuminated and viewed. This region is typically a circle of diameter 1 µm. Background scattered light is thus eliminated. The image intensity is sampled by an electronic detector behind the eyepiece and an image of the whole object is obtained by moving the microscope stage carrying the slide so that the illluminated spot is raster scanned across the whole slide.
4.5
Cameras
Cameras are used not only for standard photography, but also to record images produced by optical instruments such as microscopes. For all these uses ﬁlm has been mainly supplanted by electronic detectors, notably the charge coupled device (CCD). A CCD consists of silicon photodiodes arranged in a rectangular array. When light is absorbed on a photodiode a charge proportional to the product of the light intensity and its duration is produced and stored. This process will be described in detail in Chapter 15. The exposure time is controlled by an electromechanical or electronic shutter. After exposure the charge on each diode is ampliﬁed, digitized and stored in memory; and from this data an image of the scene can be reconstructed. A typical CCD format for a compact digital camera (‘point and shoot’) is 6.6 mm×8.8 mm, with 3.6 million pixels each of area 4 µm×4 µm. Light falling on each pixel can be focused onto its photodiode by means of a microlens array. Larger CCDs are used in the digital version (DSLR) of the single lens reﬂex cameras described below. Such CCDs will usually be 16 mm×24 mm (APSC format) in area, or 24 mm×36 mm (full frame) matching the ﬁlm size of SLR cameras. In the latter case the CCDs would typically have 12 million 8 µm×8 µm pixels. Film in traditional SLR cameras contains photosensitive crystals of silver halide which are in the range 0.1–2 µm across. The grain distribution is not uniform and an equivalent pixel size, that contains roughly equal numbers of grains, is about 3 µm. The preexisting SLR image format was 24 mm×36 mm and the focal length of the standard lenses was about 50 mm, giving a ﬁeld of view 40◦ × 27◦ . Initially, users of compact digital cameras expected a similar ﬁeld of view, and so the
Lamp and condenser
(a)
Abbe condenser
Diaphragms
Object plane
(b) Fig. 4.10 Koehler illumination. (a) shows the imaging of the left hand diaphragm (ﬁeld stop) at the object. The other diaphragm is the aperture stop. (b) shows how the source plus condenser give uniform illumination at the object.
84 Optical instruments
lenses of such cameras have focal lengths of order 12 mm. Diﬀraction imposes an irreducible limit on the angular resolution achievable of ∆θ = 1.22λ/D, where D is the lens diameter. Thus the resolution in lateral distance at the focal plane is ∆t = f ∆θ = 1.22λf /D, (4.17) which can be reexpressed either in terms of the numerical aperture ∆t = 0.61λ/NA,
(4.18)
or using the f /#, which is the ratio of a lens’ focal length to its diameter, D ∆s Object
u + ∆u
v  ∆v v
Fig. 4.11 Depth of ﬁeld. A point object is displaced a distance ∆u from the location at which its image is in focus. The image becomes a circle of diameter ∆s.
In the case of a pinhole camera the depth of ﬁeld is virtually unlimited: with a 0.5 mm hole placed 25 cm from a screen the equivalent f/# is 500.
∆t = 1.22λ(f /#).
(4.19)
Equations 4.17, 4.18 and 4.19 are equally valid for optical systems other than a camera lens. At a typical aperture of f/8, ∆t is 5 µm which is comparable to the resolution inherent in the pixel or ﬁlm granularity. Lens apertures in SLR cameras can be as large as f/1.2 giving a lens diameter of around 40 mm. One factor that bears on the image resolution is the detail which the human eye can resolve, which was found above to be ∼ 3 10−4 rad. When a photograph is viewed from a distance of 30 cm the corresponding spatial resolution will be 90 µm. Of course the eye looks at a ﬁnal image of say 20 cm width rather than the CCD or 125 ﬁlm. In order to reach that ﬁnal image size the magniﬁcation required from a 24 mm×36 m ﬁlm or CCD image is about a factor of 6. The resolution obtained on the ﬁnal image with a DSLR camera having 8 µm pixels is therefore 50 µm, with the resolution possible with ﬁne grain ﬁlm being a few times times better. The depth of ﬁeld is deﬁned to be the distance along the optical axis that the scene remains eﬀectively in focus, that is to say the image of a point object remains smaller than one pixel. Figure 4.11 shows a camera in which the ﬁlm/CCD plane is oﬀset from the image plane. From eqn. 3.8 the relation between the axial image and object displacements is ∆u = ∆v(u/v)2 . If ∆s is the pixel size then the maximum permitted image oﬀset, ∆v, in ﬁgure 4.11 is given by ∆v/v = ∆s/D, where D is the lens diameter. Combining the last two equations yields ∆u = (∆s/D)(u2 /v).
(4.20)
The image plane is quite close to the second focal plane, so that to a good approximation v may be replaced by f . Then the depth of ﬁeld is ∆u = (f /#)(u/f )2 ∆s.
(4.21)
4.5
Taking as an example a DSLR camera (f = 50 mm, ∆s = 8 µm) with u = 3 m, and f /# = f /8 gives ∆u = 8 cm. The eye tolerates a far larger defocusing than a single pixel. Both CCDs and panchromatic ﬁlm record the full visible spectrum. In the case of CCD a ﬁlter is placed over each pixel to restrict the sensitivity to either the red, green or blue parts of the spectrum. One arrangement of ﬁlters is to repeat a basic fourpixel pattern across the R G CCD: . The sensitivity of ﬁlm is denoted by an ISO number G B (speed) proportional to the ﬁlm density (darkening) for a given exposure. At very low intensities of illumination of the scene photographed both ﬁlm and CCDs record no image, and equally at very high intensities they saturate. CCDs are linear over a larger range of intensity than ﬁlm. Standard ﬁlm has the appellation ISO200, and a faster ﬁlm with twice the sensitivity ﬁlm has ISO400. The equivalent speed of a CCD depends not only on the area of the pixels but also on the level of ampliﬁcation of the charge deposited. This ampliﬁcation can be adjusted electronically and automatically to match light level, aperture and exposure time. The ISO range of DSLR cameras with large 8 µm × 8 µm pixels is typically 100 to 1600. Some CCDs have arrays consisting of alternate small and large pixels, and this provides even more ﬂexibility. Digital cameras oﬀer the huge advantage that the image can be inspected immediately, and then retained or deleted at will. In addition facilities for electronic storage, manipulation and transmission become available once the image is transfered to a PC memory. On many digital cameras the pixel charge is digitized with 8 bits giving a scale of intensity from 1 to 28 , that is from 1 to 256. When the data from the CCD is stored electronically further bits are needed for labelling. Thus for one million pixels of order 8 million bits (8Mb) are required, that is one million bytes (1MB). With the larger 12–20 million pixel CCDs with 12 bit resolution the storage required per picture is correspondingly larger. The full raw data from the CCD may be stored for later processing, but generally the memory requirement is reduced by factors up to 16 by preserving the dta in the JPEG (Joint Photographic Expert Group) format. The content of blocks of pixels are Fourier transformed and the number of coeﬃcients which are kept will depend on the degree to which the data is to be compressed. The rate of taking pictures with a digital camera depends on the rate at which the charges on the CCD are read out, digitized and transfered to memory. With a typical 40 MHz clock speed the readout from a 1 million pixels takes of order 25 ms, allowing ﬁlm mode operation.
4.5.1
Camera lens design
An early lens design that has persisted in cheap cameras is the meniscus lens shown in ﬁgure 3.31. The ﬁeld stop is in the natural location reduc
Cameras 85
86 Optical instruments
Fig. 4.12 Modern camera lens families. Reading from the top these are: Double Gaussian, Airspaced Triplet, and Double Anastigmat
3
Lambda Research Corp, 80 Taylor St., PO Box 1400, Littleton MA 014604400, USA.
ing coma and spherical aberration and ﬂattening the ﬁeld. Distortion and astigmatism are not corrected. With a symmetric pair of lenses facing a central stop, coma, distortion and lateral chromatic aberration cancel partly or completely. Thus a step forward was to have a nearsymmetric pair of achromatic doublets, retaining the meniscus shape. A nearsymmetric pair of achromatic doublets of meniscus shape facing a central stop was the basis of the rapid rectilinear camera lens. Correcting astigmatism/ﬁeld curvature as well as chromatic aberrations required the invention of high refractive index low dispersion crown glass and low refractive index high dispersion ﬂint glass. The resulting cemented triplets of a ﬂint biconcave lens sandwiched between two crown biconvexes are called anastigmats, that is, ﬂat ﬁeld achromats. A triplet of air spaced lenses, again with a biconcave between two biconvexes, has enough ﬂexibility with six surfaces to permit designers to eﬀectively remove all the aberrations simultaneously. This design is the basis a group of lenses such as the Tessar, a modern variant of which is shown in the central panel in ﬁgure 4.12 One further development of symmetric lenses overcomes higher (ﬁfth) order spherical aberration. This is the double Gaussian triplet shown in the upper panel in ﬁgure 4.12; the individual lenses will often be achromatic doublets of meniscus shape. The ﬁnal modern design shown in the lowest panel of ﬁgure 4.12 uses a symmetric pair of anastigmats. Lens designers with a particular aim can search for an existing design in a database, and if necessary, proceed from there by iteration. Current computer based packages perform analytic calculation of lens system properties beyond the simple paraxial behaviour. These packages also perform ray tracing starting from an object point anywhere in the ﬁeld of view. The rays are distributed in direction so that the points where they cross the aperture stop cover this surface in a uniform ﬁne grid. The plot of intersections of the rays across the selected image plane then gives an accurate measure of the quality of the image. This data is displayed and also stored for analysis and comparison with alternative lens designs. All the parameters of the lenses (glass properties, surface curvatures, thickness, etc.) can be varied before a ray tracing sequence. The eﬀects of including aspheric surfaces, mirrors and diﬀractive elements can also be modelled exactly. Many modern lens designs include one or more aspheric surfaces, and their inclusion gives a ﬂexibility that usually results in a design with fewer lenses for the same image quality. R As an example of the detailed analysis feasible with the OSLO design 3 suite, ﬁgure 4.13 shows the ﬁnal design and residual aberrations for a nearsymmetric camera lens.
4.5.2
SLR camera features
Single lens reﬂex cameras are the commonest ﬁlm cameras. A sketch of the components is shown in ﬁgure 4.14. The lens has a focal length of typically 50 mm and an aperture of f/# 1.2, that shown being a double
4.5
Fig. 4.13 Example of a double Gaussian lens design and residual aberrations; supR program manager, Lambda Research Corp. plied by Dr L. Gardner, OSLO
to eye
Diaphragm M
Film
to meter/autofocus Fig. 4.14 The main optical components of a single lens reﬂex camera (SLR)
Cameras 87
88 Optical instruments
4
The top right hand surface of the pentaprism is roofshaped, the ridge running parallel to the line drawn. Light travelling upward in the pentaprism ﬁrst strikes the far roof surface, is then reﬂected towards the reader to the near roof surface. From there it follows the arrowed path element, and is reﬂected horizontally to exit through the rear wall. These four reﬂections restore the image orientation.
Lenslet array Lens
4.5.3
t
Film plane
Baffle
Photodiode array
Fig. 4.15 Autofocusing system. Two ray pencils are shown which form images on the photodiode array.
f
Image plane
O
OL OR
F Aperture stop
Gaussian design. The viewﬁnding, focusing and the light metering are all ‘through the lens’ (TTL) by means of mirrors, hinged at M , which snap out of the way just before the photograph is taken. The shutter that delimits the length of the exposure can be located close to the iris diaphragm at the ﬁeld stop, and has a similar iris structure. Alternatively the shutter can be a focal plane shutter just in front of the ﬁlm, in which case it is in the form of a blind with a slit that is dragged rapidly past the ﬁlm. The front mirror has a central portion that is partly silvered so that some light travels to a second mirror and is reﬂected there to a light meter and autofocus system. Light reﬂected by the ﬁrst mirror goes through an eyepiece and a pentaprism that restores the orientation4 of the scene. In a DSLR camera a CCD replaces the ﬁlm and the image is reproduced on a visual display for the user. Figure 4.15 shows the components of a simple autofocus system. The light passes through two well separated oﬀ axis apertures in front of a lens which forms an image on the equivalent of the ﬁlm plane. Beyond this the rays fall on a transparent screen of lenslets (small lenses). These refocus the two cones of rays onto an array of photodiodes. If the image of the scene is in focus at the ﬁlm, the photodiodes illuminated are exactly a distance t apart. When the image is not in focus on the ﬁlm the separation is greater or smaller than t. The distance t is sensed electronically and the appropriate lens movement performed to regain focus.
I
Fig. 4.16 A telecentric lens with the aperture stop located in the second focal plane.
Telecentric lenses
Telecentric lenses provide a projected rather than a perspective view of small objects and this makes them of great value in machine vision and for the monitoring of production lines. The characteristic feature of the telecentric lens design is to locate the aperture stop at the second focal plane, as shown in ﬁgure 4.16 for a single lens. I is the point image formed of the point O on some object. The rays that reach I leave O in a cone of rays that has the principal ray as its axis, and because the aperture stop is located in the focal plane the principal ray is parallel to the optical axis in the object space. Then, if the object is displaced in either direction parallel to the optical axis the path of the principal ray is unaltered. Although this movement of the object causes the image at I to go out of focus gradually, the image remains centred on I. This property is necessary in checking alignment of pins on PC boards, or of laser drilled holes. The lens aperture (f/#) must be kept small enough that the depth of ﬁeld over which objects remain well deﬁned is adequate for the task. Evidently the front component of a telecentric lens has to be as wide as the area viewed at one time. A typical telecentric lens has a 5 cm diameter objective, a focal length of 5 cm and a working distance of 10–20 cm. The fully corrected telescope is some 30 cm in length; the image is formed on a CCD and the data is processed electronically. Typically the image scale varies by less than 1% for axial displacements of ±10 mm in such instruments.
4.5
4.5.4
Cameras 89
Telephoto lenses
The image size of a distant object is simply its angular size multiplied by the lens focal length. In telephoto lenses a long focal length is achieved in a compact format by moving the second principal plane well out in front of the lenses. Figure 4.17 shows the basic design consisting of a positive lens placed in front of a negative lens. The positive lens focuses a distant point on the optical axis at F+ , and this image is refocused by the negative lens at I. From eqn. 3.39 we take the expression for the equivalent focal length of a pair of lenses separated by a distance t:
P2
f
1/f = 1/f1 + 1/f2 − t/f1 f2 . A simple choice is to make f2 = −f1 , so that: f = f12 /t. In ﬁgure 4.17 the equivalent thin lens, with its second principal plane at P2 P2 , would focus the incident ray shown there along the broken line. The distance between the focal plane and the last lens surface is called the back focal length. Taking an inﬁnitely distant object and using eqn. 3.17 for each lens gives this distance: s2f = f12 /t − f1 . Telephoto lenses are relatively long so they are prone to blurred images arising from camerashake. In order to eliminate this weakness makers build in sensors to detect yaw (rotation about a vertical axis) and pitch (rotation about a horizontal axis perpendicular to the optical axis). Motors displace the lens elements, or the CCD, laterally to compensate for these motions with a response time of milliseconds. It also follows from eqn. 4.21 that the long focal length of the telephoto lens leads to a shallow depth of ﬁeld.
4.5.5
Zoom lenses
The zoom lens is used to achieve very dramatic eﬀects. In a zoom the overall focal length changes and hence the magniﬁcation, while the object viewed remains continuously in sharp focus. Figure 4.18 shows the basic components of a zoom lens, some of which will in practice be achromats, multiple achromats or have aspheric surfaces. The front lens is essentially the objective that focuses the object. Behind this is a negative lens that moves axially during the zoom. Finally there are a pair of relay lenses that bring the image to a focus on ﬁlm or an electronic detector. Placing the ﬁeld stop between the relay lenses maintains a constant f/# during the zoom. The positions of the negative lens are shown at the start (1) and end of the zoom (2), together with representative rays from the object showing that the image plane does not move. The second principal plane of the equivalent thin lens is shown before,
F+
P2
I
s 2f
Fig. 4.17 A simple telephoto lens.
90 Optical instruments
P2
f2 P1
2
f1
1 P2
Stop
P1
Fig. 4.18 Zoom lens showing the position of the mobile lens, the principal plane and ray paths: before the zoom below the optical axis (1), and after the zoom above the optical axis (2).
P1 P1 , and after the zoom, P2 P2 . After the zoom the equivalent focal length is several times larger so the image is correspondingly enlarged. A focal length zoom from 200 to 1200 mm is commonly available, giving a 6:1 change in the magniﬁcation. Zoom lenses represent the state of the art in lens design. The number of lens elements required to remove all aberrations adequately over the full zoom is rather large, 20 elements in some cases. At each air/glass or glass/air interface the fraction of incident light reﬂected is, as we shall show later, for light incident perpendicular to the surface R = (n − 1)2 /(n + 1)2 ,
(4.22)
where the glass has refractive index n. The resulting reﬂection coeﬃcient is 0.04 (0.09) for each glass/air interface where the glass has refractive index 1.5 (1.9). Thus with 40 surfaces the fraction of the incident light transmitted would be less than 0.9640 ≈ 0.20 with usual lens materials. An associated problem arises from the internal reﬂections between lens surfaces of light from bright objects in the scene photographed. If, as is not uncommon, the brightness across the view photographed varies by a factor one hundred, then multiple reﬂections of the bright objects produce blobs and a background haze that dominate the picture. Therefore, where there are many surfaces in a lens system, each surface must be given antireﬂection coatings. Such coatings, which are discussed in Chapter 9, can reduce the reﬂection coeﬃcient to a fraction of a percent per surface. Digital cameras have smaller focal lengths than ﬁlm cameras so the diﬃculties of manufacturing zoom lenses are much reduced due to the lenses being physically smaller, and widespread use is now made of aspheric lenses. Nowadays excellent zoom lenses are ﬁtted as standard on many digital cameras.
4.6
4.6
Graded index lenses
91
Graded index lenses
Graded index or GRIN lenses focus by refraction inside the lens, rather than the surface refraction of a conventional lenses. The glass composition and hence the refractive index changes within the lens. In a similar way, on a hot day, the diﬀerentially heated air above tarmac produces a mirage. There are two approaches: one is to use a radial refractive index gradient, the other an axial refractive index gradient. The ﬁrst approach is employed in lenses which couple laser diodes to optical ﬁbres or ﬁbres to ﬁbres and these lenses are usually a couple of millimetres in diameter. Because the focusing is internal the faces of graded index lenses may be simply ﬂat, parallel surfaces. Figure 4.19 shows the path of rays through a GRIN lens having a radial refractive index gradient given by (4.23) n(r) = n(0) [1 − (g r)2 ], where g is called the gradient factor. Meridional rays, that is to say rays lying in a plane containing the optical axis, follow sinusoidal paths along the lens with a wavelength 2π/g. From Snell’s law we have n(r) cos α(r) = n(r − dr) cos α(r − dr) = .... = n(0) cos α(0),
(4.24)
where α(r) is the angle the ray makes with the optical axis at radius r. At the furthest point from the axis, r = R, the ray is parallel to the axis so n(R) = n(0) cos α(0). Then using eqn. 4.23, R = sin α(0)/g. The optical path length in one complete cycle is straightforward but tedious to evaluate R l=4 n(r) dr/ sin α(r) 0
= n(0) π [1 + cos2 α(0)] /g,
(4.25)
which for paraxial rays reduces to l = 2 n(0) π /g. The optical length axially is l = n(0) p, where p is the wavelength of the sinusoids shown in ﬁgure 4.19. Hence p = 2 π/g. Representative parameters for a GRIN lenses are: 1.8 mm diameter, n0 around 1.6, g around 0.33/mm. A quantity called the pitch is deﬁned to be the length of the lens in wavelengths. The lens shown in ﬁgure 4.19 has a pitch of 1.0 and produces at one end an image of an object placed at the other end. Similarly a lens with pitch 0.25 will focus a parallel beam. Medical endoscopes have integral pitch and can be 30 cm or more long. Some radial graded index lenses provide magniﬁcation of order ×10.
Fig. 4.19 A GRIN lens with radial refractive index gradient, of pitch 1.0. Several meridional ray paths are shown.
92 Optical instruments
In the second class of graded index lenses the index varies linearly along the axial direction. The variation in refractive index is generally much greater than in the GRIN lenses, being as much as 0.15. Shaping the lens surfaces gives new eﬀects. Suppose that the lens is planoconvex and that the refractive index is increasing along the optical axis, then a ray near the axis will travel through glass of on average higher refractive index than rays through the edges of the lens. This feature provides a useful way to compensate for spherical aberration.
4.7
Fig. 4.20 Mediaeval rock crystal lens (c. AD1000) from a Viking hoard found in Sweden. Courtesy Dr. B. Lingelbach, Aalen Technical University.
An alternative way to correct aberrations is to give the lens an aspheric (departing from spherical) proﬁle. Such lenses have a long history and an example from the Gotland Museum in Sweden is shown in ﬁgure 4.20. Aspheric lenses are now widely made directly using computer controlled surface grinding techniques, or by casting from moulds prepared that way. Aspheric lenses can capture a very wide cone of rays from a source while their surface shaping removes any spherical aberration. Planoconvex lenses of focal length 1 to 5 cm with NA as high as 0.75 are very useful as the ﬁrst stages of condensers. Fully corrected lens systems can be made with fewer elements when aspheric lenses are included in the design.
4.8
Standard lens
Section
Fresnel lens
Section
Face on
Fig. 4.21 A normal lens and the equivalent Fresnel lens. On the left are sections which contain the optical axis in each case; on the right, one half of the Fresnel lens is seen as viewed along the optical axis.
Aspheric lenses
Fresnel lenses
Compared with the lenses discussed above the Fresnel lens has relatively poor optical quality, but it is cheap and can be made with a very large area. Sections through a normal lens and its equivalent Fresnel lens are shown in ﬁgure 4.21. At the right of the ﬁgure one half of the Fresnel lens is seen as viewed from along the optical axis. Each annular section of the Fresnel lens has the same surface curvature as the corresponding piece of the normal lens. However the core of the normal lens has been removed from the Fresnel lens, making the latter ﬂat in proﬁle. Each ring of the Fresnel lens focuses in the same way as the corresponding ring in the normal lens which means that the Fresnel lens produces a similar image. However the image quality is poorer because of the light loss and the scattering at the steps between rings. Large area Fresnel lenses can be conveniently cast in plastic. They are often intended to be used at a speciﬁc object distance, and are often aspheric. Each overhead projector (OHP) uses a largish area Fresnel lens to focus the light from a substage condenser through the transparency being projected. The Fresnel lens gives uniform illumination over a wide area and the transparency can sit directly on the Fresnel lens. Fresnel lenses are used anywhere that a large area beam of moderate quality is needed, while wallet sized plastic Fresnel lenses make practical magniﬁers for reading ﬁne print, and are much lighter than glass lenses.
4.8
Fresnel lenses 93
Exercises (4.1) Show that the Galilean telescope of ﬁgure 4.2 produces angular magniﬁcation Mθ = fo /fe . What is the factor by which this telescope expands a laser beam? In a Galilean telescope the objective has focal length 16 cm and diameter 44 mm, and the eyelens −2 cm focal length and diameter 10 mm. What is the angular magniﬁcation? What is the position and diameter of the image of the objective which is formed by the eyelens? In the complete optical instrument made up of this Galilean telescope and the observer’s eye what is the aperture stop? Assume the eye pupil has diameter 5 mm and is 10 mm from the eyelens. (4.2) A microscope has an objective of 2 mm diameter and focal length 10 mm, an eyelens of diameter 15 mm and focal length 25 mm, with tube length 160 mm. Calculate the magniﬁcation of each stage of the microscope. Determine the position and size of the exit pupil. What is the diameter of the ﬁeld of view? (4.3) Whereas a normal eye when fully relaxed brings objects at inﬁnity into focus, a nearsighted person’s fully relaxed eye brings objects much closer into focus. Suppose this far point is at 3 m, what focal length lens is needed to correct this nearsightedness? (4.4) A person with long sight has a near point at 1 m rather than the usual 25 cm. What focal length lens will correct this?
top of the other, the lower one facing up and the upper facing down with their rims are in contact. The mirrors are some 20 cm across and the upper one has a central hole 2 cm in diameter. Viewers see a coin resting on the top of the upper mirror at its centre and attempt to pick it up. In fact the coin is resting on the centre of the lower mirror. How is this illusion achieved? (4.6) What arrangement of two identical positive lenses will invert an image while leaving its linear size unchanged? (4.7) Using the same approach as in Section 4.5 to show that if a camera is focused at inﬁnity then the nearest point in focus is at u = Df /∆s, where ∆s is the pixel size of the detector, f the focal length of the lens and D the lens diameter. For a camera with f/# of 1.4, focal length 12 mm and pixel size 4 µm, what is u? (4.8) An eyepiece is constructed from two identical planoconvex lenses with curved surfaces facing and separated by 2f /3 where f is the focal length of either lens. What is the focal length of the eyepiece and where are the principal planes? Which lens is the ﬁeld stop? (4.9) A perfect oil immersion objective of the type shown in ﬁgure 4.8 accepts rays out to 90◦ oﬀ axis. The refractive index of the spherical lens and the oil is 1.4. What is the semiangle of the cone of rays that the next lens must be designed to accept.
(4.5) In an optical illusion two thin identically curved (4.10) Show that the NA of a GRIN lens of radius R is gRn(0) using the notation of Section 4.6. concave mirrors across are placed on a table one on
This page intentionally left blank
Interference eﬀects and interferometers 5.1
Introduction
The development and applications of the classical wave theory of light form the content of this and the succeeding six chapters. That light waves exist became accepted following Young’s observation, early in the nineteenth century, of interference between light emerging from a pair of slits both illuminated by the same monochromatic source. As mentioned in the ﬁrst chapter the wave theory of electromagnetic radiation, including light, was put on a sound theoretical basis around 1864 by Maxwell. It emerged that ray optics works in everyday situations because the wavelength of light is very small compared to everyday objects. Electromagnetic ﬁelds are vectors and add together in the same simple way as force vectors. This fundamental property is called the superposition principle. It will be used in this chapter to explain interference eﬀects in which a wave is divided up, the two parts travel diﬀerent paths, and are then superposed. Young’s double slit experiment and Michelson interferometer illustrate the two possibile procedures: respectively the division of the wavefront by apertures, and division of the wave amplitude by partially reﬂecting mirrors. The analysis of interference patterns using complex amplitudes and phasor diagrams are introduced at this point. A standard simpliﬁcation made in this and the following chapters is to drop unnecessary constants when calculating intensities where it is the variation in intensity which is of interest: for example when the electric ﬁeld is E the intensity is taken to be E 2 rather than ε0 cE 2 . Whether interference eﬀects are seen or not with a given apparatus depends on the coherence of the wavetrains being superposed. Coherence is treated at length in this chapter, and much use is made of wavepackets in discussing coherence. Practical applications of interference in devices used to measure wavelengths, distances, velocities and angular velocities are described in this chapter. Multiple beam interference produced with Fabry–Perot etalons is introduced, and is shown to give enhanced resolution and precision in wavelength measurement.
5
96 Interference eﬀects and interferometers
5.2
The superposition principle
If electromagnetic radiation from several sources is incident on any given point, the total electric ﬁeld there is simply the vector sum of the electric ﬁelds produced at that point by each source acting alone E = E1 + E2 + ....
(5.1)
Equally, adding the individual magnetic ﬁelds vectorially gives the overall magnetic ﬁeld. These statements go by the name of the superposition principle. When the point of interest is located in matter (rather than in free space) there can be diﬀerences betweeen the physical eﬀect produced by the ﬁelds of each source alone and by the total ﬁeld. For example two lasers illuminating a metal surface may melt the metal, whereas the individual beams leave it intact. Again we might choose a laser whose light can cause atoms in a gas to be excited to a level A and a second laser whose light can cause a transition of atoms from level A to a higher level B. Neither laser alone could access atomic state B but together they can do so.
P d
Slits θ
Monochromatic slit source
in ds
θ
Observing screen
Fig. 5.1 Young’s two slit interference experiment. The distances from the slits to the source and from the slits to the screen are very large compared to the slit separation. A typical slit separation might be 1 mm.
5.3
Young’s two slit experiment
Young’s apparatus is shown in ﬁgure 5.1. A narrow monochromatic source illuminates a pair of slits separated by a small distance d, and beyond the slits is a screen. The slits are located symmetrically about the axis so the waves arriving at each of them are in phase. Young observed bright and dark fringes on the screen, and he realized that these fringes were due to interference between light from the two slits. Figure 5.2 shows an analogous situation in which waves are produced on a water tank by a pair of closely spaced plungers which move at
5.3
the same frequency and in phase. Stars locate the plungers on the ﬁgure. In some directions the waves from the sources are in step and give large amplitude waves which show up as the alternating white peaks and black troughs: the waves interfere constructively. In other directions the waves arrive out of step and interfere destructively leaving the surface undisturbed, which then appears grey in the ﬁgure. Following Young we now appreciate that the electromagnetic ﬁelds in light from the two slits in ﬁgure 5.1 interfere in a similar way. The ﬁelds in a light wave oscillate at frequencies of ∼1014 Hz; frequencies which neither the eye nor any detector can follow. Rather they respond to the time average of the intensity, which is proportional to the square of the electric ﬁeld. Thus the detected intensity of light whose electric ﬁelds are like those displayed in ﬁgure 5.2 would appear as shown in ﬁgure 5.3. The regions of constructive interference, shown white, would be brightly illuminated: and the regions of destructive interference, shown black, would be in darkness. A screen placed anywhere in front of the slits would be covered in bright and dark interference fringes, so that these are called nonlocalized fringes.
5.3.1
Constr. Destr. Constr. Destr. Constr. Destr. Constr. Destr. Constr.
Fig. 5.2 Surface water waves produced by sources at the two stars. The bands of constructive and destructive interference are labelled. Constr.
Fresnel’s analysis
Fresnel, shortly after Young’s observations, used the superposition principle to add the Huygens’ waves from apertures in an obstructed light beam. The secondary waves are spherical and have an electric ﬁeld E = V0 cos (ωt − kr)/r,
Young’s two slit experiment 97
(5.2)
where the angular frequency ω and wave vector amplitude k are the same as for the incident wave, and the distance and time are measured from the origin of the secondary wave on the apertures. The factor V0 depends on the incident wave amplitude, and the factor 1/r ensures that the total power radiated remains constant with the distance from the source. Adding Huygens wavelets from a long slit produces a cylindrical wave. At any point beyond the apertures the total electric ﬁeld is the sum of the secondary wave electric ﬁelds. When the path lengths of secondary waves to the point of observation are diﬀerent there is a phase diﬀerence and this is the origin of interference eﬀects such as that seen by Young. Because the wavefronts are cylindrical, path lengths from the slit must be measured in the plane of ﬁgure 5.1. In the present case, as in may others, the path lengths of the interfering waves are almost equal, so that the eﬀect of the factor 1/r in eqn. 5.2 is an overall constant multiplying the total amplitude. Each slit in ﬁgure 5.1 is a source of Huygens’ waves and their waves arriving at the point P will have travelled distances diﬀering by d sin θ. When this distance is an integral number of wavelengths m, the waves
Destr. Constr. Destr. Constr. Destr. Constr. Destr. Constr.
Fig. 5.3 Time averaged intensity pattern computed for em waves with wavelength identical to those in ﬁgure 5.2. White indicates brightness, black darkness.
98 Interference eﬀects and interferometers
d S1 θ
P θ
S2
B
d sin
θ θ
f0
f
Fig. 5.4 A compact arrangement of Young’s two slit interference experiment. The point source is at the focal point of one converging lens. The observing plane is in the focal plane of a second converging lens. Incoming wavefronts are shown with broken lines.
arrive in phase at P and interfere constructively, d sin θ = mλ,
(5.3)
giving a central bright fringe at which m is zero. Destructive interference occurs when d sin θ = (m + 1/2)λ.
(5.4)
At small angles the separation between adjacent bright fringes is ∆θ = λ/d,
(5.5)
hence the ratio d/λ must be kept small enough so that the fringes can be seen by eye. The above simple analysis applies provided that the apertures are all narrow and the distance of the slits from source and screen are large. If the source slit is so wide that the path length from diﬀerent points on the source to a slit varies by more than a fraction of a wavelength the interference pattern will be blurred. If the two slits are themselves broad, then the pattern of interference becomes more complicated; something which is investigated in the following chapter.
5.3
Young’s two slit experiment 99
Figure 5.4 shows a compact setup with the source and screen lying at the focal planes of positive lenses. The light from the source arrives as a plane wave normally incident at the slits, and hence in phase. In addition light leaving the two slits in a direction making an angle θ to the optical axis will arrive at the point P having travelled paths diﬀering by d sin θ. This conclusion is based on the useful fact that all points on the inclined plane wavefront at S1 B are the same optical distance from the image point P in the focal plane. P lies a distance f tan θ oﬀ axis where f is the focal length of the second lens. Then the electric ﬁelds produced at P by radiation from the two slits are
Light from S2 has the earlier phase because it must leave S2 earlier than light leaves from S1 in order that they both reach P at the same time. Another general point to recall here is that a path diﬀerence of one wavelength produces a phase diﬀerence of 2π, so that φ = 2πd sin θ/λ = k d sin θ.
ω
E1 = E0 cos ωt; E2 = E0 cos (ωt − φ).
Imaginary
E1 2E c 0 os(φ /2
(5.6)
or at any time E = 2E0 cos (φ/2) exp i(ωt − φ/2), with real part (5.7)
Hence the intensity (5.8)
Detectors record the average intensity taken over many full cycles of total duration T. Their response is thus1 T I = 4E02 cos2 (φ/2) cos2 (ωt − φ/2) dt/T 0
= 2E02 cos2 (φ/2) = 2E02 cos2 (πd sin θ/λ). 1 In
detail T
2
cos (ωt) dt/T = 0
2
Fig. 5.5 Phasor diagram for Young’s two slit experiment.
4 3.5 I( θ ) / [E 0/2]
2.5 2
1.5 1
These complex amplitudes are shown on a phasor diagram in ﬁgure 5.5; that is an Argand diagram taken at the moment when E1 lies along the real axis. The superposition principle requires we add the ﬁelds vectorially giving a resultant, which at time t = 0 is
I = 4E02 cos2 (φ/2) cos2 (ωt − φ/2).
E
)
2
E1 = E0 exp iωt; E2 = E0 exp [i(ωt − φ)].
E = 2E0 cos (φ/2) cos (ωt − φ/2).
φ
3
In order to manipulate the math pictorially the complex forms of the ﬁeld are used
E = 2E0 cos (φ/2) exp (−iφ/2);
Real
φ/2
(5.9)
T
(cos (2ωt) + 1) dt/(2T ) = 1/2 + sin (2ωT )/(4ωT ). 0
4ωT is the 8π times the number of cycles completed in time T . Thus if T is the response time of a detector (> 1 ns) and ω is an optical frequency (∼1014 ) the second term is negligible, making the integral equal to 1/2.
0.5 0 2 1.5 1 0.5 0 0.5 d sin θ / λ
1
1.5
2
Fig. 5.6 Intensity in Young’s two slit experiment as a function of the direction oﬀ axis θ, expressed as a multiple of the single slit intensity. The dotted line shows the mean intensity. λ is the wavelength and d the slit separation.
100 Interference eﬀects and interferometers
This variation is shown in ﬁgure 5.6 as a function of the angle oﬀ axis. The peak intensity is four times the time averaged intensity (E02 /2) produced by a single slit. Averaging over a complete spatial cycle of the fringe pattern gives 2π Iaverage = 2 E02 cos2 (φ/2) dφ/(2π) = E02 , 0
so the average intensity over the fringe pattern is the same as the sum of the intensities for two independent slits. No light is lost or gained, it is simply redistributed.
Time averaging for monochromatic waves It will often be useful to use complex waves, making the replacement F = f cos (ωt − kz − φ) ⇒ Fc = f exp [i(ωt − kz − φ)], The choice for the complex form of cos (ωt − kz) could be either exp [i(ωt − kz)] or exp [i(kz − ωt)] which have identical real parts. Of these choices the former is met frequently in classical optics and will be used here for classical optics: the latter choice is standard in quantum mechanics and will be used for the quantum section of the book. The choice exp [i(ωt − kz)] has the nice property that the phase increases with time. If authors make diﬀerent choices in classical optics their calculations give identical results, but there can be a change of sign in intermediate nonmeasurable quantities. Notes are given to indicate where these changes occur. When the waveforms have a purely spatial dependence the choice between exp ±ikz is immaterial and is made for convenience.
(5.10)
where F is any actual electric or magnetic ﬁeld. The only measurable quantities in optics are intensities and ﬂuxes, both being products of two ﬁelds F G averaged over the time of response of the detector. In the case of monochromatic waves F G = f cos (ωt − kz − φ)g cos (ωt − kz − ψ) = (f g/2) [ cos (2ωt − 2kz − φ − ψ) + cos α ], where α = ψ − φ. The time average over many optical periods is F G = (f g/2) cos α. Compare this result with the product of the instantaneous complex ﬁelds Fc G∗c = f g exp iα, and it is seen that
F G = Re [Fc G∗c ] /2.
(5.11)
E 2 = Re [ Ec Ec∗ ] /2 = Ec Ec∗ /2,
(5.12)
Hence for intensities
while for energy ﬂow the time average of the Poynting vector is E ∧ H = Re [ Ec ∧ H∗c ] /2.
(5.13)
Visibility For simplicity the amplitudes due to the two slits have been assumed to be equal. If they are unequal the cancellation when the interference is destructive will not be complete. Suppose the amplitudes are E01 and E02 then the total electric ﬁeld is E = E01 cos ωt + E02 cos (ωt − φ),
5.3
Young’s two slit experiment 101
and the instantaneous intensity is 4 2 I = E01 cos2 ωt + E02 cos2 (ωt − φ) + 2E02 E01 cos ωt cos (ωt − φ) 2 cos2 ωt + E02 cos2 (ωt − φ) + E02 E01 [cos (2ωt − φ) + cos φ]. = E01
Intensity
The time average of this intensity is 2 2 I = E01 /2 + E02 /2 + E02 E01 cos φ,
(5.14)
where Imax and Imin are the observed maximum and minimum intensities respectively. In the case considered V =
2 1
with a minimum that does not fall to zero. A quantity called the visibility is deﬁned which expresses the degree of cancellation V = (Imax − Imin )/(Imax + Imin ),
3
0
0
0.5 1 1.5 2 Path difference in wavelengths
Fig. 5.7 Intensity distributions for fringes with visibility 1.0 (solid line) and visibility 0.5 (broken line).
2E01 E02 2 + E2 . E01 02
The visibility deﬁned in this way has the overall intensity normalized away so that its value is limited to the range between 0 to 1. Figure 5.7 shows two sets of fringes with visibilities 1.0 amd 0.5. Fringes with unit visibility are clearcut, while low visibility fringes are harder to detect because of the poor contrast between the maxima and minima of intensity.
Interference by amplitude division
When viewing the surfaces near normal incidence bright fringes of constructive interference are visible if the optical path diﬀerence between
α α
In Young’s two slit experiment wavefront division produces interfering beams. The alternative process is amplitude division in which a partially reﬂecting surface divides the light into reﬂected and transmitted beams which are subsequently both directed onto a surface where the interference pattern is observed. First consider the simple case of monochromatic illumination. Figure 5.8 shows an arrangement in which part of the incident light is reﬂected by a pair of partially reﬂecting surfaces inclined at a small angle to one another. Interference will only be seen if the two beams can be focused on the same region of the viewer’s retina. This requires that the eye is focused as shown in the diagram, at a speciﬁc depth below the mirrors. Therefore the fringes are said to be localized, in contrast to Young’s fringes which can be detected everywhere in front of the slits. Of course fringes can only be seen if light is incident from the direction that reﬂects into the viewer’s eye. A broad diﬀuse source is needed if fringes are to be seen over a broad angular range. In the familiar case of white light from the sky falling on an oil slick the direction of the bright fringes will be diﬀerent for diﬀerent wavelengths so that coloured bands are seen apparently lying on the oil.
2
5.3.2
Eye
Fig. 5.8 Construction to show the localization of the interference between reﬂections from a pair of plane reﬂecting surfaces inclined at a small angle α.
102 Interference eﬀects and interferometers
Low power microscope
Beam splitter Monochr. point source R Lens
r2 / 2R
r
Optical flat
Fig. 5.9 Setup for observing Newton’s rings. Reﬂections for the same incident ray path from the lens lower surface and optical ﬂat upper surface are drawn as solid lines.
reﬂections is an integral number of vacuum wavelengths (mλ). Between consecutive fringes the path diﬀerence changes by λ and the spacing of the reﬂecting surfaces by λ/2. Such fringes form a contour map of regions of equal thickness. They are called fringes of constant thickness and also Fizeau fringes. Newton’s rings are circular fringes of equal thickness seen when a lens is placed on top of an optical ﬂat: interference occurs between the the reﬂections from the adjacent surfaces of the lens and the optical ﬂat. The experimental setup is illustrated in ﬁgure 5.9 with a planoconvex lens whose curved face has a radius of curvature R. At a radial distance r from the point of contact, the sagitta of the lens is r2 /2R. Light incident there parallel to the optical axis will be reﬂected either at the lens or at the ﬂat and then arrive at the microscope; the two paths diﬀer by an amount r2 /R. An unexpected observation is that the centre of the pattern where the physical path diﬀerence is zero appears dark rather than bright. This comes about because a reﬂection from an optically denser medium (here the air/optical ﬂat interface) at normal incidence produces a phase change diﬀering by π from the phase change for a reﬂection from a less dense medium (here the lens/air interface). Thus bright fringes are seen when r2 /R = (m + 1/2)λ.
(5.15)
The quadratic dependence on r means that the fringes get more tightly packed the further one moves from the optical axis. Departures from fringe circularity indicate an imperfection whose importance can be estimated using the above equation. Viewing Newton’s rings provides a quick practical test of lens quality. When two ﬂat reﬂecting surfaces are aligned parallel as sketched in ﬁgure 5.10 the fringes are now found to be localized at inﬁnity: with the eye relaxed the two reﬂections arrive at the same point on the retina. AC is drawn perpendicular to the reﬂected rays and the diﬀerence in path length between the two reﬂections is A C. The two interfaces are now taken to be identical so that the phase changes are the same at each reﬂection. A bright fringe is seen when the path diﬀerence is a whole number of wavelengths 2nd cos θ = mλ, (5.16) Eye θθ
A
C
d θ
d
B
A’
Fig. 5.10 Fringes localized at inﬁnity from interference between similar parallel reﬂecting surfaces.
θ being the angle between the direction the viewer is looking and the perpendicular to the surfaces and n is the refractive index of the medium. If an extended diﬀuse source is used then each bright fringe seen by the viewer will extend to form a circle subtending a semiangle θ with respect to the surface normal. These fringes are called fringes of equal inclination or Haidinger fringes, and they are localized at inﬁnity. A positive lens can be used to project the fringes onto a screen or photodetectors. Notice that the fringe order m is very large for even a 1 mm spacing of the mirrors and that m is largest at the centre of the pattern. If the source in ﬁgure 5.10 were pointlike its two images would, like Young’s slits, produce nonlocalized fringes. It is worth asking what it is
5.4
about a broad source that suppresses these nonlocalized fringes. First note that diﬀerent regions of the source produce fringe patterns which are displaced from one another. Then note that light from these diﬀerent regions is incoherent, hence what is seen is the sum of the intensities of the fringes from all regions of the source. The overall eﬀect with a broad source is therefore uniform illumination – with one exception. The one exception is illustrated in ﬁgure 5.10 where the eye is focused at an inﬁnite distance. In that case each area of the source gives a bright fringe in the direction satisfying eqn. 5.16 and so the fringes from all areas match exactly.
5.4
Michelson’s interferometer
This interferometer pictured in ﬁgure 5.11 was designed by Michelson to produce constant inclination fringes. With it he and Morley made measurements that underpinned the theory of special relativity. Modern versions of this interferometer are widely used in spectroscopy, especially in the infrared. The source must be both broad and diﬀuse,
Source and diffusing screen
M ’1 M2 stationary M
Mc d
M1 mobile Fig. 5.11 Michelson interferometer. The two paths are indicated by grey and black arrowheads. d is the distance of M1 (M1 ’s image in M) from M2 .
a lamp positioned behind a ground glass screen is normally adequate. Light from the broad diﬀuse source undergoes amplitude division at a beam splitting mirror M which reﬂects 50%, and transmits 50% of the incident light. One beam is reﬂected from the movable mirror M1 , and the other from the ﬁxed mirror M2 . The beams returning after these reﬂections are recombined by M and focused by the positive lens onto a
Michelson’s interferometer 103
104 Interference eﬀects and interferometers
detector. The lens and detector could be an observer’s eye. Mc is a glass plate cut from the same sheet as M; its inclusion equalizes the thickness of glass traversed by the two beams and hence eliminates any chromatic dispersion. The detector receives light directly from M2 and also from the image M1 of M1 formed by M. When M1 and M2 are parallel (M1 and M2 perpendicular) there is a circular pattern of fringes of equal inclination illustrated in ﬁgure 5.12; the semiangle θ of any bright fringe satisfying eqn. 5.16 with d now being the separation of M1 from M2 . One useful feature of the Michelson is that having a virtual mirror M1 the path diﬀerence can be set to zero. A signiﬁcant advantage of the Michelson interferometer is that light from all points on the broad source contributes to the fringe formation, a point explained in the last paragraph in the last section. This makes for fringes far brighter, and easier to use, than those obtained with wavefront division.
Fig. 5.12 Fringes of constant inclination seen with Michelson interferometer when the mirror M2 and virtual mirror M1 are parallel.
When the mirror M1 is moved to increase the path diﬀerence eqn. 5.16 tells us that a fringe corresponding to given value of m will expand (θ increases) and fresh fringes enter at the centre of the pattern. Between the appearances of successive fringes the mirror M1 moves exactly one half wavelength. This simple fact is the basis for measuring mechanical movement directly in terms of wavelengths of atomic transitions, which themselves are determined solely by the laws and constants of atomic physics. Macroscopic lengths can therefore be expressed in fundamental, reproducible units. This is of course intellectually satisfying, but is now essential to many modern industries. The alignment of M2 parallel to M1 is made using monochromatic light. A piece of wire is hooked over the ground glass diﬀuser and M2 is rotated until the two reﬂected images of this wire coincide. At the moment they do so the fringes appear and are generally straight fringes of equal thickness because there is some small remaining tilt between the mirrors. Further delicate rotation of M2 will remove the tilt and give circular fringes. Using eqn. 5.16 the reader may like to show that the angular spacing between adjacent fringes increases as the gap between M1 and M2 is reduced; which suggests a way of bringing M1 into coincidence with M2 . When M1 approaches coincidence with M2 the fringes become very broad and any departures from ﬂatness in the mirrors causes the fringes to lose their circular shape. At zero separation light across the whole spectrum is in step and around this setting a white light source produces a few brightly coloured fringes.
5.4.1
The constancy of c
In the late nineteenth century scientists imagined that light travelled as waves on an otherwise undetectable aether that existed everywhere and through which matter moved without much aﬀecting the aether.
5.5
In modern terms the aether would be an absolute frame with respect to which light would have a constant velocity. If the Earth’s velocity relative to the aether were v the measured velocity of light on Earth would be changed by −v. Michelson and Morley attempted to detect evidence for the Earth’s motion through the aether using the Michelson interferometer. Suppose such an interferometer with arms of equal length L moves with velocity v with respect to the aether in a direction parallel to the length of the M2 mirror arm. As shown in ﬁgure 5.13 the length of the return path starting from mirror M via M2 is (L + D) + (L − D) = 2L, where 2D is the total displacement of M2 through the aether during the √ time the light is travelling to and fro. The return path via M1 is 2 L2 + D2 which is approximately 2L + D 2 /L because v is much smaller than c. The diﬀerence between the two paths is thus D 2 /L. This diﬀerence will reverse if the interferometer is rotated through 90 ◦ so that the motion is now along the direction of the M1 mirror arm. From this argument it would follow that there would be a fringe shift on rotating the interferometer of ∆m = (2D2 /L)/λ.
(5.17)
Now D/L = v/c, and substituting for D/L in the previous equation gives ∆m = (2L/λ)(v/c)2 .
(5.18)
Taking the actual values used, L = 10 m and λ = 500 nm, gives ∆m = 4 107 (v/c)2 . At the time of the measurement it was understood that the Earth’s orbital velocity was 30 km s−1 so that ∆m would have been 0.2, and if one uses instead the solar system’s 600 km s−1 velocity relative to the cosmic microwave background this shift should be twenty times larger. Michelson was able to exclude a fringe shift greater than 10−4 . According to the ﬁrst postulate of the special theory of relativity em radiation travels in free space at the same velocity in all inertial frames, that is to say in frames that have constant relative velocity. This neatly explains the equality of travel times irrespective of how fast the source or observer move. The concept of an aether is therefore seen to be an error.
5.5
Coherence and wavepackets
In the above introduction to interference eﬀects pure sinusoidal, that is monochromatic waves with identical polarization were the rule; after division the wave trains acquired a phase diﬀerence dependent only on their path diﬀerence to the point where they interfere. When waves have
Coherence and wavepackets
105
2D M(0)
M(2t)
M2(t)
L
M1(t) L
Fig. 5.13 Michelson–Morley experiment showing the position of the mirrors at the speciﬁed times in the conjectured aether.
106 Interference eﬀects and interferometers
identical wavelength and a ﬁxed phase diﬀerence they are said to be fully coherent. This ideal situation is only well approximated by laser beams. If such a fully coherent source were available the interference pattern would be present however large the mirror separation in Michelson’s interferometer, simply getting weaker as the light intensity from the far mirror faded away. In practice when a standard laboratory monochromatic source is employed the interference pattern disappears at a mirror separation of only a few centimetres. Such sources give partially coherent beams. A white light source is much less coherent than a laboratory monochromatic source: the fringes disappear if mirror M1 moves only a wavelength or so from the null position. The coherence of a beam depends on the form of the wavetrain of radiation produced by the source and this connection will now be investigated.
2
Resultant intensity
Resultant amplitude
When the quantum nature of electromagnetic radiation is met it will emerge that the wavepackets describe the position and motion of the quanta of radiation. Wavepackets are being presented here with an eye to this basic connection. A crude translation into quantum language would replace the word wavepacket by the word photon. However wavepackets can contain many photons, and in laser beams usually do so. 5 4 3 2 1 0 1 2 3 4 5 0
20 18 16 14 12 10 8 6 4 2 0 0
The wavetrain from any source is made up of very large numbers of wavepackets emitted by individual atoms or molecules. The electric ﬁeld distribution in a wavepacket is like that in ﬁgure 1.6(c).2 A one watt torch bulb emits around 1018 such wavepackets per second. Laboratory reference sources containing a chemically pure gas at low pressure have a spectrum consisting of spectral lines, so that the radiation is conﬁned to a few narrow wavelength intervals. The wavepackets corresponding to these spectral lines are typically of duration 10−10 s and the number of oscillations in a wavepacket is of order 105 rather than the handful shown in ﬁgure 1.6(c). Over a time interval short compared to the wavepacket duration the electric ﬁeld from the ith radiating atom would be approximately Ei = E0 cos (ωt + φi ),
1
2
3 4 5 6 7 8 Distance in wavelengths
9
10
(5.19)
where for simplicity the amplitude E0 , the angular frequency ω and the wavepacket duration are taken to be the same for all the atoms. Each atom radiates independently and so the phase φi is quite random. Summing the electric ﬁelds from all the radiating atoms gives E = E0 cos (ωt + φi ) i = E0
cos ωt
cos φi − sin ωt
i
sin φi
i
= E0 ξ cos (ωt + β). 1
2
3 4 5 6 7 8 Distance in wavelengths
9
10
Fig. 5.14 The electric ﬁeld and intensity produced when there are ten randomly phased wavepackets of the same wavelength and of unit magnitude, and each about ﬁve wavelengths long.
In any other short interval diﬀerent atoms would be emitting radiation and the factors ξ and β would have new values. Thus the electric ﬁeld E = E0 ξ(t) cos [ ωt + β(t) ]
(5.20)
with the magnitude ξ(t) and phase β(t), varying with time. It will be the case that over an interval, taken at some arbitrary time t1 , and short compared to the duration of a wavepacket the wavetrain is approximately sinusoidal. If one then chooses another short section of
5.5
the wavetrain at a much earlier or later time t2 , this wave section too is approximately sinusoidal. However the phase between the two sections is quite random because the time interval is much longer than the wavepacket length: the relative phase will not, except by chance, equal ωt1 − t2 . Figure 5.14 illustrates this, but the reader should bear in mind that typical laboratory sources emit wavepackets that are 10 5 waves long, rather than the few waves shown here. The overall electric ﬁeld and the intensity are seen instantaneously over a test region ten wavelengths long produced by a wavetrain containing ten wavepackets at any given time, each wavepacket of length ﬁve wavelengths, and all having random phases with respect to each other. As time passes this pattern will continually change. The horizontal line on the intensity plot indicates the long term mean intensity. Segments of a wavetrain at instants separated by an interval short compared to the wavepacket duration are fully coherent. However segments separated by intervals long compared to the wavepacket duration have sometimes one phase, sometimes another. Thus the fringes produced by a Michelson interferometer have good contrast when the path diﬀerence between the arms is nearly zero but will fade at large path diﬀerences to leave ﬁnally a uniformly illuminated ﬁeld of view. The total instantaneous sum of the two beams at the detector in the interferometer is, using eqn. 5.20, E = E0 cos [ ωt + β(t) ] + E0 cos [ ω(t + s) + β(t + s) ], where s is the time delay between the arms and for simplicity, and without loss of generality in the result, ξ(t) is taken to be unity. The instantaneous intensity is then I = E02 {cos2 [ ωt + β(t) ] + cos2 [ ωt + ωs + β(t + s) ] +2 cos [ ωt + β(t) ] cos [ ωt + ωs + β(t + s) ]} = E02 {cos2 [ωt + β(t) ] + cos2 [ωt + ωs + β(t + s) ] + cos [ 2ωt + ωs + β(t) + β(t + s) ] + cos [ ωs + β(t + s) − β(t) ]}. The time average of this is T 2 2 I = E0 + E0 cos [ ωs + β(t + s) − β(t) ] dt/T, 0
E02 /2
is the mean intensity of either beam alone. If the beams are where coherent so that β(t) = β(t + s) then this becomes I = E02 + E02 cos (ωs), which shows maximum interference as s is varied, going from 0 to 2E02 . In the case of incoherent beams β(t) − β(t + s) varies randomly and the integral vanishes leaving I = E02 ,
Coherence and wavepackets
107
108 Interference eﬀects and interferometers
M2
M2
M1
M1
Fig. 5.15 Segments of one wavepacket travelling in the Michelson interferometer. In the left hand panel the arms are nearly equal in length, while in the right hand panel they are of very diﬀerent length.
which is simply the sum of the individual intensities. These results bring out the key feature that for incoherent beams the time average intensity, which is what is detected, is just the sum of the individual intensities. However maximal interference is seen in the time average of coherent beams. The above analysis releases us from having to tediously sum the electric ﬁelds for incoherent sources, calculate the intensity and time average – only to ﬁnd that the result is always the sum of intensities. It justiﬁes the simple rubric: with coherent sources add the amplitudes, but with incoherent sources add the intensities. It helps in understanding the nature of coherence to consider individual wavepackets passing through a Michelson interferometer as shown in ﬁgure 5.15. In the left hand diagram the path diﬀerence d is made much smaller than the length of a wavepacket. After the reﬂection the segments of the wavepacket arrive in coincidence at the detector or observer’s eye with a phase diﬀerence 2πd/λ at the centre of the ﬁeld of view. This will be equally true for all wavepackets. Consequently they all contribute to an identical interference pattern and this is what is seen. The right hand diagram shows the opposite extreme in which the path diﬀerence is much greater than the wavepacket length. The two segments of any wavepacket arrive at the detector at diﬀerent times so they do not overlap and so interference between them is impossible. When the wavepacket segment that took the longer path arrives at the detector it can coincide with part of an entirely diﬀerent wavepacket that was emitted later by the source and took the shorter path. Such pairs of parts have quite random phase; some will interfere constructively and just as many will interfere destructively. Averaged over the response time of any detector the overall eﬀect is a uniform illumination. The path diﬀerence over which interference can be observable is called the coherence length Lc for the radiation, and evidently this is simply the
5.5
length of a wavepacket. The corresponding coherence time is τc = Lc /c. A useful indication of the degree of coherence of interfering beams is given by the visibility of the fringes observed. When the diﬀerence between the lengths of the arms of the Michelson interferometer is increased from zero the fringe visibility, deﬁned by eqn. 5.14, declines from unity when the beams are fully coherent to zero when the path diﬀerence exceeds the coherence length and a uniform intensity is seen. Two physically identical sources produce incoherent beams because the atoms in the two are emitting wavepackets quite randomly. This is equally true of the diﬀerent parts of a single source. Therefore periods of time when the interference happens to be constructive between them will be matched by equal intervals of destructive interference. Therefore, averaging over the response time of any detector, a uniform intensity results. Intensities rather than amplitudes are added. Lasers are the exception to the rule that the atoms in a source emit photons with random phases. In Chapter 11 the lasing mechanism will be described in detail. For the moment it is suﬃcient to appreciate that in lasers the the phase of each wavepacket emitted is locked to that of the existing radiation in the laser; this results in the wavepackets remaining in phase for relatively long periods of time. In essence the wavepackets fuse to give one long extended wavepacket. Lasers are therefore capable of producing beams of radiation that are coherent over correspondingly long times; a coherence time as long as a millisecond is readily achieved. This is a big jump in coherence time compared to what other sources could oﬀer in the infrared, visible and near ultraviolet. Many novel research and manufacturing possibilities were opened up by this increase in coherence time and length.
5.5.1
The frequency content of wavepackets
Any wavepacket can be duplicated with a sum of monochromatic waves at frequencies around the mean frequency of the radiation. The technique of Fourier analysis which is used in resolving a wavepacket into these frequency components and in extracting their magnitudes is a key topic in Chapter 7. Here simple arguments will be made which relate the spread of frequencies, ∆f , to the duration of a wavepacket, ∆t. Suppose that the wavepacket is the sum of waves with frequencies f = ω/2π and wave vectors k = ω/c W (x, t) =
Af cos (2πf t − kx + φf ) df .
(5.21)
Without any loss of generality we can take the peak of the wavepacket to be the origin (x = 0). At this point, the peak of the wavepacket, the
Coherence and wavepackets
109
110 Interference eﬀects and interferometers
contributing waves are all in phase so we can put φf = 0 also. Then (5.22) W (0, t) = Af cos (2πf t) df. As time passes the contributing waves at the origin gradually slip out of phase with each other and the wavepacket electric ﬁeld at the origin falls, i.e. the wavepacket moves elsewhere. If fmax and fmin are the maximum and minimum frequencies in the wavepacket then the waves at the origin cancel one another after a time ∆t/2 such that 2πfmax (∆t/2) = 2πfmin(∆t/2) + π i.e. ∆t (fmax − fmin) = 1, or ∆t ∆f = 1.
(5.23)
This relation relates the frequency spread of the constituent waves in a wavepacket to the wavepacket duration, and will be reﬁned in Section 7.3.1. Now the time during which the waves from a source remain coherent is called the coherence time, τc , and is equal to the wavepacket duration ∆t. Hence a relation between the coherence time of light from a source and the frequency spread of the light can be written down τc = 1/∆f.
(5.24)
Correspondingly the coherence length of the waves in free space from a source will be: Lc = cτc = c/∆f = λ2 /∆λ. (5.25) The coherence length so deﬁned is more precisely the longitudinal coherence length, while the transverse coherence length speciﬁes the lateral distance over which radiation from a source remains coherent. A typical source used to produce the line spectrum of an element is a low pressure gas discharge tube, and for such a source the intense lines in the atomic spectrum have wavepackets lasting of order 10−10 s. Sources once used as wavelength references have longer coherence times: nitrogencooled, low pressure krypton sources produce an orangered line of wavelength whose wavepackets are 0.75m in length.
5.5.2
Optical beats
Beating between optical beams is analogous to the beating of tuning forks of very similar frequencies. From what has been said above about the incoherence of one optical source with another it is evident that the observation of optical beats between sources must require very special experimental conditions. The analysis of how and when optical beats can occur helps to bring out the signiﬁcance of the coherence time of the radiation and its interplay with the response time of the detector used. Suppose that in Young’s two slit experiment the slits are illuminated by diﬀerent beams whose polarizations are identical but whose angular frequencies are ω1 and ω2 . Suppose also that the coherence times of
5.5
Coherence and wavepackets
111
both sources are longer than the response time of the detectors used. The electric ﬁeld at point P in ﬁgure 5.1 would be the sum of two pure sinusoidal waves during an interval short compared to the coherence time E0 cos (ω1 t − k1 s1 ) + E0 cos (ω2 t − k2 s2 ), (5.26) where ki = c/ωi and si is the path length from slit i. The sources could be beams from separate lasers. The intensity is then I = E02 [ cos2 (ω1 t − k1 s1 ) + cos2 (ω2 t − k2 s2 ) +2 cos (ω1 t − k1 s1 ) cos (ω2 t − k2 s2 ) ] = E02 [cos2 (ω1 t − k1 s1 ) + cos2 (ω2 t − k2 s2 ) + cos (Σωt − Σ(ks)) + cos (∆ωt − ∆(ks))],
(5.27)
where ∆ω is the diﬀerence (ω1 − ω2 ), Σω the sum (ω1 + ω2 ) and so on. Detector response is produced by the total ﬂux of eﬀective radiation during the response time td of the detector, that is td times the average light intensity.3 In the case of photomultipliers and photodiodes we can take 10 ns as a representative value for td . The ﬁrst three terms in eqn. 5.27 oscillating at 1014 Hz would average out over the many full cycles in the detector response time to E02 /2, E02 /2 and 0 respectively. However the ﬁnal term has an angular frequency ∆ω, which may be low enough that the detector can sample the waveform more than once in each cycle. This condition is fulﬁlled if td ∆ω 2π. Then taking td to be 10 ns the critical frequency diﬀerence below which the detector can sense the variations in the intensity is ∆fd = ∆ωd /2π = 1/td = 100 MHz. A screen covered with detector pixels would record the transient fringe movements provided the frequencies of the two stable sources were closer that ∆fd . Then eqn. 5.27 reduces to I = E02 [1 + cos (∆ωt − ∆(ks))].
(5.28)
The oscillations observed are optical beats analogous to audible beats. The fringes would be transitory, lasting for a coherence time and then changing position. If however the source frequencies diﬀer by much more than ∆fd the response of the detector to the ﬁnal term in eqn. 5.27 is also zero leaving an intensity I = E02 everywhere across the screen. That is uniform illumination with an intensity equal to that produced by the two slits separately. In summary, optical beats can only be detected when the sources are highly coherent and close enough in frequency that the beat period is much longer than the response time of the detectors involved. The beating together of radio waves shows diﬀerent features because detectors at radio frequencies respond linearly to the electric ﬁeld unlike detectors of visible radiation which respond to the intensity.
3
By eﬀective is meant radiation at wavelengths to which the detector responds. The response will usually be the release of electric charge and the response time is the length of time taken for this charge to travel through the detector plus the time for this charge signal to be processed by the associated electronics.
112 Interference eﬀects and interferometers
Quasimonochromatic sources It follows from the preceding that if the spread in angular frequency of radiation from a source ∆ω, and the response time of the detectors τd , are such that ∆ωτd 2π, then interference fringes produced by all the frequency components will superpose giving high visibility fringes. In such a cases the coherence time then greatly exceeds the response time and the source is said to be quasimonochromatic. The equations given above for evaluating time averages of intensities and energy ﬂux in radiation, eqns. 5.12 and 5.13, apply to such quasimonochromatic radiation.
5.5.3 Slit
Source aperture
θs θd
∆y
∆y
d
n θd
si
Coherence area
The ﬁnite width of a source imposes limitations on the coherence of a wavetrain transverse to the direction of travel. In ﬁgure 5.16 the source illuminating Young’s slits is ∆y wide and at a distance z from the slits, the angle subtended by the source at the slits is θs while the semiangle subtended at the source by the slits is θd . The path diﬀerence between light arriving at the upper slit from the two edges of the source is
Slit
∆y sin θd = d∆y/2z,
z
Fig. 5.16 Transverse coherence length due to ﬁnite source dimension.
and there is a similar spread in the case of the lower slit. This leads to a spread in the relative phase between the light arriving at the two slits of kd∆y/2z = πd∆y/(zλ). Thus the light arriving at the slits is coherent only if this phase diﬀerence is small. A corresponding lateral coherence length can be deﬁned, within which the slits must lie from one another in order that there can be interference dc = zλ/∆y = λ/θs . (5.29) For a circular source the corresponding lateral coherence area is then Ac = λ2 /(πθs2 ).
(5.30)
Tc = Ac (πθs2 ) = λ2 ,
(5.31)
Arranging this diﬀerently
and referring to eqn. 4.11 shows that the etendue from the source into the coherence area is exactly λ2 . The coherence area and the coherence length taken together deﬁne a coherence volume. At all points within this volume the waves have a constant phase relation and could be brought together in some apparatus to interfere. This could be in Young’s two slit experiment or a Michelson interferometer. Each wavepacket in the beam will, at any given moment on its journey, determine a coherence volume.
5.6
5.6
Stokes’ relations
113
Stokes’ relations
There is a general principle, called time reversal invariance, stating that if time were reversed then physical processes could reverse exactly. The only known violations are observed in particular weak decays of elementary particles. However on the macroscopic scale a reversal of any dissipative process, such as a bursting bubble viewed with a video in reverse, will occur with negligible probability. When time reversal invariance is applied to the reﬂection and transmission of light at an interface between media relationships are obtained between the reﬂection and transmission coeﬃcients for the electric ﬁelds. These will only hold if the absorption at the surface, a dissipative process, is negligible. The case of total internal reﬂection is also excluded. Taking the point of reference on the interface to be the origin of the spatial coordinates, the incident, reﬂected and transmitted waves there all have electric ﬁelds of the form a exp (iωt) where ω is the angular frequency. Now a = a exp (iφ) so that any alteration of the time origin will be absorbed by a change in the phase φ. Therefore time reversal changes such an amplitude to a∗ exp (−iωt), because for consistency φ must change to −φ, as well as t to −t. Two time reversals evidently restore the original waves. Figure 5.17 shows both the original and the timereversed processes: r and t are the reﬂection and transmission coeﬃcients for light incident from above, r and t are those for light incident from below. The time reversed reﬂected and transmitted waves reproduce the incident beam, while the beam at the bottom left of (b) must vanish. Then r∗ r + t∗ t = 1, ∗
∗
r t + t r = 0.
(5.32) (5.33)
The coeﬃcients can be complex if the surface is a set of thin layers of dielectric. However for a single surface the coeﬃcients are real and the equations simplify to r2 + tt = 1, r = −r.
(5.34) (5.35)
These equations are known as Stokes’ relations after their 19th century discoverer. All four relations can be derived from the requirement of conservation of energy between the incident and outgoing light beams if there is negligible absorption.
5.7
Interferometry
A vast range of research and industrial devices have been built which make use of interference eﬀects. Applications include: measuring distances with precision of a fraction of a wavelength; stabilizing laser wave
(a)
(b) r
1
r*r+t*t
t
r*t+t*r
r*
t*
Fig. 5.17 Reﬂected and transmitted amplitudes: (a) for forward process and (b) for the time reversed process.
114 Interference eﬀects and interferometers
lengths and measuring wavelengths; testing and measuring the imperfections of optical components; inertial guidance in aircraft; measurement of refractive indices of gases; and revealing ﬂuid structure and ﬂow patterns in Tokomaks and wind tunnels. Predominantly these devices involve amplitude rather than wavefront division. The underlying reason is that wavefront division requires lateral coherence across the area covering the slits and this in turn implies a small area source, and in turn this means weak illumination. By contrast, taking an example of amplitude division, the Michelson interferometer has a broad source providing strong illumination. Amplitude division lends itself to ﬂexible designs, with derivatives of the Michelson interferometer design being frequently met.
5.7.1
The Twyman–Green interferometer
Monochromatic source and Pinhole
Lens under test
Linear Fringes
Convex mirror
Tilted plane mirror
Fig. 5.18 Twyman–Green interferometer with a lens under test. The optical quality convex mirror is positioned to return the rays from the lens under test along their incident direction.
The Twyman–Green interferometer sketched in ﬁgure 5.18 is a variant of Michelson’s inteferometer, which is used to examine optical component quality, in this case the quality of a lens. The combination of a point monochromatic source located at the focus of a positive lens provides a parallel coherent beam. An optical quality convex spherical mirror is positioned behind the lens being tested and is moved relative to the lens until the rays are retroreﬂected back along their incident path. In the other arm a plane mirror is tilted so that straight line fringes are seen by eye or imaged onto a detector array. If the lens were perfect the fringes would be equally spaced and straight. Any imperfections cause distortions which can be interpreted in order to deﬁne a grinding and polishing sequence which would correct the lens. Usually the image is
5.7
Interferometry 115
captured on a CCD array and the analysis carried out by proprietary software. In many industries standards of length are maintained using gauge blocks. Gauge blocks are simple cuboids with one pair of opposite calibration faces both optically ﬂat and parallel. The material used has low coeﬃcient of thermal expansion, such as the metal Invar (64% Fe, 36% Ni) with coeﬃcient 1.3 10−7 K−1 or the ceramic Zerodur with coeﬃcient as low as 10−8 K−1 . The thickness of the gauge block between the calibration surfaces is ﬁrst measured mechanically, giving a thickness between tlo and thi , which diﬀer by ∼ 10 µm. Next the gauge block is wrung in contact with the mirror in one arm of an interferometer as shown in ﬁgure 5.19. Fringes are seen from the mirror surrounding the gauge block and from the gauge block. The lateral displacement between the fringes on mirror and gauge block, expressed as a fraction of a fringe spacing ∆ni , is measured for several wavelengths λi . Then the actual thickess of the block is t = (ni + ∆ni )λi ,
(5.36)
where the values of the integers ni are as yet unknown. Benoit’s method of exact fractions provides a way to obtain these values and hence to determine t with a precision of tens of nanometres. For one wavelength λ1 the values of n1 consistent with values of the thickness lying between tlo and thi are each taken in turn and a thickness calculated. One of these thicknesses will be correct, but the question is which one. Using each such thickness in turn, t(trial), the values of the interference order for each other wavelength are calculated, ni (trial) + ∆ni (trial) = t(trial)/λi .
(5.37)
For the correct choice of t(trial), and only for that choice, all the trial fractions ∆ni (trial) will match the corresponding measured values ∆ni . This match determines t. In practice three or more wavelengths are used to eliminate the possibility of accidental coincidences between the trial and measured values of the ∆ni ’s.
5.7.2
The Fizeau interferometer
This interferometer, shown in ﬁgure 5.20, uses amplitude division within a single arm to produce interfering beams. It too is used for optical testing. An intense coherent beam from a laser is focused by a microscope objective onto a pinhole. This spatial ﬁlter produces a beam with lateral coherence over a wide angle. Such a design allows the interferometer to accomodate very large area lenses and mirrors that are used in the space industry. A collimating lens directs the beam through an optically ﬂat reference surface and then onto the component under test. In the example shown a mirror is being tested for ﬂatness. Light returning from the
Gauge block Optical flat
Fringe pattern
Fig. 5.19 Side view of gauge block on an optical ﬂat, together with the fringe pattern seen with a Twyman–Green interferometer.
116 Interference eﬀects and interferometers
Monochr. point source
Reference surface Mirror
Lens
Surface tested
Fringes
Fig. 5.20 Fizeau interferometer used to test an optical ﬂat.
mirror passes again through the reference surface and collimating lens and ﬁnally is reﬂected by the partially reﬂecting surface onto an image plane, rather than travelling back to the pinhole. The reference surface is of high quality, ﬂat to say λ/50, and the reference plate is a slightly wedge shaped so as to throw the reﬂections from its other surface out of the ﬁeld of view. Interference fringes are formed between the waves reﬂected from the reference surface and the surface under test. If the test surface were perfect and tilted slightly with respect to the reference surface a pattern of linear equally spaced fringes would be seen. The capture of the actual fringe pattern, and its analysis, proceed as described above for the Twyman–Green interferometer.
5.7.3
BS2
Imaging
M1
Test volume
BS1 M2 Extended source
Fig. 5.21 Mach–Zehnder interferometer.
The Mach–Zehnder interferometer
The Mach–Zehnder interferometer is shown in ﬁgure 5.21. Monochromatic light from an extended diﬀuse source is divided by the beam splitter BS1 and the separated beams travel via either mirror M1 or mirror M2 to meet again at the second beam splitter, BS2. Emerging together from BS2 the beams are focused by the lens onto an image plane where their interference fringes are observed. The design permits the two paths to be widely separated: one path can cross a multimetre sized volume such as a wind tunnel or a nuclear fusion device while the other path skirts that volume. Whenever a test chamber is placed in one arm, glass plates identical to those belonging to the test chamber are inserted in the other arm in order to cancel their optical eﬀects on the fringe pattern. When the mirroring surfaces are all set at exactly 45 ◦ to the rectangular shape of the beam path the interference fringes are located at inﬁnity and appear in the focal plane of the lens. By tilting BS2 from the 45◦ orientation the fringes are moved so that they are localized within the test volume and the image plane is moved correspondingly, as shown in the ﬁgure. Photographs of the fringes then show any mechanical structure in the test volume in focus together with the fringes. The structure might be an aircraft wing in a wind tunnel. In a dynamical situation involving gas compression and ﬂow any variation in the refractive index of the gas in the test chamber will alter the optical path length and distort the fringe pattern. In this way the ﬂow lines of gas around an aircraft model become observable and are available for interpretation. Interferometers used to observe the gas ﬂow in fusion reactors employ CO 2 lasers of wavelength 10.6 µm and can detect variations in the plasma electron density of one part in a thousand (1017 m−3 in 1020 m−3 ).
5.7.4
The Sagnac interferometer
The Sagnac interferometer is unique in its ability to allow the detection of rotation. Figure 5.22 shows that the device consists of a closed optical path around which coherent beams from the laser source circulate in opposite directions. This design is called passive because the laser is not integral with the ring. A beam splitter BS divides the laser beam
5.7
to provide two equal amplitude counterrotating beams. After making a circuit round the four arms the beams exit through the same beam splitter and arrive superposed at the detector. An active ring laser is formed from a monolithic glass block with surface mirrors in place of the individual mirrors in ﬁgure 5.22. The glass is doped with lasing material so that it forms both the laser and the closed path. When the ring is rotated the path for one direction of travel is shortened and for the other lengthened; an eﬀect that can be detected optically and used to measure the rate of rotation. A simple question but awkward question is this: rotation with respect to what? The answer supplied by the general theory of relativity is that the rotation measured is that with respect to the local inertial frame, that is to say the frame of the ﬁxed stars. It is not the rotation relative to the Earth or relative to the solar system that is measured. In analysing this eﬀect we treat the ring as a circular evacuated loop of radius R. The rate of rotation with respect to the inertial frame is Ω. Then the angle rotated in the time it takes light to complete one circuit is ∆θ = 2πRΩ/c. Thus the paths for the two senses of rotation diﬀer in length by ∆s = 2R∆θ = 4πR2 Ω/c. The corresponding phase diﬀerence induced between the beams after one rotation is then ∆φ = 2π∆s/λ = 8π 2 ΩR2 /(cλ) = 8πΩA/(cλ),
(5.38)
where A is the area enclosed by the loop. This result generalizes for the case of an arbitrarily shaped loop of vector area A having angular velocity Ω to ∆φ = 8πΩ · A/(cλ). (5.39) The sensitivity to rotation can be improved by using a multiturn loop of optical ﬁbre to carry the light; the phase change is then multiplied up by the number of turns. In an active ring with the laser in the loop, the complete orbits of the counterrotating beams diﬀer by ∆s, and so the times for making one complete orbit diﬀer by ∆τ = ∆s/c = 4πR2 Ω/c2 , between the counterrotating beams. Thus the fraction diﬀerence in the orbital periods is ∆τ /τ = 2RΩ/c and the frequency diﬀerence between the counterrotating beams from the one laser is ∆f = f ∆τ /τ = 2RΩ/λ.
Interferometry 117
Detector Laser BS
Fig. 5.22 Sagnac interferometer; passive ring gyro layout.
118 Interference eﬀects and interferometers
This result generalizes to ∆f = 4Ω · A/(λP ),
(5.40)
where P is the length of the perimeter of the loop. These results, in particular eqns. 5.39 and 5.40, are valid whatever material ﬁlls the path provided that this is also rotating at the angular velocity Ω, and that the wavelength used is the wavelength in free space.4 A practical diﬃculty met with in using active Sagnac rings is that the frequencies of the two beams tend to lock together at a common frequency when the rotation frquency Ω is small. The coupling between the beams comes from back scattering at the mirrors in the optical path. In order to reduce this scattering to a low enough level that the coupling is removed it has been necessary to develop super eﬃcient mirrors with reﬂection coeﬃcients as high as 0.999 999! Compact active Sagnac interferometers arranged in a set of three with orthogonal axes are used to provide inertial guidance. The necessary dimensional stability is achieved by carving the frame containing the three units from a common block of Zerodur ceramic. These ring gyroscopes have no moving parts and oﬀer superior precision to mechanical gyroscopes, so they have been ﬁtted in high performance aircraft. More recently these have been replaced in many applications by the simpler ﬁbre optic gyros described in Chapter 16.
5.8
Standing waves
An important application of the superposition principle is to the reﬂection of plane sinusoidal electromagnetic waves incident on the surface of good conductors. Suppose the incident electromagnetic wave’s electric ﬁeld is Ei (x, t) = E0 cos (ωt + kx) (5.41) parallel to the surface. At the surface, taken to be x = 0, the incident ﬁeld is: Ei (0, t) = E0 cos (ωt). (5.42) Metals are good conductors, therefore the simplifying assumption is made that the free electrons within the conductor move instantaneously so keeping the electric ﬁeld parallel to surface extremely small. This is close to the actual behaviour in the metals copper, silver and gold up to optical frequencies. The displaced electrons generate an electric ﬁeld equal and opposite to the incident electric ﬁeld at the surface: Er (0, t) = −E0 cos (ωt).
(5.43)
Applying Huygens’ principle, this disturbance produces a reﬂected plane 4 See for example Chapter 1 of The Fibre optic Gyroscope by H. Lefevre, published in 1993 by Airtech House, Boston.
Incident and reflected wave amplitudes
5.8
1 0 1 1 0 1 1 0 1 1 0 1 1 0 1 1 0 1 1 0 1 1 0 1 0
0 τ/8 τ/4 3τ/8 τ/2 5τ/8 3τ/4 7τ/8 1 2 3 Wavelengths from mirror
4
Fig. 5.23 Travelling waves from the right reﬂected by a perfect mirror; incident/reﬂected waves shown as full/broken lines: shown at intervals of one eighth of a period.
wave propogating away from the surface: Er (x, t) = −E0 cos (ωt − kx).
(5.44)
The superposition principle determines the total electric ﬁeld to be E(x, t) = Ei (x, t) + Er (x, t) = E0 cos (ωt + kx) − E0 cos (ωt − kx) = −2E0 sin (ωt) sin (kx).
(5.45)
Figure 5.23 shows the incident wave (solid curves) and reﬂected wave (broken curves) at successive intervals of one eighth of a period (τ /8). Figure 5.24 shows the corresponding total wave amplitude. This oscillates, but at certain locations called nodes the amplitude is always zero. This is therefore called a standing wave. Starting at the conductor the nodes are spaced at intervals of one half wavelength of the incident travelling waves. Midway between each pair of nodes lie the antinodes where the ﬁeld variation is largest. Figure 5.25 shows the incident and reﬂected ﬁeld vectors at the surface of a good conductor. The relative orientation of E, B and the wave direction k is preserved after reﬂection and so the magnetic ﬁeld is not cancelled out at the surface. If the metal surface is rough the reﬂections from nearby points on the surface travel in diﬀerent directions and a mirror smooth to a fraction of a wavelength is needed in order to produce simple standing waves. Everyday mirrors consist of
Standing waves 119
Total wave amplitude
120 Interference eﬀects and interferometers
2 0 2 2 0 2 2 0 2 2 0 2 2 0 2 2 0 2 2 0 2 2 0 2 0
0 τ/8 τ/4 3τ/8 τ/2 5τ/8 3τ/4 7τ/8 1
2 3 Wavelengths from mirror
4
Fig. 5.24 The resultant standing waves from the incident and reﬂected waves shown on ﬁgure 5.23. Note that the ordinate range is twice that in ﬁgure 5.23.
Er
Br B in
Reflected wave direction Incident wave direction E in
Fig. 5.25 The orientation of the electromagnetic ﬁelds and the direction of wave propogation of the incident and reﬂected waves at the mirror.
metalcoated smooth glass surfaces, while at longer wavelengths a wire mesh is adequate for satellite dish antennae and radiotelescopes. This saves on cost and makes radiotelescopes easier to support in high winds. Standing electromagnetic waves are readily demonstrated by directing microwaves onto a metal sheet. The standing wave pattern produced by a 10 cm wavelength source will have nodes spaced 5 cm apart, which are readily observed using a diode detector. By 1890 Wiener had photographed standing waves produced with visible light. As shown in ﬁgure 5.26 a glass plate coated with photographic emulsion rested with one edge on a plane frontmetallized mirror and was tilted at a small angle α (10−3 rads) to the mirror. Plane monochromatic waves whose wavefronts were accurately parallel to the mirror surface illuminated the plate. The antinodes of the electric ﬁeld are visible in ﬁgure 5.26, spaced a distance λ/2 apart perpendicular to the mirror’s surface. This spacing is ampliﬁed by the tilt to λ/(2α) ≈ 0.125 mm along the emulsion’s surface. When the photographic plate was processed the expected, equally spaced black stripes were seen where the light had activated the silver halide grains. At the edge of the emulsion in contact with the mirror, which is where the electric ﬁeld is zero and the magnetic ﬁeld is at its largest, the emulsion was clear. Hence the photochemical eﬀect which leaves a developable image in the emulsion is caused by the action of the electric rather than the magnetic ﬁeld.
5.9
In a laser two mirrors face one another and thanks to the energy drawn from the active material standing electromagnetic waves develop between them analogous to those on a violin string when bowed. In each case there is a node at the two end points. The simplest fundamental violin string oscillation has an antinode midway along the string and a wavelength of twice the string length, 2L. Vibrations are also possible with 2,3 ... antinodes along the wire. Each distinct pattern of oscillation is called a mode, with the number of antinodes deﬁning the order, n of a mode. Figure 5.27 shows modes with n equal to 1, 2, 3 and 30, where successive displacements of the string are indicated at time intervals of τ /16. In the case of standing electromagnetic waves between mirrors the order of these modes is very much higher. For the nth mode the required wavelength (λ) is given by nλ/2 = L, i.e. λ = 2L/n.
The Fabry–Perot interferometer
121
E node
E node
α
Mirror and E node
Fig. 5.26 Wiener’s experiment producing standing waves of light at a mirror surface.
(5.46)
An optical arrangement consisting of facing parallel mirrors of high reﬂectivity is called a Fabry–Perot etalon or cavity. The electric ﬁeld in the Fabry–Perot etalon has to vanish at both mirrors and so has the form: E = 2E0 sin (ωt) sin (kx), (5.47) (5.48)
with n integral in order that the transverse electric ﬁeld vanishes at x = L as well as x = 0. In helium–neon gas lasers the active gas is contained in a 50 cm long Fabry–Perot cavity and the lasing wavelength is 633 nm, making n around 1.6 million.
5.9
The Fabry–Perot interferometer
The Fabry–Perot interferometer shown in ﬁgure 5.28 is a single arm instrument with a broad diﬀuse source and a lens to focus the Haidinger fringes formed between the glass plates. These glass plates are closely spaced with their inner faces being highly reﬂecting and optically ﬂat. Together with their supporting frame the plates form a Fabry–Perot etalon. The inner surfaces of the plates can be aligned parallel by screws that hold the plates against a spacer of Zerodur or other material having a very low coeﬃcient of thermal expansion. Interference is observed between the multiple beams produced by successive reﬂections from the highly reﬂecting surfaces of the etalon. Some representative reﬂections from one point on the source are drawn in the diagram with solid lines. Parallel rays from all points on the source such as the one indicated by a broken line will give similar sets of reﬂections and coincident fringes. It will emerge below that by using multiple reﬂections a chromatic resolving power is obtained which far exceeds that of prisms or diﬀraction gratings. A thin layer of aluminium on the etalon faces is commonly used to give a uniform reﬂectivity across the visible spectrum of 80%, while
1 (a) 0.5 n=1 0 0.5 1 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
Distance
1 (b) 0.5 n=2 0 0.5 1 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Wave amplitudes
k = 2π/λ = nπ/L,
Wave amplitudes
where
Distance
1 (c) 0.5 n=3 0 0.5 1 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
Distance
1 (d) 0.5 n=30 0 0.5 1 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
Fig. 5.27 Standing wave patterns for (a) the n = 1 mode, (b) the n = 2 mode, (c) the n = 3 mode, (d) the n = 30 mode. In (a), (b), and (c) the waves are shown at intervals of τ /16, and at intervals of τ /2 in (d).
122 Interference eﬀects and interferometers
Spacer
Diffuse source
Etalon
Lens
Screen
Fig. 5.28 Fabry–Perot interferometer. The inner faces of the etalon have large reﬂection coeﬃcients. A lens brings the parallel rays from multiple reﬂections to a common focus. One set of reﬂections is shown. Rays from other places on the source like the broken line ray give similar sets of reﬂections.
reﬂectivities as high as 99% are readily achievable with more complex coatings. The outer surface of each plate is ﬂat, uncoated and makes a small angle to the inner surface so as to remove its own reﬂections from the ﬁeld of view.
Fig. 5.29 Appearance of Fabry–Perot fringes for the same mirror spacing as in ﬁgure 5.12 with a ﬁnesse of 15.
Suppose that by some means the reﬂection coeﬃcient of the facing surfaces were simultaneously increased from the glass/air value of a few % to 90% while one watched. Initially a set of broad constant inclination circular fringes would be seen, just like those produced by the Michelson interferometer. When the reﬂection coeﬃcient increased the bright fringes would get progressively narrower, as shown in ﬁgure 5.29, but would remain ﬁxed in position. Finally when each surface reﬂected 90% of the incident light the total transmission would fall to 10%×10% = 1%, which would be concentrated in the narrowed bright fringes. Thus light from two closely spaced wavelengths whose fringes would overlap appreciably in a Michelson interferometer produce well separated fringes in a Fabry–Perot interferometer. This behaviour is now analysed quantitatively. Figure 5.30 shows the transmitted waves following multiple reﬂections in a Fabry–Perot etalon for an angle of incidence θ. The path diﬀerence between the successive transmitted rays drawn in the ﬁgure is given by eqn. 5.16 with d being the separation of the mirrored surfaces and n the refractive index of the material between them. Thus the phase delay between successive rays transmitted, in terms of the wavelength
5.9
The Fabry–Perot interferometer
123
in free space and frequency is δ = 4πnd cos θ/λ = 4πndf cos θ/c.
(5.49)
We ignore the weaker diverted reﬂections at the outer unsilvered surfaces of the etalon plates. Let the reﬂection coeﬃcient for the wave amplitude at the air/glass interface be r and the transmission coeﬃcient be t; while for the glass/air interface let these coeﬃcients be r and t respectively. Assuming that there is no absorption at the surfaces we can use Stokes’ relations between these coeﬃcients given in eqns. 5.34 and 5.35. The superposition principle gives a total transmitted amplitude
r4tt1 r4t1 r3t 1 r2t1
E = tt [1 + r2 exp (iδ) + r4 exp (2iδ) + ... = tt /[1 − r2 exp (iδ)] = (1 − r2 )/[1 − r2 exp (iδ)]
rt1
(5.50)
where we have use Stokes relation eqn. 5.34 in the last line. For reference later the reﬂected amplitude is
3
Er = r + tt [r exp (iδ) + r exp (2iδ) + ...] = −r + (1 − r2 )r exp (iδ)/[1 − r2 exp (iδ)], where both eqns. 5.34 and 5.35 have been used. Then Er = r(exp (iδ) − 1)/[1 − r2 exp (iδ)].
(5.51)
Now putting r2 = R and 1 − r2 = T in eqn. 5.50 E = T / [ 1 − R exp (iδ) ]. The transmitted intensity at an angle θ is therefore I(θ) = EE ∗ = T 2 [(1 − R cos δ)2 + R2 sin2 δ]−1 = T 2 [1 − 2R cos δ + R2 ]−1 = [T /(1 − R)]2 [1 + 4R sin2 (δ/2)/(1 − R)2 ]−1 . (5.52) Thus I(θ) = 1/[1 + 4R sin2 (δ/2)/(1 − R)2 ].
(5.53)
This function is plotted against δ in ﬁgure 5.31 for a few values of R, and it shows that as R increases the bright fringes sharpen while retaining their peak intensities. The degree of sharpness is quantiﬁed by calculating the full width at half the maximum intensity (FWHM) of bright fringes. Suppose the value of δ at one peak of intensity is 2πm and that its value when the intensity has fallen oﬀ to half maximum is 2πm + δ1/2 , then 1 + sin2 ( δ1/2 /2) [ 4R/(1 − R)2 ] = 2, whence
√ δ1/2 = 2 sin−1 [(1 − R)/2 R)] √ ≈ (1 − R)/ R,
1
r2tt1 tt1
t1
Fig. 5.30 Successive contributions to the wave amplitude transmitted by a Fabry–Perot etalon.
124 Interference eﬀects and interferometers
1 0.9 0.8
R=0.04
F=0.17
R=0.2
F=1.25
R=0.5
F=8
R=0.85
F=151
Intensity
0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 0
0.2 0.4 0.6 0.8
1 1.2 1.4 1.6 1.8 δ/2π
2
Fig. 5.31 Intensity distribution against δ/2π = 2d cos θ/λ. The intensity is normalized to the incident intensity.
because all the √ angles are small. Thus the FWHM of each fringe, 2δ1/2 , is 2(1 − R)/ R. The corresponding FWHM in frequency of each fringe is obtained using eqn. 5.49 2∆f1/2 = cδ1/2 /(2πnd).
(5.54)
A convenient measure of the sharpness called the ﬁnesse is deﬁned as the ratio between the phase change separating successive fringes and the phase change across the fringe at half maximum, 2δ1/2 . The ﬁnesse is thus √ F = π/δ1/2 = π R/(1 − R). (5.55) When the reﬂection coeﬃcient R is close to unity, it follows5 that RF ≈ exp (−π) ≈ 0.04, which is similar in magnitude to the reﬂection coeﬃcient at an uncoated glass/air interface. Thus F can be treated as the number of round trips within the etalon, beyond which the wave intensity becomes insigniﬁcant. Any lack of ﬂatness in the etalon surfaces will be ampliﬁed by this factor so that the ﬁnesse expected with a given reﬂection coeﬃcient can only be actually attained if the ﬂatness is better than λ/F. A reﬂection coeﬃcient of 0.9 and a ﬁnesse of 30 represent standard values. The narrowness of the Fabry–Perot fringes makes this interferometer a useful spectrometer. The chromatic resolving power of a spectrometer 5 ln RF
ln (1 − S) π
√
= ln (R) π R/(1 − R). Now putting S = 1 − R we get ln RF = (1 − S)/S. Then because S is very small, ln RF ≈ −π.
5.9
is deﬁned to be the ratio of the wavelength to the smallest separation in wavelength that can be resolved with the apparatus. In the case of the Fabry–Perot instrument two wavelengths, λ and λ − ∆λ, will just be resolvable when their mth order bright fringes are separated by 2δ1/2 . If, close to the centre of the pattern, the interference order for wavelength λ is m, then mλ = (m + δ1/2 /π)(λ − ∆λ). Then to a good approximation πm∆λ = λ δ1/2 , and the chromatic resolving power CRP = λ/∆λ = πm/δ1/2 .
(5.56)
Now using the deﬁnition of ﬁnesse from eqn. 5.55 CRP = Fm. Using eqn. 5.16 to replace m and recalling that near the centre of the pattern cos θ ≈ 1 then gives CRP = 2ndF/λ.
(5.57)
Taking a 10 mm spacing in air and a ﬁnesse of 30 gives a resolving power of 106 at 633 nm wavelength. Now the diﬀerence in wavelength between individual spectral lines produced by most sources is large enough that adjacent circular fringes due to two spectral lines can correspond to very diﬀerent values of m in eqn. 5.16; that is to say they are from diﬀerent orders. The wavelength interval within which there is no overlap of orders and hence no ambiguity in assigning the order of a line of unknown wavelength is called the free spectral range. Clearly the increase in δ over the free spectral range is 2π, and the corresponding frequency change can then be obtained from eqn. 5.49 ∆ffsr = c/(2nd).
(5.58)
∆λfsr = λ2 ∆ffsr /c = λ2 /(2nd).
(5.59)
Then With a 1 mm etalon spacing and a wavelength of 633 nm the free spectral range is only 0.2 nm which shows that one needs to preselect a narrow wavelength interval in order to determine an unknown wavelength. Preselection can be achieved in a two stage spectrometer, with the ﬁrst stage being a grating or prism spectrometer or alternatively a bandpass ﬁlter of the sort to be described in Chapter 9. A twostage spectrometer using a constant deviation prism of the type described in Section 2.2.1 is shown in ﬁgure 5.32. Light from a slit source is focused into a parallel beam by the collimator and falls on the constant deviation prism. This gives an outgoing beam that is observed by a telescope. For any
The Fabry–Perot interferometer
125
126 Interference eﬀects and interferometers
Monochromatic source and slit Collimator
Telescope
FabryPerot etalon
Constant deviation prism
View of fringes
Fig. 5.32 Two stage spectrometer using a constant deviation prism and a Fabry– Perot etalon. In the insert the Fabry–Perot fringes across two spectral lines are shown.
given orientation of the constant deviation prism the telescope receives a narrow range of wavelengths centred on the wavelength for which that orientation produces minimum deviation. This window in wavelength can be scanned through the spectrum simply by rotating the constant deviation prism. The etalon is placed to intercept the beam from the prism and the resulting image produced by the telescope is observed by eye or let fall on a detector array. The inset in ﬁgure 5.32 shows the ﬁnal image. With the Fabry–Perot etalon removed this would consist of repeat images of the source slit, one for each spectral line resolved by the prism. With the Fabry–Perot etalon in place one now sees through each slit image a diametral slice of the circular Fabry–Perot pattern for the spectral line involved. If the spectral line is in fact two close lines of wavelengths λ1 and λ2 which cannot be resolved by the prism, then that particular slit image will bear two interlaced sets of Fabry–Perot fringes satisfying m1 λ1 = d cos θ1 and m2 λ2 = d cos θ2 . Another arrangement for carrying out spectroscopy with a Fabry– Perot etalon is shown in ﬁgure 5.33. In this case the wavelength of the incident beam is limited to less than the free spectral range by an interference ﬁlter of the sort described in Section 9.7. The detector views the central spot and a scan through the free spectral range in wavelength is made by altering the product nd in eqn. 5.16. One method is to pump gas into the volume between the etalons which requires knowledge of the pressure and temperature of the enclosed gas in order to evaluate the refractive index. The alternative scanning technique is to move one
5.9
of the etalon plates while keeping it parallel to its stationary partner. Movements of a few microns are adequate. The preferred scanning technique is to use piezoelectric crystals, which undergo a change in length in the direction of an applied electric ﬁeld. Three such crystals placed at 120◦ intervals around the rim of the spacer provide suﬃciently smooth parallel displacement of the moving plate. The size of the pinhole in front of the detector and etendue of the Fabry–Perot spectrometer will now be determined; the latter will be compared with the etendue of other spectrometers in the following chapter. If the etalon separation is such that at wavelength λ a fringe sits at the exact centre of the pattern,
The Fabry–Perot interferometer
127
Drive to piezoelectric spacer Pinhole
Source + filter Lens
Etalon
Lens Pinhole + detector
Fig. 5.33 Scanning Fabry–Perot spectrometer using a piezoelectric spacer.
2nd = mλ. The pinhole is made large enough to accept the half width of the central fringe. Suppose the angular radius of the pinhole as seen from the lens is θ, then 2nd cos θ = [m + δ1/2 /(2π)]λ. Subtracting the ﬁrst of these equations from the second and using an approximation adequate at these small angles, cos θ = 1 − θ 2 /2, gives ndθ2 = δ1/2 λ/(2π). Using eqn. 5.56 to replace δ1/2 , θ2 = m∆λ/(2nd) = ∆λ/λ,
(5.60)
which is the inverse of the CRP. Hence the solid angle subtended by the pinhole at the lens is Ω = πθ2 = π ∆λ/λ. (5.61) The etendue is then the product of this solid angle and the area of the beam at the lens T = (πD/2)2 (∆λ/λ). (5.62) where D is the diameter of the clear portion of the etalon.The reader will recall that the etendue is invariant along a beam, and may like to check that taking the product of the area of the pinhole times the solid angle of the beam at the pinhole gives an identical result.
Exercises (5.1) In Young’s experiment the slit spacing is 0.1 mm and monochromatic light of wavelength 633 nm is
used. What is the fringe spacing if the image focusing lens has focal length 1.5 m?
128 Interference eﬀects and interferometers
Source and pinhole
Slits
Gas filled tube
Evacuated tube
Fringes
Fig. 5.34 The Rayleigh refractometer.
(5.2) The ﬁrst lens in Young’s experiment shown in ﬁgure 5.4 has focal length 0.3 m. How narrow should the source aperture be in order to ensure that the slits are coherently illuminated? (5.3) A source emits light of mean wavelength 500 nm. The wavepackets are 1 m long. What is the spread in frequency and in wavelength? What fraction is the frequency spread of the mean frequency? What is the corresponding fraction in wavelength? (5.4) Two optically ﬂat glass plates are viewed in light of wavelength 633 nm. Straight line fringes are seen, spaced apart at intervals of 1.5 cm across the surface. What is the angle between the plates? (5.5) Show that at latitude 50◦ N the Earth’s rotation causes a frequency splitting of 176 Hz between counterrotating 633 nm laser beams in an active ring laser. The beam circuit can be assumed circular of radius 1 m. (5.6) Suppose a highly coherent laser beam of wavelength 633 nm is interrupted by an electronic shutter so that random pulses each of duration 10 fs are produced. What is the coherence length and what are the spreads in wavelength and frequency of the beam?
slowly into one arm. This causes the fringes to move across the ﬁeld of view; up or down in the diagram. If m fringes pass through the centre of the ﬁeld while the gas is entering, show that the refractive index of the gas, n, is given by n − 1 = mλ/L. (5.8) In the Rayleigh refractometer the slits need to be well separated, which means the fringes are very close together. Suppose the wavelength of light is 500 nm, the slit separation is 2 cm and an imaging lens of 10 cm focal length is used to bring the beams together. What is the fringe separation and how would you view them? (5.9) Calculate the ﬁnesse of a Fabry–Perot etalon whose plates have a reﬂection coeﬃcient of 0.95 at 500 nm. If the plates are 3 mm apart in air what is the free spectral range? What is the potential chromatic resolving power obtainable at this ﬁnesse? What degree of optical ﬂatness in the surfaces of the etalon is needed in order to exploit this potential? The etalon has a clear region of diameter 2.5 cm. Suppose the etalon is used as shown in ﬁgure 5.33 with the imaging lens having a focal length of 5 cm. What pinhole diameter is required on the imaging screen to cover the central fringe? What is the system etendue as deﬁned in Section 4.2?
(5.7) The accompanying ﬁgure 5.34 shows a Rayleigh refractometer. Light from coherently illuminated slits goes through separate hollow airtight arms of length L each with glass entry and exit windows. The separate beams are brought together by a lens (5.10) Find an expression linking the free spectral range, with fringes appearing in the image plane. Both ﬁnesse and FWHM of the fringes of a Fabry–Perot arms are initially evacuated and gas is introduced etalon.
Diﬀraction 6.1
Introduction
Diﬀraction is taken to mean any interference eﬀect due to the interruption of a wavefront by apertures or obstacles, often disposed in regular arrays. The pattern of illumination is very diﬀerent when the plane of observation is near to the diﬀracting surface and when it is a large distance away. Near the diﬀracting surface the pattern is the geometric shadow with fringes close to the shadow edges, and as the plane of observation moves further away this pattern changes smoothly into one that has lost any obvious resemblence to the geometric shape of the apertures and obstacles producing it. In the limit that the source and image plane are inﬁnitely far from the diﬀracting surfaces, the pattern is called Fraunhofer diﬀraction. In practice this limiting condition is simple to produce: the source is placed at the focus of one positive lens and the plane of observation in the focal plane of another positive lens. Then the source and observing plane are eﬀectively at an inﬁnite distance from the diﬀracting apertures. Diﬀraction at ﬁnite distances is called Fresnel diﬀraction. Both Fresnel and Fraunhofer diﬀraction will be treated using the the Huygens–Fresnel picture of secondary waves introduced in the previous chapter. Fraunhofer diﬀraction is easier to analyse and is the basis of many research tools and technological applications; it will therefore receive more attention than Fresnel diﬀraction. The ﬁrst section below contains a discussion of the theoretical basis of the Huygens–Fresnel picture of diﬀraction. In the following sections the analysis of diﬀraction at a single long wide slit is presented, and then that for a rectangular slit. This is followed by a treatment of diﬀraction by arrays of equally spaced, identical slits. In the limit of very large numbers of slits these arrays are called diﬀraction gratings. Next the diﬀraction at a circular aperture is described, which leads to the formula previously used in Chapter 4 to calculate the resolving power of optical systems. Grating spectrometers are described and their performance compared to prism and Fabry–Perot instruments. After this Fresnel diﬀraction at a long broad slit is treated. It is shown how the diﬀraction pattern evolves, as the distance from the aperture increases: from a geometric shadow to a Fraunhofer pattern. Fresnel diﬀraction at a circular aperture is then treated, including a discussion of Fresnel zones and zone plates. Then a section is devoted to the schemes used in optical lithography to acheive the high resolution in electronic chip manufacture. After this
6
130 Diﬀraction
some comments are made about the limitations of the theory underlying the analysis presented in this chapter when applied to regions very close to the apertures. In the ﬁnal part of the chapter simple Gaussian laser beams and the role of diﬀraction in their evolution in optical systems are described.
6.2
A
rin
The simple Huygens–Fresnel picture of interference between secondary waves presented in the previous chapter has a fundamental diﬃculty. Apparently the construction with secondary waves could equally well lead to a backward going wave as to a forward going wave. This diﬃculty was resolved by Kirchhoﬀ’s analysis of wave propagation at apertures, which is described in Appendix C. Kirchhoﬀ obtained the following expression for the spatial part of the amplitude at P of a secondary wave originating from a point A in an aperture which is illuminated by a small monochromatic source at S of area dS as shown in ﬁgure 6.1
θin θout rout P
S (source)
Huygens–Fresnel analysis
Opaque screen
Fig. 6.1 Path lengths and angles used in Kirchhoﬀ’s analysis.
E = C exp [ik(rout + rin )] [(cos θin + cos θout )/2] dS/(rout rin ).
(6.1)
Here k is the wave number of the radiation, rin is the length of the path from the source to A and θin the angle this makes with the normal to the aperture surface at A. rout and θout are the corresponding parameters for the path from A to the point P where the amplitude is observed. C is a constant that depends on the source intensity. The inclination factor (cos θin + cos θout )/2 is unity in the forward direction and falls to zero in the backward direction, and it is this which eliminates a backward propogating secondary wave. 1/rin and 1/rout are the range factors in spherical waves which ensure that the total ﬂux remains constant as they expand. If the incident wave is planar then eqn. 6.1 reduces to E = D exp [ikrout ] [(1 + cos θout )/2] /rout ,
(6.2)
where D is another constant that depends on the source intensity. Wherever the inclination and range factors change little over the aperture and image region they only aﬀect the overall magnitude and not the pattern of interference. In such cases they can often be factored out of the analysis.
6.3
Single slit Fraunhofer diﬀraction
Figure 6.2 shows the standard experimental layout used to produce Fraunhofer diﬀraction. Incident plane waves are produced by placing a monochromatic source at the focus of a positive lens. These waves are incident normally on an opaque plane sheet pierced by a single long slit. Beyond this sheet another positive lens images the diﬀraction pattern onto a screen in its focal plane. The use of lenses brings the source and image plane in from inﬁnity, so making a compact experimental setup with which to observe Fraunhofer diﬀraction.
6.3
Single slit Fraunhofer diﬀraction
The incident wave across the broad slit of width d gives rise to Huygens’ secondary waves whose resultant at the screen can be calculated. The path lengths to P will diﬀer by the distance each slit element is from
d A θ
P θ
A’ B θ
d sin
θ
f0
f
Fig. 6.2 Fraunhofer diﬀraction at a ﬁnite width slit.
the line AB, drawn perpendicular to the rays travelling to P. For light from an element of the slit of width dx at a distance x from A the extra path length compared to light from A is x sin θ and the corresponding phase delay is φ(x) = 2πx sin θ/λ = kx sin θ. In the paraxial approximation the phase delay is kxx /f where x is the lateral coordinate of P: an expression symmetric in x and x which will prove useful later. Thus the Huygens wave from an element (x,dx) of the slit gives a contribution to the wave at P dEp (x) = E0 dx exp [ i(ωt − ks − φ) ], where s is the optical path length from A to P, and E0 is a constant expressing the contribution to the ﬁeld per unit width of the slit. Summing these contributions gives the total amplitude Ep (θ), from which we get the intensity at P I(θ) = Ep∗ (θ)Ep (θ). Evidently the common factor exp i(ωt − ks) disappears in the intensity
131
132 Diﬀraction
Im C α R
R
S
E0
φ
dx
α
Re
O Fig. 6.3 Phasor diagram for Fraunhofer diﬀraction at a ﬁnite width slit. The short arrow is a representative phasor due to a short segment of the slit width. The resultant phasor amplitude for the whole slit is the chord length OS.
so it is only needful to add up the phasors E0 dx exp (−iφ). Then Ep (θ) =
d
E0 exp (−ikx sin θ)dx
(6.3)
0
= E0 [1 − exp (−ikd sin θ)]/(ik sin θ) = E0 d exp (−ikd sin θ/2) sinc(kd sin θ/2),
1
I( θ ) / I(0)
0.8
where as usual sinc(x) = sin x/x. Thus the intensity is
0.6
I(θ) = (E0 d)2 sinc2 (kd sin θ/2).
0.4 0.2 0 3
2
1 0 1 d sin θ / λ
2
3
Fig. 6.4 Intensity distribution for Fraunhofer diﬀraction at a ﬁnite width slit.
(6.4)
This calculation of Ep (θ) at P is expressed diagramatically in the phasor diagram, ﬁgure 6.3. The contribution to the amplitude at P from an element of the slit located at (x,dx) has magnitude E0 dx and has phase angle kx sin θ. Thus its phasor has length E0 dx and is inclined at an angle φ = kx sin θ to the real axis. Adding all the phasors vectorially gives a circular arc which turns through an angle α = kd sin θ and has radius R = E0 d/α = E0 /k sin θ. The resultant amplitude at P is the chord of this arc Ep (θ) = 2R sin (α/2) = E0 d sinc(kd sin θ/2)
(6.5)
6.4
Diﬀraction at a rectangular aperture 133
as before. Whenever the arc makes one or more complete circles the resultant intensity is zero. For this to be the case kd sin θ = 2nπ, which reduces to d sin θ = nλ
(6.6)
where n is a nonzero integer. On approaching the forward direction sin θ → 0, and applying l’Hopital’s rule Limitθ→0 [ sin θ/θ ] = 1.
(6.7)
Thus there is a maximum in intensity in the forward direction for which eqn. 6.4 gives I(0) = (E0 d)2 . (6.8) Then eqn. 6.4 can be rewritten compactly I(θ) = I(0) sinc2 (kd sin θ/2).
(6.9)
Figure 6.4 shows this intensity distribution calculated for Fraunhofer diﬀraction at single long slit. The outer bright fringes are half as wide as the central one. Their peaks lie nearly midway in angle between the minima and the ﬁrst two have intensities only 0.047 and 0.017 of the forward intensity. Actual diﬀracting screens have twodimensional apertures and the analysis for onedimensional slits is easily extended to two dimensions. The resulting patterns are simple only where there is some symmetry in the apertures, and of these cases the rectangular and circular apertures are treated below.
6.4
Diﬀraction at a rectangular aperture
Up to this point it has been assumed in calculating the eﬀects of interference and diﬀraction that the slit length, L, is so large that λ/L is eﬀectively zero. This excludes any diﬀraction in the direction of the slit length so that the diﬀraction pattern is two dimensional, lying in the plane perpendicular to the slit length. The calculation of diﬀraction will now be extended to a rectangular aperture with limited length and width. Monochromatic plane waves are incident normally on an aperture of width dx in the xdirection and dy in the ydirection. Let ex , ey and ez be the orthogonal unit vectors in and perpendicular to the aperture plane. Then consider the Huygens’ wave from an element of the aperture of area dx dy located at r = xex + yey . The phasor amplitude at a point P which lies in the direction given by the wave vector k = kx ex + ky ey + kz ez
134 Diﬀraction
will be dE(k) = E0 exp (−ik · r)dx dy = E0 [ exp (−ikx x) dx ] [ exp (−iky y) dy ]. The total amplitude at P is obtained by integrating this expresion over x and y. These integrals are independent and have been evaluated in the previous section. Reusing eqn. 6.9 gives for the total intensity at P I(k) = I(0)sinc2 (kx dx /2)sinc2 (ky dy /2),
(6.10)
where I(0) is the intensity in the forward direction. The diﬀraction pattern is simply the product of those for the two dimensions separately. This distribution is shown in ﬁgure 6.5 for a rectangular aperture twice as tall as it is wide. Correspondingly the fringes are twice as wide as they are tall. The ﬁgure shows how the image would appear on a photographic negative or CCD that has been deliberately overexposed in order to enhance the weaker intensity peaks that lie both along and oﬀ the axes. Fig. 6.5 Intensity pattern for Fraunhofer diﬀraction at a rectangular aperture.
6.5
Diﬀraction from multiple identical slits
The experimental arrangement shown in ﬁgure 6.2 is a template for producing Fraunhofer diﬀraction using an opaque sheet with any choice of apertures. A very useful arrangement is to have a row of identical slits, d wide, regularly spaced a distance a apart centretocentre, as shown in ﬁgure 6.6. Analysis of the diﬀraction produced by such an array will prepare the ground for the discussion of diﬀraction gratings and spectrometers. In the remainder of this section we suppose that the slits are suﬃciently long that the variation in the diﬀraction pattern is all along a line perpendicular to the slit lengths. It is seen in ﬁgure 6.6 that the path length from a slit centre to P changes by a sin θ between successive slits. Correspondingly the phase diﬀerence of the secondary waves arriving at P from the slit centres is β = ka sin θ.
(6.11)
The phasor addition of the amplitudes at P is illustrated for the case of three slits in ﬁgure 6.7. In the upper diagram the dotted line arcs show the phasor contributions of the elements in each slit; the resultants of individual slits are the phasors drawn with open arrowheads; and their resultant is the long phasor ending in a solid arrowhead. In the lower diagram the phasors of the three slits are drawn for four choices of β: namely 0, 2π/3, π, 4π/3 and 2π. If the amplitude at P with a single slit open is unity, then these four choices for β yield amplitudes 3, 0, 1, 0 and 3 respectively, while the corresponding intensities are 9, 0, 1, 0 and 9. As β changes between these conﬁgurations the intensity is either falling or rising monotonically. Some simple conclusions can be inferred from ﬁgure 6.7 about the pattern of fringes seen when there are N slits.
6.5
Diﬀraction from multiple identical slits
135
θ a P θ a θ a
f
θ a sin
Im
Fig. 6.6 Fraunhofer diﬀraction at an array of equally spaced identical slits.
• There are principal maxima whose intensity is N 2 times that of a single slit maximum. • There are N − 2 much lower intensity subsidiary maxima between each pair of principal maxima. The same results will now be obtained analytically. The wave at P due to the mth slit is
β Re O
β β=O
o
Ep (θ) exp [ −i(m − 1)ka sin θ ], where Ep (θ) is the single slit contribution given in eqn. 6.5. Making use of eqn. 6.9 the intensity at P due to all N slits is 2
IN (θ) = (E0 d) sinc
2
∗ (kd sin θ/2)XN XN ,
(6.12)
β = 120 β = 180 β = 240
where N
1 − exp (−iN ka sin θ) 1 − exp (−ika sin θ) m=1 sin (N ka sin θ/2) , = exp [−i(N − 1)ka sin θ/2] sin (ka sin θ/2)
XN =
exp [ −i(m − 1)ka sin θ ] =
where we have used the result N m=1
xm−1 = (1 − xN )/(1 − x).
β = 360
o
o
o
o
Fig. 6.7 Phasor diagrams for Fraunhofer diﬀraction at three equally spaced identical slits. The upper panel shows the the slit phasors and their resultant. Below the alignments for intensity minima and maxima appear.
1
1
0.8
0.8 I2(θ) / 4 I(0)
I( θ) / I(0)
136 Diﬀraction
0.6 0.4 0.2
0.4 0.2
0 2 1.5 1 0.5 0 0.5 d sin θ / λ
1
1.5
0 2 1.5 1 0.5 0 0.5 d sin θ / λ
2
1
1
0.8
0.8 I4(θ ) / 16 I(0)
I3(θ ) / 9 I(0)
0.6
0.6 0.4 0.2
1
1.5
2
1
1.5
2
0.6 0.4 0.2
0 2 1.5 1 0.5 0 0.5 d sin θ / λ
1
1.5
2
0 2 1.5 1 0.5 0 0.5 d sin θ / λ
Fig. 6.8 Fraunhofer diﬀraction patterns for one, two, three and four slits. The slit widths and spacing are the same in each case. The single slit pattern is shown as a dotted line in the multislit plots.
Thus ∗ XN XN =
sin (N ka sin θ/2) sin (ka sin θ/2)
2 .
Substituting this result in eqn 6.12 gives the intensity at P 2 sin (N ka sin θ/2) 2 2 IN (θ) = (E0 d) sinc (kd sin θ/2) . sin (ka sin θ/2) Writing this more succinctly IN (θ) = I(0) sinc2 (α/2)
sin (N β/2) sin (β/2)
2 (6.13)
where I(0) = (E0 d)2 is the forward intensity due to a single slit, and where we repeat that α = kd sin θ; β = ka sin θ,
6.6
Babinet’s principle 137
d being the slit width and a the slit spacing centre to centre. We see that the intensity pattern for N slits contains a multiple slit pattern MN (θ) = sin2 (N β/2)/ sin2 (β/2) which modulates the single slit intensity pattern I(0) sinc2 (α/2). MN (θ) has zeroes wherever N β/2 = mπ, where m is an integer, with the important exception that whenever m/N is also equal to an integer p, β/2 = pπ.
(6.14)
In this case l’Hopital’s rule gives sin2 (N β/2)/ sin2 (β/2) → N 2 ,
(6.15)
with the result that the principal maxima are N 2 times brighter than the single slit maximum. These principal maxima occur at angles given by a sin θ = pλ. (6.16)
Principal maximum (N)
First minimum (0)
Diﬀraction patterns with one, two, three and four identical equally spaced slits are shown in ﬁgure 6.8, all with the same slit width and spacing. In the multislit diagrams the single slit pattern scaled up by a factor N 2 is shown as a dotted line. The intensities at the principal maxima touch this envelope, while the subsidiary maxima are much weaker. Figure 6.9 shows phasor diagrams when there are a large number of slits, in this case 37 slits. Reading from the top panel down these diagrams relate to the forward principal maximum, the adjacent minimum, the ﬁrst subsidiary maximum, and another principal maximum lying midway between principal maxima. In each case the resultant phasor amplitude is written in brackets. The resultant phasors are drawn with full arrowheads for the two subsidiary maxima. With each slit contributing unit intensity the ﬁrst subsidiary maximum has intensity D2 = (2N/3π)2 , which is 0.045 that of a principal maximum. Thereafter the subsidiary maxima decline in intensity with those midway between the principal maxima having about the same intensity as a single slit maximum.
6.6
Babinet’s principle
The diﬀraction pattern produced by an opaque screen with any arrangement of apertures is related to that produced by the complementary
First subsidiary maximum (D)
D = 2N/3π
Middle subsidiary maximum (1)
Fig. 6.9 Phasor diagrams for Fraunhofer diﬀraction produced by a large number of slits.
138 Diﬀraction
screen: this complementary screen is transparent wherever the ﬁrst screen is opaque and opaque wherever the ﬁrst screen is transparent. Suppose that at P, lying in the focal plane of the second lens in ﬁgure 6.2, the light amplitude with the ﬁrst screen in place is A1 ; and A2 when the complementary screen replaces it. Also suppose the intensity at P when both screens are removed is A0 . Then clearly A0 = A1 + A2 . However A0 is only nonzero in the exact forward direction, that is when P lies on the optical axis. Elsewhere A2 = −A1 . It follows that the intensities produced by complementary screens are identical except in the forward direction and that there they add to give the unobstructed intensity. This is known as Babinet’s principle.
6.7
y
s
x k
φ θ
z
Fraunhofer diﬀraction at a circular hole
The circular aperture, of radius r, shown in ﬁgure 6.10 is used to produce Fraunhofer diﬀraction in the standard experimental arrangement. Light from a point s on the aperture with wave vector k travels to P in the image plane. There is phase diﬀerence of k·s at P compared to light also travelling to P from the centre of the aperture. The vectors involved are s = s cos φ ex + s sin φ ey , k = k cos θ ez + k sin θ ex ,
Fig. 6.10 Fraunhofer diﬀraction at a circular aperture. The vectors k and s point from the hole’s centre toward a point on the image and to an element of the aperture respectively.
where ex , ey and ez are unit vectors along orthogonal axes drawn in the diagram. Thus k · s = ks sin θ cos φ, and the wave at P is
r
2π
E = E0
exp (iks sin θ cos φ)s dφds, 0
(6.17)
0
where E0 includes constants and also the complex exponents that vanish when the intensity is calculated. Integration over φ gives a Bessel function of order zero1 r E = 2πE0 J0 (ks sin θ)sds, 0 1 Table of Integrals, Series and Products by I.S. Gradshteyn and I.M. Ryzhik, edited by A. Jeﬀrey, 5th edition 1994; published by Academic Press, London.
2π
exp (iz cos φ)dφ = 2πJ0 (z), 0
r
zJ0 (z)dz = rJ1 (r). 0
6.8
Diﬀraction gratings 139
and the integral over s yields a Bessel function of order one E = 2πr2 E0 J1 (kr sin θ)/(kr sin θ).
I( θ) / I(0)
Finally the intensity at P is 2 4
I(θ) = 4π r
1
E02 [J1 (kr sin θ)/(kr sin θ)]2 .
where D is the diameter of the hole. Lord Rayleigh proposed what is now the standard deﬁnition of the limit of the resolving power of a lens system: using an optical system with entrance pupil of diameter D, two point objects are resolvable if their angular separation exceeds 1.22λ/D. At the limit the intensity maximum of one object’s image would lie at the ﬁrst minimum of the other object’s image. This Rayleigh criterion was already used in Chapter 4 to evaluate the resolution of optical instruments.
6.8
Diﬀraction gratings
It follows from the analysis of Fraunhofer diﬀraction by multiple slits that when the number of slits becomes large the principal maxima become very narrow, while the subsidiary maxima are so weak as to be undetectable. Opaque screens with large numbers of identical equally spaced slits are known as diﬀraction gratings and are widely used in studying the spectra emitted by sources. The key point is that the principal maxima are so narrow that the principal maxima of spectral lines of closely similar wavelength are separate and distinct: they are resolved. The chromatic resolving power of spectroscopic devices is deﬁned as CRP = λ/∆λ,
(6.20)
where ∆λ is the smallest diﬀerence in wavelength at which it is possible to separate two spectral lines at wavelength λ. The practical limit occurs for gratings when the principal maximum for wavelength λ + ∆λ coincides in angle with the minimum adjacent to the same maximum for wavelength λ. In the case of the pth order maximum produced by light of wavelength λ on a grating with N slits N pλ = N a sin θ,
5 0 5
0
5
10
inθ krs Fig. 6.11 The intensity distribution for diﬀraction at a circular hole. It is also shown in projection on the roof of the box as it might appear on a photographic negative with only the Airy disk visible.
θ
(6.19)
0 10
sin
sin θ = 0.61λ/r = 1.22λ/D,
0.5
kr
Noting that in the forward direction this expression reduces to π 2 r4 E02 , the intensity in the diﬀraction pattern of a circular hole can be rewritten as 2 2J1 (kr sin θ) I(θ) = I(0) . (6.18) (kr sin θ) This function is plotted in ﬁgure 6.11. The bright central spot is named the Airy disk, after the 19th century astronomer who was the ﬁrst to calculate this distribution. 84% of the light falls within the Airy disk and its angular radius θ is given by
10 10
5
140 Diﬀraction
and at the adjacent minimum (N p + 1)λ = N a sin (θ + ∆θ).
(6.21)
If this minimum coincides with the pth order principal maximum for wavelength λ + ∆λ it follows that N p(λ + ∆λ) = N a sin (θ + ∆θ).
(6.22)
Subtracting eqn. 6.21 from eqn. 6.22 gives N p ∆λ − λ = 0, whence λ/∆λ = N p.
(6.23)
It is worth emphasizing that N is the number of grating lines illuminated by the source being studied. Lines out of the beam can play no part in diﬀracting the beam. There can be confusion when fringes of one order overlap those of the adjacent order; which will happen if the range of wavelengths in the incident radiation is large. In order to avoid overlaps between the the ﬁrst and second order principal maxima the spread of wavelengths must be such that 2λmin > λmax , ∆y
Source f
Detector
where λmin and λmax are the shortest and longest wavelengths. The widest permissible spread is called the free spectral range of the grating, which in this case is λfsr = λmin /2 (f irst order).
(6.24)
For any higher order, p, overlaps with both the adjacent orders, p − 1 and p + 1, must be avoided. In this case λfsr = λmin /(p − 1) (higher order). α θ Grating Na
Fig. 6.12 Reﬂection grating.
(6.25)
Gratings met in research and industry are nearly always reﬂection gratings because these are straightforward to manufacture with line densities up to thousands per millimetre. Figure 6.12 shows a parallel beam being reﬂected from the faces of the reﬂective elements, which take the place of slits in a reﬂection grating. The path diﬀerence between light reﬂected from adjacent elements is now a(sin θ − sin α) where α is the angle of incidence and θ the angle of reﬂection. This is to be compared to a delay of a sin θ in the case of a transmission grating with light incident normally. Therefore the sole change required in the analysis carried through for a transmission grating to make it applicable to a reﬂection grating is to replace sin θ everywhere by (sin θ − sin α). The equation giving the angular location of principal maxima becomes a(sin θ − sin α) = pλ.
(6.26)
6.9
6.9
Spectrometers and spectroscopes 141
Spectrometers and spectroscopes Frame
An instrument used to view spectra by eye is called a spectroscope, while one employed with any sort of electronic detector is called a spectrometer. One simple design is shown in ﬁgure 6.13. A massive cylindrical frame supports a central rotating table designed to carry the dispersing element, which could either be a reﬂecting grating or a prism. Two arms protrude from the frame. One arm is ﬁxed rigidly to the frame and carries a collimator with a variable width entry slit. Light from a source illuminating this slit emerges as a parallel beam onto the grating on the central table. The second arm carries a telescope which receives the light from the reﬂecting grating. It brings the light to a focus at a slit, behind which the detector is placed. This telescope arm can rotate independently about the central vertical axis, and carries a vernier that travels, as the arm moves, over a graduated angle scale that runs around the edge of the frame. A comparison will now be made between the chromatic resolving powers obtainable with such a spectrometer using in one case a grating and in the other a prism. The chromatic resolving power of a prism is, unexpectedly, limited by diﬀraction. In ﬁgure 6.14 a parallel beam containing light of two nearby wavelengths is dispersed by a prism at minimum deviation: these wavelengths are λ and λ + ∆λ. The prism has vertex angle α and it is assumed the whole face is illuminated down to the base, which has length s. BD and BE are wavefronts for the two wavelengths after dispersion, and w is the width of these beams. When working at minimum deviation ˆ = (α + δmin )/2 ≡ θ. Then eqn. 2.26 gives the refractive the angle CBE index n = sin θ/ sin (α/2). (6.27) Diﬀerentiating eqn. 6.27 with respect to wavelength gives dn/dλ = cos θ[ dδmin /dλ ]/[ 2 sin (α/2) ].
(6.28)
The two wavelengths will be resolvable provided that the change in the deviation in θ between them is greater than the angular width of the diﬀraction peak for a slit of width w. In this limit ∆δmin = λ/w.
(6.29)
If ∆λ is the precise change in wavelength that produces a change in minimum deviation λ/w we have ∆δmin /∆λ = λ/(w∆λ).
(6.30)
∆λ is small so we can replace dδmin /dλ by ∆δmin /∆λ in eqn. 6.28 giving dn/dλ = λ cos θ/[ 2w∆λ sin (α/2) ]
(6.31)
Rearranging this equation gives the chromatic resolving power of the prism λ/∆λ = [2w sin (α/2)/ cos θ](dn/dλ) = s(dn/dλ).
(6.32)
Grating Rotating telescope
Scale
Fixed collimator Slit Source
Fig. 6.13 Simple spectrometer or spectroscope design. C δ/2 α
E
D w
(α + δ)/2
(π α )/2 s
∆δ B
Fig. 6.14 Separation of wavelengths λ and (λ + ∆λ) by a prism at minimum deviation. The incoming and outgoing beams are then symmetric. For clarity δmin is abbreviated to δ on this diagram.
142 Diﬀraction
2
For reference a comparison is made between the chromatic resolving power of a 5 cm wide grating with 1200 lines/mm and that of a prism of base length 5 cm made of DF ﬂint glass2 with high dispersion dn/dλ = 10−4 nm−1 . Then using eqn. 6.23 and the above equation, we have
See Table 3.1.
(λ/∆λ)grating = 60000 (λ/∆λ)prism = 5000 Gratings are cheaper to make and simpler to handle so they are preferred for most applications. The chromatic resolving power of gratings falls far short of that obtained with a Fabry–Perot etalon. However the free spectral range of an etalon is so narrow that often a two stage instrument must be used to avoid confusion between orders: the ﬁrst stage uses a prism or grating to select a narrow band of wavelengths which are made the input to the etalon.
6.9.1
Normal to grating plane
φ
Normal to reflecting slit
Fig. 6.15 Blazed grating surface.
Grating structure
The ﬁrst useful gratings were produced by mechanically engraving lines on metal. Light is reﬂected oﬀ one face of the Vshaped grooves, which act as the slits. Defects, in the form of cyclic variations in the depth or spacing of grooves, were hard to avoid and resulted in low intensity satellites close to the principal maxima, known as ghosts. These could be confused with real but weak lines in the spectrum. Nowadays the position of the engraving tool is monitored and controlled using an interferometer which makes it possible to eﬀectively eliminate such cyclic or random errors. The master metal gratings are used as moulds from which polymer replica gratings are cast, after which a ﬁlm of aluminium is deposited on the replicas. The newer holographic gratings are produced photographically. First a polymer sheet coated with photoresist is exposed to an interference pattern formed by intersecting UV laser beams and the resist is broken down at locations of high intensity in the interference pattern. Afterwards the resist surface is chemically etched to remove the degraded material, leaving a rippled surface that forms the grating. Finally aluminium is deposited on the grating. This holographic process gives ﬁner and more regular line spacing, and in addition the resulting gratings are freer of random defects than the ruled gratings. 3600 lines/mm is a standard line density achieved in holographic gratings compared to 1200 lines/mm with ruled gratings. Most of the light incident on a grating will end up in the central zero order fringe which is at the same location for all wavelengths. The light going into this central white fringe is wasted as far as any spectroscopic study is concerned. The situation can be improved in the case of ruled gratings by shaping the scribing tool so that the groove crosssection has the appearance shown in ﬁgure 6.15. With this proﬁle the normal to each reﬂecting facet is now inclined at an angle φ with respect to the normal to the plane of the grating and hence the peak of the single slit
diﬀraction pattern lies in this directon also. This process is known as blazing and φ is called the blaze angle. Blazing leaves the directions of principal maxima unchanged because the plane containing the slit centres is unchanged. The diﬀraction pattern (for ten slits) is shown both without and with blazing in ﬁgure 6.16. Evidently blazing at an appropriate angle can improve the brightness of the ﬁrst principal maxima by a very big factor. The wavelength whose ﬁrst principal maximum on one side lies exactly at the centre of the rotated single slit envelope is called the blaze wavelength, and is given by λblaze = a sin φ.
Spectrometers and spectroscopes 143
1 0.8 I( θ ) / I(0)
6.9
0.4 0.2 0 2 1.5 1 0.5 0 0.5 d sin θ / λ
(6.33)
Holographic gratings can be blazed to a limited extent by etching the grating with an ion beam.
1
1.5
2
1
1.5
2
1 0.8
Etendue
The etendue is calculated for the instrument shown in ﬁgure 6.12. Collimator and telescope are represented as single lenses with identical focal lengths f , the collimator slit width along the direction of dispersion is ∆y, and its length is w. The grating is taken to have N lines with interval a and to be a square of side length W = N a. Etendue is invariant through the system so it can be calculated at any convenient aperture; here the beam at the collimator slit is considered. Only light passing through the area of the grating is useful, and its area projects to cover an area W 2 cos α of the collimator lens. This active area at the lens subtends a solid angle at the collimator slit Ω = W 2 cos α/f 2 .
(6.34)
Ideally the slit width should subtend an angular width at the grating equal to the FWHM of a principal maximum. If it is any larger the maxima would be smeared to a greater eﬀective width; if it is any narrower it would restrict the light etendue unneccessarily. The next step is to determine this ideal slit width. Diﬀerentiating eqn. 6.26 with respect to wavelength gives dα = p dλ/(a cos α). (6.35) Now the lateral displacement at the collimator slit plane of dy produces a change in the angle of incidence on the grating dα = dy/f. Using the previous two equations to eliminate dα gives dy = pf dλ/(a cos α). Hence for a resolution ∆λ the slit width must be as small as ∆y = pf ∆λ/(a cos α),
(6.36)
I( θ ) / I(0)
6.9.2
0.6
0.6 0.4 0.2 0 2 1.5 1 0.5 0 0.5 d sin θ / λ
Fig. 6.16 The diﬀraction pattern seen with unblazed (upper panel) and blazed (lower panel) gratings.
144 Diﬀraction
and the slit area is then A = wpf ∆λ/(a cos α).
(6.37)
Using eqns. 6.34 and 6.37 to substitute for Ω and A in eqn. 4.11 gives the etendue of the spectrometer T = [W 2 cos α/f 2 ] [wpf ∆λ/(a cos α)] = W 2 wp ∆λ/(f a).
(6.38)
Using eqn. 6.26 again to replace p, we get T = W 2 (w/f ) [∆λ/λ] (sin θ − sin α).
(6.39)
The luminous ﬂux through the system is simply obtained by multiplying the etendue by the radiance incident on the input slit in the wavelength range (λ,∆λ) F = T I(λ)∆λ = I(λ) [∆λ2 /λ] W 2 (w/f ) (sin θ − sin α).
(6.40)
Once the chromatic resolving power and the incident luminance are chosen, the slit length and grating area should both be increased, and the f/# reduced as far as feasible. Of course the light from the source needs to illuminate the whole slit and its fan of rays should be wide enough to cover the whole grating; otherwise the potential etendue calculated above will not be available.
6.9.3 Entry slit
Grating 2θ Mirrors
Exit slit
Fig. 6.17 Czerny–Turner spectrometer components.
Czerny–Turner spectrometer
A very widely manufactured spectrometer design is based on the Czerny– Turner mounting shown in ﬁgure 6.17. The grating rotates about an axis through its centre and perpendicular to the plane of the diagram. Concave mirrors are used to collimate the light from the entry slit and focus the diﬀracted light onto the exit slit. The use of mirrors avoids any chromatic aberration that could arise with lenses, and the symmetric layout means that coma cancels between the two reﬂections. Folding the optical paths keeps the instrument relatively compact. When the grating is at a given orientation the ﬁrst principal maximum for some wavelength determined by the slit location will pass through the exit slit onto the detector. Therefore the spectrum can be scanned across the detector by rotating the grating. The entry slit is often of ﬁxed width, while the exit slit has a variable width which can be changed to alter the chromatic resolving power. In order to optimize the light entering from the source the source can be imaged with a lens so as to just ﬁll the input slit. The cone of rays from the source should also be made just wide enough to ﬁll the grating. Light passing alongside the grating is not only wasted but also gets reﬂected to give a background haze falling on the output slit.
6.9
In common with other spectrometers the Czerny–Turner can be used as a monochromator. In this role the source illuminating the input would ideally have a ﬂat continuous spectrum (I(λ) is constant). Then the output slit itself becomes a source of narrow bandwidth radiation that can be tuned in wavelength by rotating the grating.
6.9.4
Paraboloid mirror
Blazed grating
Littrow mounting
This simple design for a grating spectrometer uses a single lens, or mirror, as shown in ﬁgure 6.18. The blazed grating is oriented so that the incoming light is incident at the blaze angle and then the reﬂected light returns almost parallel to the incident beam. The mirror is an oﬀaxis section of a paraboloid, and by giving this a slight tilt the reﬂected beam is cast a little below the plane of the diagram so that the detector does not overlap the source. Littrow mountings suﬀer from astigmatism and coma, but are extremely compact. A grating in the Littrow mounting can be used as a monochromator to tune the wavelength of a dye laser, as shown later in ﬁgure 14.10.
6.9.5
Spectrometers and spectroscopes 145
Entry slit
Fig. 6.18 Littrow mounting with a blazed grating with the light incident on the grating perpendicular to the reﬂective surfaces.
Echelle grating
Michelson appreciated that the chromatic resolving power of a grating λ/∆λ = pN is proportional to the order p as well as to the number of lines N . He reasoned that it should therefore be possible to obtain a high chromatic resolving power with a coarse grating by using a high order of diﬀraction. Today the most practical grating whose design is based on this principle is the echelle grating, and this is illustrated in ﬁgure 6.19. It is an extreme form of a blazed grating with deep, coarse steps and is used in a Littrow mounting for which eqn. 6.26 becomes 2a sin φ = pλ,
(6.41)
where φ is the blaze angle. Echelle gratings provide a very useful intermediate level in chromatic resolving power between the Fabry–Perot spectrometer and the grating spectrometer. Thus for example with the R2 grating, which has a blaze angle, φ, such that tan φ = 2, and 316 lines/mm, the diﬀraction order at 500 nm wavelength is around 11 and the chromatic resolving power with a 10 cm width grating is thus 350 000. Such wide line spacing is hard to achieve holographically and echelle gratings must be engine ruled under interferometer control. A two stage spectrometer using a prism and an echelle grating in series, in which the directions of dispersion for the two instruments are at rightangles not only covers a wide range of wavelengths but also achieves good resolution across this range. These instruments are used in
d
φ t a
Fig. 6.19 Echelle grating showing incoming and outgoing beam directions in a Littrow mounting.
146 Diﬀraction
satellite surveys of the Earth and the atmosphere. For some applications it is useful to combine the dispersing elements by forming the grating directly on the surface of the prism; such a device is called a grism.
6.9.6
CCD array
Collimator mirror Input slit
Grating
Fig. 6.20 Spectrometer with focusing grating and CCD detector array.
Automated spectrometers
Simple modern spectrometers contain a ﬁxed grating and a CCD array to capture the whole spectrum simultaneously. Figure 6.20 illustrates the basic elements of such a spectrometer. The CCD array is two dimensional being made wide enough to cover the image of the slit length, with typically 10 µm×10 µm pixels. If a standard gratings were used the image of the input slit would trace out a curved surface as the input wavelength changes. Instead special gratings are constructed in order to keep the whole image focused on the plane surface of the CCD array. These are holographic gratings in which the slit spacing changes uniformly across its length, and this feature gives some supplementary focusing which ﬂattens the image plane. These spectrometers are made to be handheld and interface to a PC with immediate display of the spectrum. Even so the resolution can be as good as 0.2 nm over the visible spectrum.
6.10
Fresnel and Fraunhofer diﬀraction
Figure 6.21 contrasts the conditions for observing Fresnel and Fraunhofer diﬀraction. When, as shown in the upper diagram, the incident and outgoing waves at the aperture have plane wavefronts the phase of the light arriving at P depends linearly on the position at which the Huygens’ wave originated across the slit and it is this feature that makes analysis straightforward. Linear dependence of phase on the position across the slit can be regarded as the distinguishing feature of Fraunhofer diﬀraction. Much more common is the situation shown in the central diagram, where the viewing plane is at a ﬁnite distance from the slit and the diﬀraction is known as Fresnel diﬀraction. P is the point at which the light is observed and is a perpendicular distance r from the plane of the slit, and a transverse distance ρ from a point S on the slit. Then the distance SP is s = r2 + ρ2 ≈ r + ρ2 /2r. (6.42) If it is assumed that plane monochromatic waves are incident on the slit then the relative phase of light arriving at P from S is k [r+ρ2 /2r], which varies quadratically with the position of the point of origin across the slit. This quadratic dependence is characteristic of Fresnel diﬀraction and makes it relatively more complicated to analyse than Fraunhofer diﬀraction. At large enough distances the wavefronts converging at P become suﬃciently ﬂat that the diﬀraction approaches the Fraunhofer limit. At this point the term ρ2 /2r has become small compared to the wavelength. The explicit criterion for Fraunhofer diﬀraction to apply
6.11
Single slit Fresnel diﬀraction 147
for a slit of width w is that (w/2)2 /(2r) ≤ λ/8, Fraunhofer
or equivalently r ≥ w2 /λ = rR ,
(6.43)
where rR is known as the Rayleigh distance. Thus far the incoming waves at the slit were assumed to be plane. If the source is instead at a ﬁnite distance rs then the condition for Fraunhofer diﬀraction to apply is that both r ≥ rR and rs ≥ rR , (6.44) so that the curvature of both incoming and outgoing waves is negligible. If either condition is violated the phase of the light arriving at P depends quadratically on ρ, giving Fresnel diﬀraction. In the case of a circular aperture of radius r rR = πr2 /λ. (6.45) In the lower diagram in ﬁgure 6.21 the observation plane lies at the image plane of the source, and the aperture is located anywhere between source and image plane. According to the deﬁnitions given, the diﬀraction pattern observed with this experimental layout is also, surprisingly, an example of Fraunhofer diﬀraction. In order to understand this conclusion, ﬁrst imagine that a negative lens is placed immediately before the aperture and that it has the correct focal length −f to give a parallel beam. In addition imagine a second, positive lens of focal length f to be placed immediately after the aperture. With this new setup there will be Fraunhofer diﬀraction because the incident and emerging waves are planar at the aperture. The new setup diﬀers from the original in having coincident lenses of equal and opposite powers at the aperture. Thus it is optically equivalent to the original setup: the image will be in in the same place and the same size as before. Hence the original setup in the lower diagram produces Fraunhofer diﬀraction. To summarize: in all cases where observation is made in the image plane of the source Fraunhofer diﬀraction is observed, which includes the standard setup shown in ﬁgure 6.2. For all other arrangements there is Fresnel diﬀraction. In Fresnel/Fraunhofer diﬀraction the phase of light arriving at the observing plane is quadratically/linearly dependent on the distance across the aperture of the point at which the Huygens’ secondary wave originates.
6.11
Single slit Fresnel diﬀraction
The phasor diagram for Fresnel diﬀraction diﬀers markedly from that for Fraunhofer diﬀraction shown in ﬁgure 6.3. In that plot the change in phase angle between phasors contributed by successive elements of the slit was constant because of the linear dependence of phase on position across the slit. By contrast with a quadratic dependence of the phase
P Focal plane
Plane waves incident
Fresnel
S ρ
P r Fraunhofer
S Source
P Image plane Aperture
Fig. 6.21 Examples of Fraunhofer diﬀraction (upper and lower panels), and Fresnel diﬀraction (centre panel).
148 Diﬀraction
angle on position the arc formed by the phasors will curl up, and the result looks as shown in ﬁgure 6.22. The calculation of the intensity distribution for Fresnel diﬀraction at a linear slit shown in the middle panel of ﬁgure 6.21 starts from the expression for the Huygens’ wave arriving at P from S, an element of the slit of width dρ at a distance ρ across the slit dE = exp (iks)dρ/s, where the factor 1/s allows for the falloﬀ in amplitude as the wave spreads out with distance. Thus the total wave at P is
1.5
S(u)
Z2
+0.5
1.0 2.0 0.5
0.5
+0.5
0.5
C(u)
2.0 1.0
0.5
Z1 1.5
Fig. 6.22 Phasor plot for Fresnel diﬀraction at a slit. The resultant amplitude is the chord length between the points with values of u corresponding to the two edges of the slit. The tighter turns of the Cornu spiral, which are omitted here, would converge at the crosses. Points where u = ±0.5, ±1.0, ±1.5... are indicated by ×s.
E=
exp (iks)dρ/s.
Using the approximation from eqn. 6.42 and taking constant factors outside the integral E = [ exp (ikr)/r] exp (ikρ2 /2r)dρ, where the change in the factor 1/r across the slit is ignored. Making a change of variable to u = 2/λr ρ (6.46)
6.11
and dropping the constant multipier gives u2 u2 π π 2 u du + i u2 du, cos sin E(u1 , u2 ) = 2 2 u1 u1
Single slit Fresnel diﬀraction 149
(6.47)
where u1 and u2 are the values of u at the two edges of the slit. This equation contains the Fresnel deﬁnite integrals u π u2 du; C(u) = (6.48) cos 2 0 u π (6.49) S(u) = sin u2 du. 2 0
1.4 Geometric Edge
1.2
The amplitude given by eqn. 6.47 can be rewritten (6.50)
Thus the intensity at P is, apart from constants, I(u1 , u2 ) = [ C(u2 ) − C(u1 ) ]2 + [ S(u2 ) − S(u1 ) ]2 .
Intensity
E(u1 , u2 ) = [ C(u2 ) − C(u1 ) ] + i [ S(u2 ) − S(u1 ) ].
1 0.8 0.6 0.4
(6.51)
These results become more approachable once they are displayed graphically. In ﬁgure 6.22 the trajectory of the function C(u) + iS(u) as u runs from −∞ to ∞ is drawn on an Argand diagram with representative values of u indicated along the path. The total phasor amplitude given by eqn. 6.47 is that part of the loop between the points where u is equal to u2 and u1 . The magnitude of the resultant amplitude at P in ﬁgure 6.21 is then the length of the chord joining the two ends, while the light intensity at P is the square of this chord length. The quadratic dependence of phase on position across the slit has the following eﬀect: the angle which the curve in ﬁgure 6.22 makes with real axis increases quadratically with the parameter u, and hence the phasor curve coils up more and more tightly as u increases. Only the ﬁrst few loops are shown; asymptotically they spiral into the end points Z1 and Z2 in ﬁgure 6.22. The length Z1 Z2 represents the amplitude of the electric ﬁeld at P when the obstructing screen is removed, namely the amplitude of a freely propagating wave. A line from the origin to either end point represents half this amplitude and corresponds to the situation in which one slit jaw is moved oﬀ to inﬁnity so that one side of the incident wave is unobstructed, and the other half of the incident wave is blocked. P would then lie exactly at the edge of the geometric shadow. This gives an intensity at P exactly one quarter what it would be in the freely propagating wave. If P moves laterally into the shadow then one end, u2 , of the phasor remains ﬁxed at the upper end point Z2 while the other, u1 end of the phasor moves away from the origin along the curve toward Z2 . Thus the amplitude (chord length) and intensity (chord length squared) diminish steadily as P moves deeper into the shadow. If P moves in the other direction away from the geometric shadow, the u2 end of the phasor remains at Z2 , while the u1 end now
0.2 0 6
5
4
3
2 u
1
0
1
2
Fig. 6.23 Fresnel diﬀraction at a linear edge. In the upper panel the intensity distribution is shown around the edge of the geometric shadow. The phasor giving the highest intensity is drawn on the Cornu spiral in the lower panel.
150 Diﬀraction
moves away from the origin along the curve toward the lower end point Z1 . In this case the phasor length oscillates and so will the intensity. As P moves steadily away from the shadow edge the u1 end of the phasor moves round loops that grow tighter around Z1 and the intensity oscillates with gradually diminishing swings. The intensity variation near the edge of the geometric shadow is shown in the upper panel of ﬁgure 6.23, normalized to the unobstructed wave intensity. Note that the illumination is highest just outside the geometric shadow, being around 40% brighter than in the unobstructed wave. The phasor producing maximum intensity is drawn onto the Cornu spiral in the lower panel of ﬁgure 6.23. How Fresnel diﬀraction changes as the plane of observation moves away from the slit until it lies at the Rayleigh distance from the slit is illustrated in ﬁgure 6.24. On the left are shown three planes at selected distances from the slit, and on the right the light intensity distributions observed on those planes: the correspondence between each surface and its intensity curve is indicated by a shared line style. On the surface closSlit Intensity ∆ u = 2.0
rR
∆ u = 6.0
∆ u = 1.414
Slit width
Fig. 6.24 Fresnel diﬀraction at a slit. On the left are shown the planes and on the right the patterns seen in those planes. Plane and plot are drawn with matching line styles.
est to the slit indicated by the full line in ﬁgure 6.24 ∆u = u2 − u1 = 6, which is a long section of the Cornu spiral. As P moves across the illuminated region of the surface indicated by the solid line the length of the chord on the Cornu spiral oscillates strongly in length, and the intensity is shown by the full line curve. The pattern still has some resemblance to the geometric shadow. At a surface further oﬀ, indicated by the broken line, the arc length ∆u = 2 is shorter and the chord length changes less violently as P moves across through the illuminated region. The corresponding intensity curve is shown as a broken line.
6.12
Fresnel diﬀraction at screens with circular symmetry
The third, dotted line surface is at the Rayleigh distance from the slit, and here ∆u = 1.414. As the point of observation P moves across the illuminated region the dotted intensity curve is traced out. This has a shape approximating to the Fraunhofer diﬀraction pattern for a single slit. Notice that the minima of intensity are not zero, and will only fall to zero in the Fraunhofer limit.
6.11.1
rs
Lunar occultation
When the Moon’s surface passes across the line joining the observer on Earth to a star the Moon is said to occult the star. This alignment is shown in ﬁgure 6.25. At occultation the star’s image at the Earth’s surface will be diﬀracted. Although the distances from the Earth to Moon, rm , and from Moon to star, rs , are literally astronomical the diﬀraction is Fresnel diﬀraction. This is because the aperture is inﬁnitely wide so that the Rayleigh distance given by eqn. 6.43 is also inﬁnite. Thus the fringe pattern is that shown in the lower panel. From ﬁgure 6.22 we see that the ﬁrst maximum of intensity occurs where u ≈ 1.2, and the second where u ≈ 2.4. Then using eqn. 6.46 the fringe spacing is ∆ρ ≈ 1.2 λrm /2. (6.52) Taking the wavelength to be 500 nm gives a fringe spacing of about 12 m over the Earth’s surface. The fringes travel at the speed of the Moon’s shadow so that the signal produced at a light sensitive detector will be a series of pulses corresponding to the intensity maxima.
6.12
Star
Fresnel diﬀraction at screens with circular symmetry
The prediction for the diﬀraction pattern produced by a circular disk provided a convincing early test of the wave theory of light. In 1818 Fresnel submitted a paper on wave theory in a competition judged by a panel appointed by the French Academy of Sciences. One panel member, Poisson, calculated from Fresnel’s theory that when a circular obstacle is illuminated by a point source there should be a bright spot at the exact centre of the geometric shadow. This prediction appeared to be absurd and easy to refute: Arago, the panel chairman, did the experiment and then saw that the bright spot was really present. This spot became known as Poisson’s spot. The analysis of Fresnel diﬀraction at a circular aperture is presented here in a more qualitative way than that used for a rectangular aperture. A more complete discusion would become complicated and not add any signiﬁcant insights. Figure 6.26 shows a screen illuminated by light from a point source shining through a circular hole. Huygens’ waves originating from an annulus of the wavefront at the hole will arrive at the point P on axis with
Moon rm
Earth Intensity
Fig. 6.25 Lunar occultation.
151
152 Diﬀraction
equal phase delay k(s1 +s2 ) where s1 is the distance from source to annulus and s2 the distance from the annulus to P. With the approximation made in eqn. 6.42 s1 + s2 = r1 + r2 + ρ2 /2r1 + ρ2 /2r2 . Then the phase becomes
aperture s1
k(r1 + r2 ) + kρ2 /2r1 + kρ2 /2r2 ,
screen s2
point source
ρ P
and dropping the piece common to all annuli, k(r1 + r2 ), leaves the relative phase φ = kρ2 (1/r1 + 1/r2 )/2 = kρ2 /2R, (6.53) where R = r1 r2 /(r1 + r2 ).
r1
r2
Fig. 6.26 Fresnel diﬀraction at a circular aperture.
(6.54)
Now imagine that the surface of the incoming wavefront is divided into annular zones such that between adjacent zones the path length to P changes by one half wavelength. Then the Huygens’ waves from adjacent zones will arrive at P with a phase diﬀerence of π, so they tend to cancel one another’s contributions to the amplitude at P. Annular zones drawn in this manner are called Fresnel zones and the outer radius of the mth Fresnel zone is given by kρ2m /2R = mπ, whence ρ2m = mRλ.
(6.55)
The zones therefore have equal areas π(ρ2m − ρ2m−1 ) = Rλ in this approximation. Consequently the cancellation of contributions to the amplitude from succesive Fresnel zones is quite precise. The total amplitude at P is ρ0 E= 2π exp [iρ2 k/2R]ρdρ. 0
Making the substitution ξ = ρ2 k/2R gives ξ0 E = λR exp (iξ) dξ = iλR [ 1 − exp (iξ0 ) ],
(6.56)
0
which indeed oscillates around zero as ρ0 and ξ0 increase. When a more precise calculation is made, starting from Kirchhoﬀ’s expression for the secondary wave in eqn. C.8, the corrections to an individual zone almost cancel. On the one hand the zone area increases slightly with increasing m; on the other hand the amplitude at P falls oﬀ with m, both because the distance from the zone increases and because
6.12
Fresnel diﬀraction at screens with circular symmetry
153
the inclination factor grows smaller. If there are n zones exposed in the aperture the total amplitude can be written as follows E = 0.5E1 + (0.5E1 + E2 + 0.5E3 ) +(0.5E3 + E4 + 0.5E5 )
+ .... + 0.5En−1 + En + .... + 0.5En
(n even) (n odd)
.
Contributions from odd and even zones have opposite sign so that the contributions which are bracketed together cancel, leaving E = 0.5(E1 + En ).
(6.57)
Now imagine that the aperture is made wide enough that the path to P from the whole spherical wavefront emitted by the source is unobstructed. Then the ﬁnal, nth, zone will be in the backward direction and will have an inclination factor zero. This zone therefore makes zero contribution at P, so the previous equation reduces to E = 0.5E1 , showing that if the wavefront is completely unobstructed the wave amplitude on axis is only half as large as when a circular hole exposes just the central Fresnel zone. It is now possible to use this analysis to explain the origin of Poisson’s spot. An aperture exposing exactly n Fresnel zones has as its complement a disk that exactly covers n zones. Suppose that the electric ﬁeld at the centre of the pattern with the disk in place is Edisk . Babinet’s principle requires that the sum of the complementary wave amplitudes equals the amplitude for a freely propagating wave on axis. Thus Edisk + 0.5(E1 + En ) = 0.5E1 . It follows that Edisk = −0.5En , which is nonzero. Therefore there is always a bright spot at the centre of the geometric shadow of a circular disk illuminated by a point source. A clean ball bearing makes an excellent circular disk; but if the outline of the disk used departs from the circular by an area as small as a single Fresnel zone Poisson’s spot is lost.
6.12.1
Zone plates
A zone plate is a ﬂat circular screen which has alternate transparent and opaque annular zones, √ the outer radius of the mth zone counting from the centre is ρm ∝ m. An example is shown in ﬁgure 6.27 with the opaque even zones. If such a zone plate is placed between a point source and a screen, as in ﬁgure 6.26 such that eqn. 6.55 is satisﬁed, ρ2m = mRλ
(6.58)
Fig. 6.27 Zone plate seen from beam direction.
154 Diﬀraction
then the zones on the zone plate precisely match the Fresnel zones. The contributions to the amplitude at P due to light passing through each clear zone are then all in phase. If there are n/2 transparent zones the central intensity is roughly n2 /4 larger than that due to the central zone alone and n2 /2 times larger than the intensity with the unobstructed incident radiation. In eﬀect the zone plate focuses the light from the point source at P. Using eqn. 6.54 to replace R in eqn. 6.58 gives r1 r2 ρ2m /m = λ, r1 + r2 so that 1/r1 + 1/r2 = mλ/ρ2m . If the distances are measured with the Cartesian sign convention this becomes 1/r2 − 1/r1 = mλ/ρ2m , which is identical to the thin lens formula with ρ2m /mλ being the equivalent focal length. Moving oﬀ axis the focusing property is soon lost, and in addition the image obviously suﬀers chromatic aberration.
Mask
Wafer
Point source Aperture stop
Fig. 6.28 Imaging a mask onto a wafer using Koehler illumination. Principal rays are drawn from the edges of the wafer.
While the focusing property of a zone plate is mainly a curiosity for light its use is important in nearby regions of the electromagnetic spectrum for which lenses are diﬃcult or impossible to construct; that is, for Xrays and short wavelength UV light. In addition zone plates are used to focus electron beams used in the commercial fabrication of electronic circuits on silicon wafers. The focal lengths are then about 1 mm and the zones’ widths are about 1 nm.
6.13
Microprocessor lithography
The progress exhibited in Moore’s law, that the number of transistors in processors doubles every 18 months, has been due in large measure to improvements in optical lithography. Each layer of circuitry on a silicon wafer requires the following sequence of operations that make up the lithographic process. First the wafer is coated with a photoresist which when irradiated changes to a form that can be chemically etched away. Then the features required for that layer are transferred from an optical mask (made by optical or electron beam lithography) onto the wafer by a lens system using Koehler illumination as shown in ﬁgure 6.28. The arrowed rays are the principal rays from the two edges of the mask and these are incident normally at the wafer surface. This telecentric arrangement has the advantage that the image will not be displaced laterally by height variations across the chip surface. After exposure, etching and subsequent cleaning the wafer is ready for further processing which may involve deposition of material or removal of material in the regions exposed by the etching. The smallest achievable feature according to the Rayleigh criterion is ∆s = (1.22λ/D)u = 0.61λ/NA,
(6.59)
6.14
where u and D are the image distance and exit pupil diameter respectively. Ever ﬁner features can be produced by reducing the wavelength λ: from 1997 the 248 nm KrF laser has been used and more recently the 193 nm ArF laser. As can be seen in ﬁgure 1.16 the internal transmittance of fused silica falls steeply in this region so that further reduction in wavelength requires a change in approach to less wellstudied materials and to mirrors. Shortening the wavelength also reduces the depth of ﬁeld given by eqn. 4.21. Taking the image to lie in the focal plane, the depth of ﬁeld obtainable with the above resolution is ∆u = 0.305λ/(NA)2 .
(6.60)
At a wavelength of 248 nm and a NA of 0.6 the depth of ﬁeld is only 0.15 µm, which is less than the surface height variation over a wafer. Consequently several exposures may be needed at diﬀerent depth settings to adequately expose the resist across all the wafer. The telecentric illumination ensures that this does not lead to any lateral shift of the image. Several methods are used to circumvent the diﬀraction limit so that features of less than λ/2 are consistently realised, and simple examples of these techniques are now described. One method used is to immerse the region between the ﬁnal lens and wafer in a liquid of refractive index n and refocus to keep u unchanged. The wavelength in the liquid is lower by the factor n than the wavelength in air, and hence the resolution is similarly improved. Figure 6.29 shows an example of the technique known as phase shift masking (PSM). In the upper panel a section of a mask with opaque sections formed by chrome deposited on quartz is shown together with both the amplitude pattern and intensity distribution at the wafer. Through diﬀraction the structure of the mask has been totally obscured. In the lower panel, one clear region of the mask is treated to give a phase shift of π, which might be obtained by coating with molybdenum silicide or by etching away a layer of quartz. Now the structure of the mask is resolved in the intensity distribution on the wafer. An array of sophisticated types of PSM are deployed in practice. The eﬀect of diﬀraction when features drop below one wavelength in size is to round corners, extend tracks and broaden long tracks. Therefore instead of making the mask exactly the same shape as that required on the wafer in the upper panel of ﬁgure 6.30, the mask is shaped instead as shown in lower panel. This is called optical proximity correction. Detailed optical modelling is needed to optimise the shapes used.
6.14
Near ﬁeld diﬀraction
The Huygens–Fresnel addition of secondary waves fails when used to determine the electric ﬁeld amplitude of electromagnetic waves very close to the edges of obstacles. This is because the underlying theoretical justiﬁcation for the Huygens–Fresnel analysis made by Kirchhoﬀ no longer
Near ﬁeld diﬀraction 155
Mask
Mask
π
Phase shifter
Fig. 6.29 Phase shift masking. The mask in the lower panel has a coating over one open region. The amplitudes and intensities at the wafer are shown with broken and full lines respectively.
Fig. 6.30 A mask with optical proximity corrections is shown in the lower panel. This yields a satisfactory approximation to the shape shown above when projected onto the wafer.
156 Diﬀraction
applies. Kirchhoﬀ assumed that the electric ﬁeld across any aperture is exactly what it would be in the absence of any obstacle, and that over the area of the obstacles it is zero. Reﬂections of waves from obstacles are thus neglected although they are important close to the edges of obstacles. Secondly Kirchhoﬀ’s analysis assumes that the electric ﬁeld is a scalar quantity. This is adequate only if the points at which interference is observed are far enough from the aperture that the light from all points across the diﬀracting surface arrives nearly parallel. Then scalar addition yields a good approximation to the total vector ﬁeld amplitude. However the waves from the two edges of a slit to a nearby point travel in very diﬀerent directions. Their electric ﬁeld vectors will not usually be parallel and should not be added as if they were scalars. Near to the diﬀracting surface it is necessary to solve Maxwell’s equations with the appropriate boundary conditions imposed by the edges of the apertures; for example, one requirement is that at the surface of a good conductor the tangential component of the electric ﬁeld vanishes.
6.15
R
Fig. 6.31 Gaussian beam conﬁned in a symmetric confocal Fabry–Perot cavity. The curved broken lines are wavefronts.
Gaussian beams
A Fabry–Perot etalon with a large spacing between mirrors is called a cavity and may have mirrors which are ﬂat or curved. Figure 6.31 shows a longitudinal section through a beam conﬁned in a Fabry–Perot cavity with, in this case, mirrors of radius R spaced a distance R apart. This is called a confocal cavity because the focal points of the mirrors coincide, and symmetric because the radii are equal. The beam boundary does not have straight line edges, rather it has an outline that is hyperbolic with a waist. The standing waveforms that develop have curved wavefronts, shown as dotted lines in the ﬁgure, and their intensity falls oﬀ as the edge of the mirror is approached. These standing waves are examples of solutions of Maxwell’s wave equation subject to the requirement that there is a node of the electric ﬁeld at the mirror surface and that outside the mirror area the intensity vanishes. Light within a laser is usually conﬁned within a Fabry–Perot cavity, and consequently the solutions to Maxwell’s equations for radiation conﬁned in such a cavity have special interest. For the present we can note that in a laser the cavity contains an active material which establishes and maintains the standing waves of electromagnetic radiation in the cavity. Part of this wave escapes through one partially transmitting mirror to form the external laser beam. Figure 6.32 shows a longitudinal slice through such a laser beam after passing through a perfect converging lens. Although there is no aberration the beam does not come to a point focus but has the same characteristic waist as the beam in the cavity. The solution that has been selected for analysis is the simplest of all the possible waveforms that can occur in a Fabry–Perot cavity. This waveform has a radial distribution which falls oﬀ from the optical axis with a Gaussian proﬁle, and has no azimuthal variation. It is
6.15
called the Gaussian or TEM00 mode. Here TEM indicates that electric and magnetic ﬁelds are transverse to the direction of the waves, while the subscript zeroes specify that the proﬁle has no radial or azimuthal nodes. Because of its simple compact shape this is the preferred laser beam shape. Usually the mirror edges or some internal aperture restrict the area of the mode contained in the cavity, lightly clipping the tail of the Gaussian. However the other broader modes spill much more over the mirror edges and are not built up by repeated passes through the active medium.
2 w0
2θ
f Fig. 6.32 Gaussian beam focused by an aberrationfree lens. The curved dotted lines are wavefronts and the broken lines are asymptotic to the beam outline at a large distance along the axis.
The electric ﬁeld distribution in a Gaussian beam of angular frequency ω and wave number k is E = E00 exp [i(ωt − kz)] with3 E00 = (w0 /w) exp (iφ) exp (−r2 /w2 ) exp [ −ik(z + r2 /2R) ],
(6.61)
where w, φ and R are complicated functions of the distance, z, along the beam axis; r is the radial distance oﬀ axis. This and any other exact solution to Maxwell’s equations with boundary conditions has, builtin, the eﬀects of diﬀraction caused by these same boundaries. Thus the radial conﬁnement in the third term on the right hand side in eqn. 6.61 is accompanied by an angular divergence apparent in the ﬁnal term of that equation. Imperfections of lenses or mirrors increase the divergence of the beam, while the aberrationfree waveform of eqn. 6.61 is said to have diﬀraction limited divergence. The evolution of a Gaussian beam as it travels is more complex than that of a beam following ray optics because ray optics ignores diﬀraction. Rather than derive the Gaussian 3 The details of the derivation can be found in Chapter 4 of the fourth edition of Principles of Lasers by O. Svelto, published by Plenum Press, New York (1998).
Gaussian beams
157
158 Diﬀraction
waveform an attempt is made here to interpret the terms in eqn. 6.61 in three steps. First note that the radial distribution of the TEM00 mode is a Gaussian: E00 [radial] = exp (−r2 /w2 ), (6.62) √ where w/ 2 is the root mean square radial width. This wave can be pictured as a plane wave launched at z = 0 with w = w0 . Thereafter the mode undergoes diﬀraction and at a distance along axis much greater than the Rayleigh distance, πw02 /λ the beam boundary asymptotically approaches a cone in shape. The angular spread is then θ = λ/w0 π, R r2/2R
(6.63)
where λ is the wavelength. In detail the radius of the Gaussian beam evolves with distance as follows: w2 = w02 + z 2 θ2
z r
Wavefront
Fig. 6.33 Gaussian beam wavefront.
= w02 [ 1 + (zλ/πw02 )2 ] .
(6.64)
Secondly the spreading waveform tends to a spherical shape with a radius of curvature R. Referring to ﬁgure 6.33 the sagitta at a radius r oﬀ axis is r2 /2R. Thus the phase at the point (z,r,φ) is the same as that at the point (z + r2 /2R, 0, φ). Consequently the wave dependence on distance is contained in eqn. 6.61 in a term E00 [motion] = exp [−ik(z + r2 /2R)].
(6.65)
Collecting the terms in eqn. 6.61 with an explict dependence on r gives exp {−[ikr2 /2][1/R − 2i/(kw2 )]}. 4
Here the imaginary part would have the opposite sign if the choice of complex wave were exp [i(kz − ωt)] rather than exp [i(ωt − kz)].
(6.66)
Evidently the curvature of the wavefront is not exactly 1/R but has acquired this complex form4 1/q = 1/R − 2i/kw2 = 1/R − iλ/πw2 .
(6.67)
It follows that at the plane z = 0, q0 = iπw02 /λ.
(6.68)
If the radius at launch is made very small we would have q = z, and for ﬁnite apertures this becomes q = z + iπw02 /λ.
(6.69)
Thirdly, and ﬁnally, the total energy in the wave must remain constant as it expands. This requires a normalization factor in the Gaussian waveform E00 [normalization] = q/q0 = (w/w0 ) exp (iφ),
(6.70)
6.15
Gaussian beams
159
where φ = tan−1 (zλ/πw02 ) = tan−1 (zθ/w0 ).
(6.71)
Only a single parameter is required in addition to the wavelength to characterize the Gaussian wave fully. The parameter conventionally chosen is the width of the beam w0 at its waist which has been taken to lie at z = 0. The alternative expressions for q, eqns. 6.67 and 6.69, provide a connection between R, z and w. Using this link we have R = z [ 1 + (w0 /zθ)2 ] = z [ 1 + (πw02 /zλ)2 ].
(6.72)
As required the wavefronts become ﬂat at the waist at z = 0. Rearrangement of eqns. 6.64 and 6.72 gives the following useful expressions for w0 and z in terms of R and w w02 = w2 / [ 1 + (πw2 /λR)2 ] 2 2
z = R/ [ 1 + (λR/πw ) ] .
(6.73) (6.74)
These expressions can be used to calculate the location of the waist and its width when the width and curvature of the beam are known at an arbitrary location along the beam.
0 < (1 − L/R1 )(1 − L/R2 ) < 1,
Symmetric confocal
Planar
Cocentral
4 3 2 1 0 1 1  L/R 1
2
3
4
Fig. 6.34 Stability diagram for Fabry– Perot cavities.
(6.75)
where L is the mirrors’ separation and R1,2 are the mirror radii of curvature.5 Otherwise the beam will diverge steadily at each pass. The region of stability is shown shaded in ﬁgure 6.34. Both the symmetric confocal and planar cavities are borderline cases but are simple examples to analyse. In practice some displacement to within the shaded region ensures stability. There are plans to store light at intensities of 100 kW in 4 km long Fabry–Perot cavities in the gravitational wave detectors described in Chapter 8. In the case of a symmetric confocal cavity the waist is midway between the mirrors and the mirrors are located at z = ±R/2. Applying eqn. 6.74 to these mirror surfaces gives R = ±πω02 /λ,
1  L/R 2
In Section 5.9 we saw that monochromatic light within a Fabry–Perot cavity whose wavelength was such that successive reﬂected waves are in phase can remain in the cavity for a number of reﬂections comparable to the ﬁnesse; which can easily reach a few hundred. If in addition, the waveform reproduce its radial and azimuthal distribution after a complete round trip, rather than spreading, conditions are then excellent for storing the beam. The requirement that ensures that a Gaussian beam reproduces itself after each pass to and fro in the cavity is that
4 3 2 1 0 1 2 3 4
(6.76)
5
This condition is derived in exercise 6.13.
160 Diﬀraction
so that the phases at the mirror given by eqn. 6.71 are ±π/4. In order for the waves to duplicate their form after two traversals of the etalon we must have 2kL = 2pπ + π, (6.77) Equation 6.78 was the condition for standing waves found with the simpler analysis of a Fabry–Perot etalon carried through in the previous chapter. In that analysis it was implicitly assumed that the mirrors and the plane waves were of inﬁnite extent laterally.
where p is integral. All the more complex modes beyond the Gaussian TEM00 mode either share this condition or have 2kL = 2pπ.
(6.78)
It is important to note that the modes with a particular value of p are a mix of modes with diﬀerent numbers of longitudinal nodes and diﬀerent transverse distributions. The condition given by eqn. 6.78 is also that for the TEM00 mode in a plane mirror etalon. We see that all the modes of the symmetric confocal cavity have wavenumbers which are integral multiples of π/2L, whatever their angular distribution. This simple distribution in the mode wavenumbers makes the symmetric confocal cavity particularly useful for spectroscopy: if a monochromatic beam is incident having a relatively broad angular range it will be able to couple eﬃciently to a combination of these equal frequency cavity modes. With the other stable cavity conﬁgurations this simplicity is lost. If for example the mirrors of a 10 cm symmetric confocal cavity are moved to be just 11 cm apart, moving inside the zone of stability in ﬁgure 6.34, the cavity modes essentially form a continuum. When therefore nonconfocal cavities are used in spectroscopy a Gaussian beam is required. Scanning across a wavelength range can be made through changing the etalon spacing or by ﬁlling it with gas and altering the gas pressure. Only the latter method is open when using a symmetric confocal cavity. As noted above, lasers are constructed with the active material inside a Fabry–Perot etalon. Generally a simple TEM00 mode is preferred because it is compact and, as we see below, its behaviour in optical systems can be calculated by matrix methods parallel to those used for paraxial ray beams. Consequently the confocal cavity is avoided because the higher order modes, having the identical frequencies, would so easily be excited with the desired Gaussian mode.
6.15.1
Matrix methods
The matrix formulation for tracing paraxial rays through an optical system that appears in Section 3.6 can be extended quite simply so as to apply to Gaussian beams. The matrices deduced in Section 3.6 which describe the action of optical components apply equally to Gaussian beams. A new complex parameter has emerged which is the analogue of the radius of curvature of a wavefront in standard ray optics 1/q = 1/R − iλ/(πw2 ).
(6.79)
6.15
In the case of paraxial rays R = y/α, where y is the distance the ray lies oﬀ axis and α is the ray slope at the same location. Referring back to Section 3.6 an optical element for which the matrix operation is a b , M= c d will produce a new radius of curvature R = (cα + dy)/(aα + by) = (c + dR)/(a + bR).
(6.80)
Correspondingly for a Gaussian wavefront the eﬀect on the complex curvature, q, is q = (c + dq)/(a + bq). (6.81) A simple example of applying this formalism is now given. The Gaussian beam is incident on a lens of focal length f with the lens placed at the beam’s waist of radius r. Then 1/q = −iλ/(πr2 ). The matrix describing the operation of the lens on the beam is 1 −1/f , M= 0 1 so that the wave out of the lens has q = 1/ [ (−iλ/πr2 ) − 1/f ] . Rearranging this result gives 1/q = −iλ/πr2 − 1/f. It follows that the radius of curvature of the emerging wave is −f . In addition the waist radius of the wave leaving is identical to that entering the lens, r. The location of the beam waist beyond the lens is obtained using eqn 6.74 z0 = −f / [ 1 + (λf /πr2 )2 ] which shows the waist is closer to the lens than the ray focus. The beam waist is given by eqn. 6.73, r02 = r2 / [ 1 + (πr2 /λf )2 ]. Then if πr2 λf the waist radius is approximately λf /πr.
Gaussian beams
161
162 Diﬀraction
Exercises (6.1) Check the set of (bulleted) conclusions given in Section 6.5 by drawing phasor diagrams for the case of four identical slits.
focal length 1 m, and θ is ﬁxed at 60◦ . The slit lengths are also 5 cm long. What value should α have in order that the detector intercepts the ﬁrst order maximum for 632 nm? What is the potential chromatic resolving power achievable with the grating, and how narrow should the slits be in order to attain this? What is the etendue of the instrument?
(6.2) A diﬀraction grating 4 cm long has 1000 lines/mm. What is its potential chromatic resolving power? What blaze angle will make 500 nm the blaze wavelength? What is the free spectral range of the grating? (6.11) Show that the etendue of the TEM00 Gaussian mode is λ2 . (6.3) What aperture diameter would a telescope require
(6.4)
(6.5)
(6.6)
(6.7)
to resolve stars that are 1.3 arcsec apart at a wave (6.12) Using the expressions for the radius of the wavelength 500 nm? front and the diameter of a Gaussian beam, show that λR/πw2 = πw02 /zλ. Hence prove eqns. 6.73 What is the frequency of the pulses that are reand 6.74. ceived by a detector on Earth which is set up to observe lunar occultation of a distant star? The (6.13) Using eqns. 3.32 and 3.33 show that the transfer Earth’s radius is 6.38 106 m. matrix for a complete round trip in a cavity with concave mirrors of radius R1 and R2 set a distance What width diﬀraction grating of 200 lines/mm is apart L is required to just resolve the two sodium D lines at 589.592 and 588.995 nm in ﬁrst order diﬀraction? a b If a lens of 0.5 m focal length is used to image the , S= c d fringes what will the physical separation of the lines be? What f/# lens is required? with A grating has two slits each width d and separated a = 1 − 2L/R1 − 4L/R2 + 4L2 /R1 R2 , centretocentre by a. How many maxima will lie under the ﬁrst maximum of the corresponding sinb = −2/R1 + 4L/R1 R2 − 2/R2 , gle slit envelope? c = 2L − 2L2 /R1 , If a grating is immersed in water how does this d = −2L/R1 + 1. change the angular position of the maxima?
(6.8) If alternate slits of a grating are covered with a layer of transparent material that has an optical thickness 0.5 wavelengths longer than that of the same thickness of air what will be the eﬀect on the Fraunhofer diﬀraction pattern?
Then show using eqn. 6.81 that if a Gaussian mode propagating in this cavity reproduces itself after one round trip, that is q = q,
(6.9) All the ten zones of a zone plate are transparent. Alternate zones starting with the second from the centre are covered with the material mentioned in the previous question. If a parallel monochromatic beam is incident normally on the zone plate what is the intensity at the bright central point at focal plane compared to that observed with a conventional zone plate of the same dimensions?
Now use the fact that the q for a Gaussian mode has a nonzero imaginary part to prove that
(6.10) A 5 cm square reﬂection grating with 1200 lines/mm is used as in ﬁgure 6.12 with lenses of
Finally show that eqn. 6.75 follows from these limits.
q = (d − a)/2 ±
[ (d − a)2 /4 + bc ].
(a + d)2 < 4. For this step you need to recall that for any transfer matrix ad − bc = 1.
Fourier optics 7.1
Introduction
The technique of Fourier analysis will ﬁrst be introduced in an elementary way. This relates the time distribution of any wavetrain to its frequency distribution in a precise and simple manner. Equally it provides the relationship between spatial and spatial frequency distributions. Diﬀraction of light at apertures acquires a new interpretation: the spatial distribution of the apertures in coordinate x is being in eﬀect Fourier transformed into the equivalent distribution in spatial frequency, kx = k sin θ, where θ is the outgoing direction of the light. Optical transforms, correlations and power spectra are then discussed. The application of the Michelson interferometer to study the correlations between a light beam at two instants was discussed in Chapter 6. This autocorrelation will be used to give a quantitative measure of coherence. Fourier transforming an autocorrelation gives the frequency distribution; its application to spectroscopy is described here. Parameters for quantifying image resolution are explained. The view of diﬀraction as a Fourier transformation process led Abbe to the modern theory of image formation, and hence to a better understanding of the resolution of optical instruments. Abbe’s theory and its applications will therefore be described here. The use of acoustooptic cells to modulate and deﬂect light, and holography are also described, with applications to information processing.
7.2
Fourier analysis
Pure sinusoidal waves such as cos (kx) are mathematical abstractions because they would extend to inﬁnity. They have simple mathematical properties, for example their integrals and diﬀerentials to all orders are well deﬁned and simple. In the late eighteenth century Fourier showed that any repetitive wave (whether inﬁnite in space or time) can be expanded as a sum of pure sinusoidal waves with numerical coeﬃcients called a Fourier series. This analysis can be extended to wavetrains that are ﬁnite in extent. Suppose the function f (x) repeats its form at equal intervals λ in coordinate x, then Fourier showed that the following expansion is generally
7
164 Fourier optics
valid f (x) = c0 /2+c1 cos kx+c2 cos 2kx+...+s1 sin kx+s2 sin 2kx+..., (7.1) 1
λ/2
cos mkx cos nkx dx −λ/2
where k = 2π/λ. These discrete sinusoids are orthogonal over the range −λ/2 to λ/2,1 so that the Fourier series coeﬃcients are easy to extract, λ/2 cm = (2/λ) f (x) cos (mkx) dx, (7.2) −λ/2 λ/2
λ/2
cos (m + n)kx dx/2
=
sm = (2/λ)
−λ/2 λ/2
cos (m − n)kx dx/2
+ −λ/2
= λ/2 if m = n; 0 if m = n.
λ/2
cos (m − n)kx dx/2 −λ/2
λ/2
−
cos (m + n)kx dx/2 −λ/2
= λ/2 if m = n; 0 if m = n.
f (x) = 1 for x < λ/4; f (x) = 0 for λ/4 ≤ x < λ/2.
λ/2
=
(7.3)
Notice that when f (x) is an even (odd) function of x the sine (cosine) coeﬃcients are all zero. One simple example is a periodic square wave with repeat distance λ, which is deﬁned in the region between −λ/2 and λ/2 by
sin mkx sin nkx dx −λ/2
f (x) sin (mkx) dx. −λ/2
This is a symmetric function of x so that the sm ’s all vanish, while λ/4 cm = (2/λ) cos (mkx) dx = (2/mπ) sin (mπ/2), −λ/4
giving
λ/2
f (x) = 1/2 + (2/π)[ cos kx − cos (3kx)/3 + cos (5kx)/5 − ...]. (7.4)
cos mkx sin nkx dx −λ/2
λ/2
sin (m + n)kx dx/2
= −λ/2
The contributions from the ﬁrst two, four and six terms are shown in ﬁgure 7.1. A more compact way of writing the Fourier series is
λ/2
sin (m − n)kx dx/2 = 0.
−
f (x) =
−λ/2
∞
Fm exp (−imkx),
(7.5)
−∞
with
Fm = (1/λ)
λ/2
f (x) exp (imkx) dx. −λ/2
where F±m = (cm ± ism )/2, and F0 = c0 /2. Waves from actual sources like the pulse shown in ﬁgure 1.6(c) have ﬁnite length and are called nonperiodic. As the following example shows, they are the limiting forms of periodic waves. We rewrite the square repetitive wave as f (x) = 1 for x < λ/4M ; f (x) = 0 for λ/4M ≤ x < λ/2M with M = 1. Now let λ and M increase by the same factor so that λ/M remains ﬁxed at a ﬁnite value d. This leaves unaﬀected the pulse centred on the origin, while increasing the pulse to pulse separation through the increase in λ. If λ and M are increased to inﬁnity while preserving their ratio then what remains is a single square pulse of width d/2 at the origin – which is a nonperiodic wave. At the same time the change
7.2
Fourier analysis 165
in the argument, kx = 2πx/λ, between successive terms in the Fourier series tends to zero. In this limit the sum of terms becomes an integral, where Fm is replaced by F (k)dk, F (k) being a continuous function of k. It follows, after some manipulation, that ∞ f (x) = F (k) exp (−ikx) dk/2π, (7.6) −∞
with
∞
F (k) =
f (x) exp (+ikx) dx.
(7.7)
−∞
where the replacement k = −k has been made in the second integral. From this separation we see that if f (x) is real the two quantities being integrated must be complex conjugates of one another, that is to say ∗
F (−k) = F (k) and F (−k) = F (k).
(7.9)
After putting F (k) = F (k) exp [iα(k)] eqn. 7.8 becomes for real f(x) ∞ F (k) cos [ kx − α(k) ] dk/π. (7.10) f (x) =
F (k) =
where
The combination of eqns. 7.6 and 7.7 require the choice of exp [i(ωt − kx)] for the complex waveform. Equally the combination of equations presented in this note require the complex waveform to be exp [i(kx − iωt)]. 1.5
f (t) exp (−iωt) dt,
(7.12)
−∞
and if f(t) is real f (t) = 0
∞
2 terms used
1 0.5 0 0.5
1
0
1.5
1
x/λ
4 terms used
1 0.5 0 1
0
1.5
x/λ
6 terms used
1
f(x)
1
0.5 0
0.5
1
0
1
x/λ
Fig. 7.1 Partial summations of the Fourier series for a square repetitive wave.
∞
F (ω) =
f (x) exp (−ikx) dx. −∞
0.5
−∞
F (k) exp (+ikx) dk/2π,
−∞ ∞
0
Thus far Fourier analysis has been applied to spatial distributions. It is equally valid and valuable in relating distributions in time and frequency, which are the other pair of conjugate variables appearing in a general wavefunction f (ωt − kx). For these ∞ f (t) = F (ω) exp (+iωt) dω/2π, (7.11)
∞
f (x) =
f(x)
0
It is important to note that the signs of the exponents in eqns. 7.6 and 7.7 could be reversed so that
f(x)
This pair of equations are complementary, and the pair of variables linked, in this case x and k are called conjugate variables. F (k) is known as the Fourier transform of f (x) and taking its Fourier transform, with the sign in front of i reversed, returns f (x). For convenience the relationships will also be written in this way: F (k) = FT(f (x)) and f (x) = FT(F (k)). Rewriting the expression for f (x) ∞ 0 f (x) = F (k) exp (−ikx) dk/2π + F (k ) exp (−ik x) dk /2π 0 −∞ ∞ = F (k) exp (−ikx) dk/2π 0 ∞ F (−k) exp (+ikx) dk/2π, (7.8) +
F (ω) cos [ (ωt + α(ω) ] dω/π,
(7.13)
166 Fourier optics
where F (ω) = F (ω) exp [iα(ω)]. Equations 7.10 and 7.13 show that, as expected, the expansion of measureable, and therefore purely real quantities requires only contributions of positive frequency with real coeﬃcients. Another useful result is that the Fourier transform of a Gaussian distribution is also a Gaussian √ FT[ exp (−x2 /2σ 2 )/ 2πσ ] = exp (−σ 2 k 2 /2). (7.14) The major consequence of the preceding analysis is that it becomes straightforward to calculate the results of any frequency dependent physical process which is caused by a ﬁnite wavetrain, provided that eﬀects add linearly. First the wavetrain is resolved into its pure sinusoidal components with their numerical coeﬁcients. Then the eﬀects of the physical process are calculated for each frequency component. Finally these are summed to give the total eﬀect.
The Dirac delta function The use of this function, δ(x), of a variable x simpliﬁes the Fourier analysis of repetitive pulses. It is deﬁned such that ∞ δ(x − a)f (x) dx = f (a). (7.15) −∞
2
Strictly δ(x) is a functional because it maps an integral onto a number, whereas a function maps a number onto a number.
The function δ(x−a) has unusual properties: it is zero everywhere except for a positive spike at x = a which is inﬁnitely narrow, while at the same time the area under this spike is unity.2 Two useful results follow
∞
−∞
δ(bx) = δ(x)/b b = 0, exp [ ik(x − x ) ] dk/2π = δ(x − x ).
(7.16) (7.17)
Taking the Fourier transform of a function f (x) twice, and using eqn. 7.17, gives f (x ) exp (ik(x − x)) dk dx /2π = f (x), which conﬁrms that the original function is obtained. The reader may now like to prove that FT[f (x − x0 )] = exp(+ikx0 ) FT[f (x)], FT[F (k − k0 )] = exp(−ik0 x) FT[F (k)],
(7.18)
which are called the shift theorems.
7.2.1
Diﬀraction and convolution
If the reader refers back to eqn. 6.3 it is seen that the expression for the amplitude of the diﬀracted wave is simply the Fourier transform
7.2
Fourier analysis 167
Table 7.1 Table of Fourier transforms. f (x) f (ax) δ(x) exp (−x), x >√0 exp (−x2 /2σ2 )/( 2πσ) Π(x): = 1 for x < 0.5, = 0 for x > 0.5
F (k) F (k/a)/a 1 (1 − ik)/(1 + k 2 ) exp (−k 2 σ2 /2) sinc(k/2)
of the wave amplitude at the diﬀracting aperture, in that case for a slit of width d. The conjugate variables are the coordinate across the screen, x, and the lateral component of the wave vector kx = −k sin θ. With a plane wave incident normally on the screen exp (ikx x) is the phase factor required for light travelling in a direction making an angle θ with the forward direction, and this factor constitutes the Fourier transforming factor. Diﬀraction of the same incident beam by a screen whose transmission coeﬃcient is f (x) gives an amplitude in the direction θ which is its Fourier transform ∞ E(kx ) = f (x) exp (ikx x) dx. −∞
For this reason the lens following the screen is called the Fourier lens and its focal plane is called the Fourier transform plane or simply the transform plane. In Section 6.5 the intensity of the diﬀraction pattern for multiple broad slits turned out to be the direct product of the pattern for a single broad slit and the pattern for multiple (inﬁnitely) thin slits. Figure 6.8 illustrates this. This type of simpliﬁcation is very general. The transmission coeﬃcient for multiple identical slits can be described using two functions. One function, h(x), describes an individual slit centred at x = 0: h(x) = 1 for x < a/2 and zero elsewhere. The other function describes the placement of the slit centres at x1 , x2 ,... : g(x) = m δ(x − xm ). The transmission coeﬃcient is then explicitly ∞ f (x) = g(x )h(x − x ) dx , (7.19) −∞
an expression that is called a convolution and which is written in shorthand form as f (x) = g(x) ⊗ h(x) = h(x) ⊗ g(x). (7.20) With the example given of multiple identical slits f (x) = [ δ(x − xm )h(x − x ) ] dx = h(x − xm ). m
m
This convolution is shown pictorially in ﬁgure 7.2. When the functions are more complicated than delta functions the results are less easy to
Σδ 0 Pulse ⇓
0
0
Fig. 7.2 The ﬁgure shows the convolution of a sum of delta functions with a square pulse. The resultant is a set of square pulses centred on the delta function locations.
168 Fourier optics
visualize. The simpliﬁcation in the diﬀraction pattern of an array of identical apertures emerges when the Fourier transform of the convolution is taken: ∞ ∞ F (k) = g(x )h(x − x ) exp (ikx) dx dx . −∞
−∞
Putting y = x − x gives ∞ F (k) = g(x ) exp (ikx ) dx −∞
∞
h(y) exp (iky) dy = G(k) H(k),
−∞
(7.21) where G(k) = FT[g(x)] and H(k) = FT[h(x)]. Alternatively this result can be written FT [ g(x) ⊗ h(x) ] = G(k) H(k), (7.22) which is known as the convolution theorem. The factorization of the diﬀraction pattern of several broad slits into the broad single slit pattern, H(k), and the multiple thin slit pattern, G(k) seen in eqn. 6.13 gives one example of the operation of the convolution theorem. Two other examples are illustrated in ﬁgure 7.3. This ﬁgure shows the Fraunhofer diﬀraction patterns for two, and for ﬁve randomly placed circular apertures. The patterns seen are products of the pattern for a single circular aperture multiplied by the multislit pattern. Optical transfer function
Fig. 7.3 The diﬀraction patterns for two and ﬁve identical circular apertures. The small black circles indicate the hole size and spacing for the two cases.
Any optical system produces a more or less blurred image of a point object due in part to aberrations and in part to diﬀraction. We suppose that the illumination is incoherent, as it would be for a camera viewing a landscape or a telescope viewing a star ﬁeld. The spread of light intensity around the ideal image point is called the point spread function or PSF; and the ratio of the peak intensity to the peak intensity in the limit with no aberrations is called the Strehl ratio. When the Strehl ratio is of order 0.5 or larger it is given approximately by S = 1 − [ 2πσ(λ)/λ ]2
(7.23)
where σ(λ) is the rms wavefront error; that is the departure of the wavefront from an unaberrated shape. Excellent optical systems for which the Strehl ratio is larger than 0.8, corresponding to a wavefront error of λ/14 or less, are conventionally called diﬀraction limited. The PSF determines how easy it will be to detect a point source against a background using an optical system and detector such as a CCD array. The observer would only be able to detect the source provided that the charge deposited in the pixels inside the image signiﬁcantly exceeds the noise in adjacent pixels. The noise includes contributions from background and from the detector itself in the absence of any radiation. If the PSF is reduced by a factor n then the size of the image of the point source is similarly reduced. Its signal is then concentrated
7.3
Coherence and correlations 169
in n2 fewer pixels so that the signal to noise ratio or SNR increases by the same factor. If the PSF is the same over the whole image then the intensity, I, across the image is related to that across the object, O, through a convolution I(x) = O(x) ⊗ PSF(x) = PSF(x − mx )O(x )dx , (7.24) where m is the magniﬁcation. Fourier transforming the PSF resolves the smearing of the image as a function of spatial frequency, k/2π = 1/λ, so that the eﬀects of smearing on small and large scale features in the image are separated. This Fourier transform is called the optical transfer function or OTF. Then from the convolution theorem we have FT[I] = OTF FT[O].
(7.25)
The OTF is in general complex and can be expanded thus:
where MTF is called the modulation transfer function, and PTF the phase transfer function, both being functions of the spatial frequency and both real. If features of some spatial frequency k in the object have visibility (contrast) V (k), this is degraded in the image to MTF(k) ⊗ V (k). One simple method used to measure the MTF is to record an image of test screen. This screen carries parallel lines, spaced λ apart, the intensity varying sinusoidally like cos (2πx/λ). The visibility of the image is then compared with that of the original for diﬀerent choices of the spatial frequency 1/λ. Figure 7.4 shows some representative plots of the MTF: one for a perfectly corrected optical system, labelled C; and others for two imperfect systems, labelled A and B. The MTF for C falls to zero at a spatial frequency determined by diﬀraction at the aperture stop. System A is superior for low spatial frequencies but would be the poorer choice for high spatial frquencies, so that B is better for recording ﬁne detail. This comparison shows that the OTF gives a more complete representation of optical performance than a single number such as the resolving power or Strehl ratio. When coherent illumination is used the amplitudes at the image plane are added, rather than the intensities. The interference eﬀects exploited in microprocessor lithography and described in Section 6.13 are one example. These eﬀects invalidate an analysis based simply on adding intensites.
7.3
1
(7.26)
Coherence and correlations
The coherence of light beams and sources, ﬁrst introduced in Chapter 5, will now be given a quantitative basis in terms of correlations between
MTF
OTF = MTF exp (iPTF),
0.5 A
C
B 0
Spatial frequency: 1/λ
Fig. 7.4 Examples of the dependence of the modulation transfer function on spatial frequency for a well corrected lens (C) and for two imperfect lenses. Of these two B performs better at higher spatial frequencies.
170 Fourier optics
beams or of correlations within a single light beam (autocorrelations). Fourier analysis provides the key by which the spectrum of a source can be extracted from the autocorrelations of the light emitted. The instantaneous intensity produced by a pair of interfering beams of light with complex ﬁelds E1 and E2 is I(t) = E1 (t) + E2 (t)2 = E1 (t)2 + E2 (t)2 + 2Re [ E1∗ (t)E2 (t) ]. Electronic detectors and our eyes measure an instantaneous intensity which is the average over a time T, very much longer that the period of the light oscillations. Thus the measured quantities are I(t) = I1 (t) + I2 (t) + 2Re [ E1∗ (t)E2 (t) ], with the bar indicating a time average. The interference T term contains what is called a correlation of E1 and E2 , namely 0 E1∗ (t)E2 (t)dt/T . Usually beams of any interest have constant intensity so that I = I1 + I2 + 2Re(E1∗ (t)E2 (t)), where I1 and I2 are constants. It will be assumed that the beams involved are stationary, which means that their ﬂuctuations (due to the random emission of wavepackets with random phases) are not changing in character with time. It follows that the correlations deﬁned here do not depend on when they are measured. From here on, unless otherwise speciﬁed, all beams with constant intensity are assumed to be stationary. The size of the correlation term depends not only on how similar the beams are, but also on how intense they are. In order to remove the √ dependence on intensity the correlation is divided through by I1 I2 to give g (1) (1, 2) = E1∗ (t)E2 (t)/ I1 I2 (7.27) which is called the degree of ﬁrst order coherence between the beams. If the interference is between light from the same source at times separated by τ I = 2I0 + 2Re(E ∗ (t)E(t + τ )) (7.28) T ∗ with the inteference term being an autocorrelation 0 E (t)E(t + τ )dt/T . The degree of ﬁrst order coherence is in this case g (1) (τ ) = (E ∗ (t)E(t + τ ))/I0 .
(7.29)
For stationary beams g (1) (1, 2) is independent of time and g (1) (τ ) depends only on the interval, τ , between measurements.
7.3
Bearing in mind that the maximum value of the real part of any complex quantity x is x and its minimum value is −x, it follows that the maximum and minimum intensities are
1
0.5
Imax/min = 2I0 ± 2(E(t)E ∗ (t + τ )).
0 0
Hence the visibilty is
0.5
1
0.5
1
0.5
1
(1)
(τ ),
(7.30)
so that the visibility provides a simple measure of the the modulus of the degree of ﬁrst order coherence. Figure 7.5 shows examples of how the intensity would vary across an interference pattern produced by beams of equal intensity. The xcoordinate could represent either the position across a Young’s two slit pattern or the position of the mobile mirror in Michelson’s interferometer. In the upper diagram there is full coherence and g (1)  is unity, with g (1) having whatever phase shift there is between the wavetrains. Then in the second diagram there is partial coherence with g (1)  being 0.5 and ﬁnally in the third diagram there is incoherence with g (1) being zero.
Power spectra
The power per unit area of a light beam in free space is given by Poynting’s formula eqn. 1.19 N (t) = E(t)2 /Z0 , measured in W m−2 , where E(t) is the real electric ﬁeld. The total energy radiated per unit area is obtained by integrating over time ∞ E= E(t)2 dt/Z0 . (7.31) −∞
The electric ﬁeld at any time can be reexpressed in terms of its Fourier transform e(ω) which is a function of the angular frequency ω. That is to say ∞
E(t) =
e(ω) exp (+iωt)dω/2π.
(7.32)
−∞
Thus the total energy radiated per unit area becomes ∞ e(ω)e∗ (ω ) exp (i[ω − ω ] t) dt dω dω/(4π 2 Z0 ) E= −∞ ∞ = e(ω)e∗ (ω )δ(ω − ω )dω dω/(4π 2 Z0 ) ∞ −∞ = e(ω)e∗ (ω) dω/(2πZ0 ), (7.33) −∞
where the fact that E ∗ = E has been used in writing the ﬁrst line. Comparing eqns. 7.31 and 7.33 we see that the total energy radiated
Intensity
1
V = (E(t)E ∗ (t + τ ))/I0 = g
7.3.1
Coherence and correlations 171
0.5 0 0 1 0.5 0 0
Fig. 7.5 Fringe visibility for two (top) fully coherent, (centre) partially coherent, and (bottom) incoherent sources.
172 Fourier optics
can be expanded in an identical manner in terms of either the temporal distribution, E(t), or of its Fourier transform, the frequency distribution, e(ω). This result is known as Parseval’s theorem and provides the justiﬁcation for treating e(ω)2 /Z0 , measured in W m−2 Hz−1 , as the actual distribution of electromagnetic energy in frequency (rather than angular frequency). It is therefore called the spectral energy distribution. Using the fact that E is real and eqn. 7.9 we have e(−ω) = e(ω), and then eqn. 7.33 becomes ∞ E=2 e(ω)2 dω/(2πZ0 ), (7.34) 0
with, as seems reasonable, only positive frequency components contributing. If the integral is restricted to positive frequencies then the integrand is twice as large. The corresponding energy per unit angular frequency interval is then e(ω)2 /(πZ0 ), while P (ω) = e(ω)2 /(πZ0 T )
(7.35)
is the mean power per unit area per unit angular frequency during the time T that the beam is on. Line width and bandwidth The classical view of atomic and molecular transitions is that the electric ﬁeld emitted by an individual, isolated atom undergoes damped harmonic oscillations E(t) = E(0) exp (−γt/2) cos (ω0 t),
As we shall see in Chapters 12 and 13, quantum theory provides fuller interpretations of both γ and ω.
(7.36)
which is shown in the upper panel of ﬁgure 7.6, and is zero for negative values of t. The intensity decreases with time like exp (−γt) with the mean value of t being 1/γ; this parameter is known as the lifetime of the state radiating. As might be expected, ω0 /2π will emerge as the central frequency of the radiation. The Fourier transform of the distribution is ∞ e(ω) = E(0) exp (−γt/2) cos (ω0 t) exp (−iωt)dt 0 ∞ { exp [ −γt/2 − i(ω + ω0 )t ] = [ E(0)/2 ] 0
+ exp [ −γt/2 + i(ω0 − ω)t ] }dt = [ E(0)/2 ]/[γ/2 + i(ω + ω0 )] + [ E(0)/2 ]/[γ/2 + i(ω − ω0 )]. The ﬁrst term can be dropped in the last line because the factor (ω + ω0 ) in the denominator makes its magnitude negligible at optical frequencies. Then we have e(ω) = [ E(0)/2 ] /[γ/2 + i(ω − ω0 )].
(7.37)
Thus the power spectrum normalized to unit total power is P (ω) = [ γ/2π ]/[γ 2 /4 + (ω − ω0 )2 ],
(7.38)
7.3
Spectral lines emitted by laboratory gas discharge sources are further broadened by collisions of the radiating atoms with other atoms and by Doppler shifts arising from their own motion. When an atom which is emitting a wavetrain collides with another atom there will be an unpredictable change in phase between the wavetrain before and after collision. This broadens the frequency distribution while retaining a Lorentzian shape. In the atmosphere the interval between collisions, τcoll , is around 5 10−9 s which can be converted to a wavelength spread by using the relation given in eqn. 5.24: we get ∆λ = λ2 /cτcoll = 10−3 nm.
√ According to the kinetic theory of gases τcoll ∝ 1/(P T ) where P is the pressure and T here the absolute temperature. When the atom moves with velocity v towards the observer the angular frequency of the radiation emitted shifts from the rest value ω0 to ω0 + ω0 v/c. The kinetic theory predicts the velocity distribution of atoms in a gas to be F (v)dv = exp (−mv 2 /2kB T ) dv, where kB is the Boltzmann constant and m the atomic mass. Then the frequency distribution of the intensity of radiation is, apart from a constant factor, P (ω)dω = exp [ −m(ω − ω0 )2 c2 /(2ω02 kB T ) ] dω, which has the Gaussian shape mentioned above. The constant factor can be chosen so that the integral is unity. We use the result for a Gaussian ∞ √ exp (−x2 /2β 2 ) dx = 2πβ; (7.39) −∞
then the standard form having unit area under the curve is √ P (ω) = exp (−(ω − ω0 )2 /2σω2 )/ 2πσω , (7.40) with a width parameter σω equal to ω02 kB T /mc2 . For gas atoms in the atmosphere the width at 500 nm wavelength is around 1 GHz in frequency or 10−3 nm in wavelength. The angular frequency distribution of the electric ﬁeld is e(ω) = exp [ −(ω − ω0 )2 /4σω2 ].
(7.41)
E(t)
1
0
1 0
P(ω)
which is called a Breit–Wigner or Lorentzian line shape, and this too is shown in ﬁgure 7.6. The power radiated peaks at an angular frequency ω0 and has dropped to half its peak value at angular frequencies ω0 ±γ/2, so that γ is known as the line width. Lifetimes of excited states of isolated atoms undergoing electric dipole transitions are typically around 10−8 s which gives natural line widths around 0.1 GHz. These natural line widths are appropriate to stationary, isolated atoms. Atoms in gases at suﬃciently low temperature and low pressure approximate to this ideal.
Coherence and correlations 173
5
γ
10
FWHM
ω = ω0
Fig. 7.6 Lorentzian spectrum. In the upper diagram the electric ﬁeld distribution in time is shown; in the lower diagram the em energy spectrum. Radiation at optical frequencies would typically have tens of thousands of oscillations under the decay curve.
174 Fourier optics
The corresponding time distribution is the Fourier transform of this ∞ E(t) = exp (+iωt) exp [−(ω − ω0 )2 /(4σω2 )] dω/(2π) −∞ = exp (+iω0 t) exp (+iωt) exp [−ω 2 /(4σω2 )] dω/(2π),
E(t)
1
0
1 5
0
σωt
5
where the ﬁrst of the shift theorems eqn. 7.18 has been used. The remaining integral is given by eqn. 7.14, so that apart from a constant E(t) = exp (−σω2 t2 + iω0 t),
(7.42)
P(ω)
and the actual (real) electric ﬁeld is 2.36 σω
FWHM
ω = ω0
Fig. 7.7 Gaussian spectrum. In the upper diagram the electric ﬁeld distribution in time is shown; in the lower diagram the em energy spectrum.
E(t) = exp (−σω2 t2 ) cos (ω0 t).
(7.43)
This is shown in ﬁgure 7.7 together with the frequency distribution of the electromagnetic energy. Notice that the Lorentzian and Gaussian shapes are quite diﬀerent with the former having a longer tail. With gas sources the Doppler and collision (also called pressure) broadening usually exceed the natural line width while the overall line shape is often close to a Gaussian. The existence of a general relationship between the spread in frequency and in time was broached earlier in Section 5.5.1. Fourier analysis has shown that when the radiation intensity has a frequency distribution of pure Gaussian shape the time distribution of the intensity is also a Gaussian and the widths satisfy the relation σt σω = 1.
(7.44)
This is known as the bandwidth theorem. Now the full width at half maximum (FWHM), ∆t = 8ln(2)σt = 2.35σt , so the above equation when written in terms of the FWHM of frequency and time becomes ∆f ∆t = 0.88.
(7.45)
Measurement errors broaden both the angular frequency and time distributions so that in general the product of the widths is only increased by measurement error. It is also the case that for any distribution other than a Gaussian the product of the width of the distribution and its Fourier transform is always larger than for Gaussians. These points provide the justiﬁcation for writing ∆f ∆t ≈ 1 when the measurement errors are small, and ∆f ∆t ≥ 1 in general. Whenever waveforms have the limiting widths given by eqns. 7.44 or 7.45 they are said to be transform limited. Where the broadening is the same for all atoms or molecules in a source, whether it is due to collisions in a gas or electrostatic forces between nearby atoms in a uniform crystal the broadening is called homogeneous. Where the broadening varies for diﬀerent subsets of the atoms,
7.3
as with Doppler broadening or when a crystal has inhomogeneities, this is called inhomogeneous broadening. The term bandwidth is also used in the analysis of the response of detector systems consisting of a detector of radiation and the electronics for amplifying and frequency ﬁltering the current from the detector. Suppose that h(t) is the current due to unit radiation intensity on the detector system, and that the Fourier transform is H(ω). Then the bandwidth of the system is deﬁned as +∞ B= H(ω)2 dω/2π, (7.46) −∞
where the deﬁnition uses H(ω)2 , rather than H(ω) because it is the power that the detector produces which is relevant. As an example consider the case of a system which has a response time τ , so that it eﬀectively integrates the input over a time τ . For this system we can put h(t) = 1/τ for − τ /2 < t < +τ /2. If the input waveform is cos (ω0 t), then the response as a function of angular frequency is +τ /2 H(ω) = exp (−iωt) cos (ω0 t) dt/τ −τ /2 +τ /2
=
−τ /2
{exp [ i(ω0 − ω)t ] + exp [ −i(ω + ω0 )t ]} dt/(2τ )
= sinc[ (ω0 − ω)τ /2 ]/2,
(7.47)
where, as before, the term with denominator (ω + ω0 ) can be neglected. Therefore the bandwidth expressed in terms of frequency is ∞ B= sinc2 [ (ω − ω0 )τ /2 ] df /2. −∞
Substituting g for (ω − ω0 )τ /2 this becomes3 ∞ B= sinc2 (g) dg/(2πτ ) = 1/(2τ ).
(7.48)
−∞
Following the same steps for a Lorentzian response, which for unit intensity input would give an output intensity exp (−t/τ )/τ , the bandwidth is B = 1/(4τ ). (7.49) 3 Integral 3.821/9 in the 5th edition of Table of Integrals, Series and Products by I. S. Gradshteyn and I. M. ∞Ryzhik,2 edited 2by A. Jeﬀrey, and published by Academic Press New York (1994). {sin (ax)/x } dx = aπ, a > 0. −∞
Coherence and correlations 175
176 Fourier optics
7.3.2
Intensity
A
C
Fourier transform spectrometry
Michelson interferometers are widely employed to measure spectra in the visible and especially the infrared part of the electromagnetic spectrum using a technique called Fourier transform spectrometry. The apparatus is arranged as shown in ﬁgure 5.11 to study a source’s spectrum. Mirrors M1 and M2 are set parallel in order to give circular fringes and a detector is placed behind a small circular hole centred on the focused fringes. The intensity seen by the detector is recorded continuously during a scan in which the mobile mirror is moved at constant speed. This recorded intensity pattern is called an interferogram, and we shall see that taking its Fourier transform yields the power spectrum of the source.
B
Path difference
Fig. 7.8 Schematic of interferogram for the sodium D lines.
A simple example of an interferogram which illustrates how information on spectra can be extracted is shown schematically in ﬁgure 7.8. It is an interferogram recorded with the D lines of a sodium gas source at 588.995 nm and 589.592 nm, all other spectral lines being ﬁltered out. Eﬀectively there are two incoherent sources each giving a separate intensity fringe pattern at the detector. These two intensities add to give the total intensity detected and recorded as the interferogram. At points of maximum visibility (A,C) the fringe patterns are in step: their peaks in intensity coincide and their troughs in intensity coincide. Correspondingly at a minimum of visibility (B) they are exactly out of step. Between successive maxima of visibility the path diﬀerence must change by a certain number of wavelengths, ξ, for the longer wavelength and by ξ + 1 wavelengths for the shorter wavelength. ξ is determined to the nearest integer simply by counting the number of waves under the envelope between A and C. Knowing this value (ξ + 1)λshort = ξλlong , so ∆λ = λshort /ξ, which provides a measurement of ∆λ if λshort is already known. The chromatic resolving power obtained with the Michelson interferometer in this way depends on the total change in path diﬀerence during the scan. It will be possible to just resolve two wavelengths λ and λ − ∆λ provided that two adjacent maxima of intensity, A and C, lie within the scan length (full range of the path diﬀerence), which we take to be xw = nλ. Then at the limit of resolution (n + 1)(λ − ∆λ) = nλ. Rearranging this equation gives an estimate of the chromatic resolving power λ/∆λ = n = xw /λ. (7.50) The next step in the discussion will be to consider the analysis of more typical spectra consisting of many lines, and for which Fourier analysis is essential.
7.3
The intensity at the detector, assuming that the beams from the mirrors have equal intensity, is given by eqn. 7.28 I = 2I0 + 2(E(t)E(t + τ )) where E(t) and E(t+ τ ) are the values of the real electric ﬁeld at times t, and (t + τ ) respectively during the scan. Only the autocorrelation term is of interest here. Writing the electric ﬁeld in terms of its frequency components ∞ E(t) = e(ω) exp (+iωt) dω/2π, −∞
the autocorrelation term is E(t)E(t + τ ) = e(ω)e(ω ) exp (+iωt) exp [+iω (t + τ )] dω dω dt/(4π 2 T ), where T is the duration of the measurement. Using the equality exp [ +i(ω + ω )t ] dt = δ(ω + ω ), and recalling that because E(t) is real e(−ω) = e∗ (ω), the autocorrelation simpliﬁes thus ∞ E(t)E(t + τ ) = e(ω)e∗ (ω) exp (−iωτ ) dω//(2πT ) −∞ ∞ = e(ω)e∗ (ω) exp (−iωτ ) dω/(πT ). (7.51) 0
We have seen in eqn. 7.35 that e(ω)2 /(πZ0 T ) is simply the power spectrum P (ω), so that taking the Fourier transform of eqn. 7.51 gives P (ω) = FT{E(t)E(t + τ )}/Z0 .
(7.52)
This relation between the power spectrum and the Fourier transform of the autocorrelation provides the required link betwen the interferogram and the power spectrum. Equation 7.52 is one version of the Wiener– Khinchine theorem. This result can be reexpressed in terms of the wave number k = ω/c and the path diﬀerence x P (k) = P (ω)(dω/dk) = (c/Z0 )FT{E(t)E(t + x/c)} .
(7.53)
Figure 7.9 shows a reproduction of a typical spectrum and interferogram obtained in an undergraduate experiment using a mercury lamp.4 4 If the number of samplings is N , an N × N matrix inversion is needed to extract the Fourier transform of the interferogram. When, as is typically the case, N is of order 10 000 this process is very time consuming. J. W. Cooley and J. W. Tukey, Math. Computing 19, p.297 (1965), invented a fast Fourier transform technique that reduces the number of inversions to N ln(N ). However the spectrum obtained is discrete, being determined at the wavelengths xw , xw /2, xw /4, ... , 2xw /N only.
Coherence and correlations 177
178 Fourier optics
Intensity
1 546nm
436nm
0.5 404nm 0 10
12 14 16 18 Wavenumber in µ m 1
20
5 10 Scan coordinate in µ m
15
Detector current
4
2
0 0
Fig. 7.9 In the upper panel the spectrum of a mercury lamp with broad lines at 404, 436 and 546 nm is shown, and in the lower panel the interferogram. We shall see later that the wave number is directly proportional to the energy change in a molecular/atomic transition, which enhances the usefulness of this display.
There are limitations on the information about a spectrum that can be extracted from an Michelson interferogram. Firstly the scan length, xw , restricts the interferogram to a ﬁnite window, and secondly the detector only samples the intensity at discrete steps, xs apart, and not continuously. Figure 7.10 helps to illustrate this, showing in the upper panel that no useful information can be extraced for radiation of wavelength longer than the scan length xw . The lower panels show that sampling a wavetrain of wavelength 2xs /3 can give an identical set of measurements to sampling one of wavelength 2xs . This eﬀect is known as aliasing. It can be inferred that measurements every xs along the interferogram only sample waves of wavelengths greater than 2xs adequately. This inference is given precision by the Whittaker–Shannon sampling theorem: it states that sampling at intervals xs over an inﬁnite scan length would uniquely determine a spectrum containing only waves with wavelength longer than 2xs . These limitations can be quantiﬁed when the Fourier transform is examined. What is being measured is not I(x) = E(t)E(t + x/c) but the product I = I C W = I(x) δ(x − mxs ) W (x) m
where C is a comb function consisting of a sum of delta functions at intervals xs , and W is a window function, a square pulse of height unity and width xw . According to the convolution theorem the Fourier transform of I is the convolution P (k) = FT(I) ⊗ FT(C) ⊗ FT(W ).
(7.54)
The Fourier transform of the square window broadens any line at k0 in the following way
xw /2
FT[W ] = −xw /2
exp (ikx) dx/xw = sinc(kxw /2),
(7.55)
and if the spectrum is monochromatic with k = k0 , then FT(I) = δ(k − k0 ) and (7.56) FT[I] ⊗ FT[W ] = sinc[ (k − k0 )xw /2 ]. Two spectral lines will be regarded as just resolved if the maximum of the sinc function of one falls at the ﬁrst minimum of the sinc function of the other line. This happens when the separation in wave number, ∆k, between the lines is such that ∆k = 2π/xw .
(7.57)
Then the chromatic resolving power is λ/∆λ = k/∆k = xw /λ,
(7.58)
7.3
Coherence and correlations 179
which reﬁnes the estimate given in eqn. 7.50. xw
By contrast the eﬀect of the comb is to give aliases of any spectral line. The Poisson summation theorem gives δ(x − mxs ) = (1/xs ) exp (−iks nx), (7.59) m
n
xs
where both summations run from −∞ to +∞ and ks = 2π/xs . This equality is a good approximation in the present case when the number of samples is several hundred. Then FT[C] = (1/xs ) exp [ i(k − nks )x ] dx n δ(k − nks ). = (1/xs )
(7.60)
n
Suppose again that the spectrum is monochromatic so that FT[I] = δ(k − k0 ). It follows that δ(k − k0 − nks ). (7.61) FT[I] ⊗ FT[C] = (1/xs ) n
These aliases are displaced in wave number from the actual line by integral multiples of 2π/xs . Aliases from radiation of wavelengths below 2xs which could simulate spectral lines above that wavelength must be eliminated by inserting an optical ﬁlter which removes radiation of wavelengths shorter than 2xs . The Michelson spectrometer has two important advantages over the grating spectrometer which make it the preferred instrument in many situations. The ﬁrst, Jacquinot advantage, is that the etendue is about one hundred times larger than for a comparable5 grating spectrometer. This means that weaker sources can be studied and also weaker lines identiﬁed. The second, Felgett advantage, is that the Michelson detector receives all the wavelengths throughout the scan, whereas the detector in a grating spectrometer receives a restricted wavelength range. With N samples the time required to examine the same spectrum with a grating spectrometer in the same detail as with a Michelson is longer by a factor N , which is usually 1000 or more. When studying infrared spectra the radiation from any part of the apparatus falling on the detector can give a background illumination that overwhelms the signal of interest in the case of grating and prism spectrometers. At room temperatures the spectrum of black body radiation peaks at a wavelength around 10 µm. The superior etendue of the Michelson is then a prime advantage. One simple technique used to remove this background is to pulse the radiation studied by placing a shutter in front of the source. Then the pulsed part of the radiation received on the detector is selected electronically using a lockin ampliﬁer.
Fig. 7.10 In the upper panel the scan length, xw , is much shorter than the wavelength. In the lower panels aliasing is shown between sine waves of wavelength 2xs and and 2xs /3 for a sampling interval xs . Samplings are indicated by dots.
5
See exercise 6.10 and the example immediately below
180 Fourier optics
In addition, because the Michelson spectrometer measures correlations it is correspondingly less sensitive to a background which is constant with time.
f
f θhole
D
θhole
Fig. 7.11 aperture.
Michelson
spectrometer
Example 7.1 The etendue and chromatic resolving power are the parameters most useful when comparing diﬀerent types of spectrometers. In order to determine the etendue we need to consider the area and angular spread of the beam falling on the detector. One essential restriction is that the circular aperture deﬁning the area of the detector exposed should contain just the central fringe and no more. If there is a maximum of intensity at the exact centre of the fringe pattern for wavelength λ when the optical path diﬀerence between the two arms is x, then x = nλ. The adjacent minimum is at an angle θ given by (n − 1/2)λ = x cos θ = x(1 − θ 2 /2 + ...). Taking the diﬀerence of the last two equations gives, to an approximation adequate for the small angles involved, λ = xθ2 . Referring to ﬁgure 7.11 the angular radius θhole of the aperture in front of the detector is made small enough to accept only the central fringe 2 θhole = λ/x.
Next let the focal length of the exit lens be f so that the area of the hole deﬁning the area of the detector exposed is 2 = πf 2 λ/x. A = πf 2 θhole
The other parameter deﬁning the etendue is Ω, the solid angle subtended at the hole by the exit lens. If D is the lens diameter Ω = πD2 /4f 2 . Using eqn. 4.11 the etendue is T = A Ω = π 2 D2 λ/4x.
(7.62)
A typical instrument might have a scan length of 2 cm and a lens of 3 cm diameter. At 1 µm wavelength, for such an instrument, eqn. 7.58 gives a chromatic resolving power of 2 104 . The etendue would be 0.11 mm2 sr, which is very much larger than the etendue of a grating spectrometer.
7.4
7.4
Image formation and spatial transforms 181
Image formation and spatial transforms
The connection between diﬀraction and the resolution of detail seen in optical images was discovered in the mid 19th century by Abbe whilst he was trying to understand why larger but less wellcorrected lenses used in microscopy gave more detail than smaller, better corrected lenses. Koehler illumination, illustrated in ﬁgure 4.10, is essentially coherent. Figure 7.12 shows a grating, AB, illuminated by monochromatic plane waves and imaged by a wellcorrected lens. For simplicity the object
Plane waves incident on grating
Lens
A
’
B
A
’
B Transform plane
2f
f
Image plane
f
Fig. 7.12 Abbe’s insight into image formation. The lower order diﬀractive beams from the grating AB are shown as shaded bands. A B is the grating image. Rays travelling from A to A are drawn as broken lines; rays from B to B as solid lines.
distance is taken to be twice the focal length, giving unit magniﬁcation. Parallel beams are shown emerging from the grating in the directions of the principal maxima, and these produce bright lines in the transform plane, that is to say in the focal plane of the lens. Apart from these maxima the transform plane is dark. As they travel onward from the transform plane the beams spread out again to form the image of the grating, A B , at twice the focal length from the lens. Abbe pictured image formation as a two step process: in the ﬁrst step the grating diﬀracts the incident light into parallel beams which form the Fourier transform of the grating; in the second step the lens performs a second Fourier transform to recover the image of the grating. Abbe pointed out that even with a perfect lens the image cannot be exact because the lens only accepts the lower order (small angle) diﬀracted beams. If the lens is so small that it only captures the zeroth order diﬀracted beam then the image plane is of uniform brightness and all detail is lost. If instead the ﬁrst and zeroth orders are captured the image is a grating of the correct spacing but with a sinusoidally varying
182 Fourier optics
intensity across each element of the grating. We can see this by taking a distribution in the transverse component of the wave vector kt which only contains these orders Aδ(kt ) + B[ δ(kt − k0 ) + δ(kt + k0 ) ]. Fourier transforming this expression gives a distribution in the lateral coordinate, x, exp (−ikt x){Aδ(kt ) + B[ δ(kt − k0 ) + δ(kt + k0 ) ]} dkt = A + B[ exp (−ik0 x) + exp (+ik0 x) ] = A + 2B cos (k0 x). The ﬁnest resolvable detail on an object will be detail for which the lens has just big enough diameter to capture the ﬁrst order diﬀracted beam. If this limiting detail has linear dimension ∆x then k0 ∆x = k∆x sin θ = 2π, where θ is the semiangle subtended by the lens at the object. Then ∆x sin θ = λ.
(7.63)
This, Abbe’s view of what is resolvable by an optical system, provides a complementary approach to the Rayleigh criterion given in Section 6.7.
f
f
Fig. 7.13 Spatial ﬁlter consisting of a pinhole and lenses. Only the plane wave component from the incoming distorted wave focuses at, and passes through the pinhole.
Notice that the nth diﬀraction order for a grating of a given pitch coincides in angle with the ﬁrst order for a grating with an ntimes ﬁner pitch, so we can say that higher order beams carry information about the ﬁne detail of the grating. Put another way: higher diﬀraction orders correspond to features with higher spatial frequencies across the image and are needed to recover the sharp edges of the slits in the image. Of course most objects viewed with a microscope are more complex than gratings, but the conclusion remains valid that information on the ﬁne detail in the image is carried by light diﬀracted at large angles. Although this wide angle light may be very weak it is essential for resolving ﬁne detail. The fact that it makes a signiﬁcant contribution despite its faintness was what had escaped notice before Abbe’s investigations. The diﬀractive orders are directly accessible in the transform plane and it is here that manipulations are most easily made to enhance the properties of images formed with a coherently illuminated object.
7.5
Spatial ﬁltering
Any changes made to the high or low order spatial frequency components of an image by selectively obscuring regions of the transform plane is known as spatial ﬁlter. Blocking the high order spatial frequencies will
7.5
smooth the image. This will for example remove noise in the form of dust spots on a picture or the dot structure in pictures produced by halftone printing. Conversely, removing low order spatial frequencies will enhance the outlines of images. Similar image editing facilities are provided by packages for manipulating graphics on PCs. The simple ﬁltering shown in ﬁgure 7.13 produces a pure plane wavefront from a distorted wavefront, such as that from a diode laser. The incoming beam is focused on a pinhole located on axis in the focal plane of the ﬁrst lens. Only the plane wave component in the incident beam passes through the pinhole; all the other components in the beam focus away from the pinhole. The waves emerging from the pinhole are spherical and the second lens converts these to a plane wave. Biological specimens studied with microscopes are mostly transparent and the structures of interest may only diﬀer from the surrounding material in having a slightly diﬀerent refractive index. One drastic approach is to dye the specimen with a material that is selectively absorbed by some structures; but this is not always feasible and it may damage the specimen. An alternative is to use dark ﬁeld illumination which is shown in the lower diagram in ﬁgure 7.14. The left hand diaphragm forming the aperture stop in the condenser is perfectly opaque apart from a narrow clear annular aperture. In the plane conjugate to this diaphragm on the objective side an annular stop is placed which exactly covers the image of the annular aperture. This plane is the transform plane and the annular stop introduced blocks the zero order light. An observer viewing a blank slide in the ﬁeld of view would see a perfectly dark ﬁeld. If a specimen slide is then inserted the higher spatial frequencies in its image, which are normally swamped by the zero order beam, can now produce an image. In this image the edges of structures are particularly clear. A more subtle technique earned a Nobel prize for Zernike in 1953 and this is illustrated in the upper diagram of ﬁgure 7.14. The annular condenser aperture just described is again used, but now a transparent phase plate replaces the annular stop in the objective. This phase plate is made thicker over the same annular region which was opaque in dark ﬁeld illumination to give a phase delay of π/2 for light travelling through the annulus relative to light missing it. Suppose the amplitude of light at the transform plane is E0 sin (ωt) with normal illumination and a clear slide. When a specimen is inserted the light passing through any given area undergoes a small relative phase shift , determined by how much the local refractive index deviates from the mean refractive index. The corresponding amplitude is E0 sin (ωt + ) = E0 sin (ωt) + E0 cos (ωt), and the intensity is E02 (1 + 2 ) ≈ E02 , showing that the weak image has been swamped. Things are quite diﬀerent when the phase plate is inserted because the zero order component is phase shifted to E0 cos (ωt).
Spatial ﬁltering 183
Condenser
Objective
Phase plate
(a)
Annular aperture
Specimins
Annular blocking
(b)
f
f
Fig. 7.14 The upper diagram shows phase contrast illumination, and the lower diagram shows dark ﬁeld illumination.
184 Fourier optics
Thus the total amplitude becomes E0 (1 + ) cos (ωt) and the intensity becomes E02 (1 + )2 ≈ E02 (1 + 2). This gives a detectably large change in intensity for even a small change in refractive index. A
7.5.1
A Object plane Knife ’ edge S S
Slit source
S
0 1 2
S
Image plane A’
’
A’
Fig. 7.15 Schlieren photography. The diﬀraction orders at the plane of the knife edge are numbered 0, 1 and 2.
This technique is used to make visible the variations in density in a ﬂuid in motion around large scale objects, and the apparatus used is sketched in ﬁgure 7.15. A monochromatic slit source is focused by a mirror so as to throw a parallel beam through the volume of interest, which is shown shaded in the ﬁgure: this might, for example, be a wind tunnel. Afterwards the light falls on a second mirror whose focal plane (the transform plane) is S S and which images the object AA at A A . A knife edge is placed at S S parallel to the slit source and positioned to block oﬀ part of the zero order as well as all the higher diﬀraction orders on one side. Regions of refractive index diﬀering from the mean then appear as bright and dark streaks, just as in the phase contrast microscope. Mirrors are preferred to lenses for the focusing because the area of the ﬁeld of view is comparatively large. The mirrors shown are oﬀaxis paraboloids whose aberrations remain small over the required angular range.
7.5.2
Intensity
4
2
0 0
Schlieren photography
Mirror
Mirror
Location of target
0.5 Radial distance from star
1
Fig. 7.16 The image intensity of a bright point source produced by a Gaussian shaded aperture (broken line) and that for a sharp aperture boundary (solid line).
Apodization
Attempts to observe faint objects near brighter ones are made diﬃcult because the diﬀraction rings around the image of the brighter object can easily swamp the fainter one’s image. The rings outside the central Airy disk can be suppressed by placing a ﬁlter over the telescope objective whose transmission coeﬃcient falls with a Gaussian dependence on the radial distance oﬀ axis. The diﬀraction pattern is the Fourier transform of this shape, which is also a Gaussian. Figure 7.16 contrasts the diﬀraction patterns for a clear circular aperture and a Gaussian shaded aperture. While the Gaussian is broader its lack of any outer rings means that there is a much improved chance of detecting a target against the glare of a brighter companion. Apodization is the term used to describe this procedure.
7.6
Acoustooptic Bragg gratings
Ultrasound waves in a solid or liquid produce sinusoidal variations in density and hence similar changes in refractive index across the material. Incident light will diﬀract from such density gratings and at the same time undergo a minute Doppler shift in frequency because the grating is travelling at the speed of sound. Both the deﬂection and frequency shift have signiﬁcant applications in information processing. The techniques, their analysis and some applications will be reviewed here.
7.6
The crystals used to excite the ultrasound waves such as quartz are chosen for their high piezoelectric coeﬃcients at ultrasound frequencies of up to many gigahertz. Figure 7.17 shows an incident plane wave light beam being diﬀracted from a Bragg cell. The Bragg cell consists of a piezoelectric crystal which generates ultrasound waves, an acoustooptic material through which they travel, and ﬁnally an absorber that suppresses reﬂections. Choices for the acoustooptic material are extradense ﬂint glass and lithium nibate. If the acoustic wave has suﬃcient width W the incident light will be diﬀracted entirely into the zero order and ﬁrst order on one side. This can be understood with the help of the lower panel in ﬁgure 7.17. The directions of the waves being diﬀracted in ﬁrst order from planes of maxima in refractive index are shown, θ being the angle which the incident light beam makes with the acoustic wavefronts inside the cell. Diﬀraction maxima will satisfy the usual relation pλ = Λ(sin θ + sin α ),
Acoustooptic Bragg gratings
Piezoelectric transducer
θ θ
w
Absorber
θ’
θ’
Λ θ’
where θ and α deﬁne the incoming and outgoing wave directions, where Λ and λ are the respective wavelengths of acoustic and of light waves in the crystal, and where p is integral. There is an additional requirement for observing a maximum when the acoustic wave is broad: light diﬀracted at all points along a given acoustic wavefront must be in phase, which means that the angle of reﬂection (α ) equals the angle of incidence (θ ). Then the requirement for a diﬀraction maximum becomes
185
θ’
θ’ θ’
Fig. 7.17 The upper diagram shows an acoustooptic Bragg cell diﬀracting a laser beam. The lower diagram shows the scattering of light from a pair of acoustic waves inside the Bragg cell.
2Λ sin θ = pλ . Expressing this in terms of the angle at which the light waves are incident in the air and their wavelength in air this becomes 2Λ sin θ = pλ. Lastly the angle of incidence is made suﬃciently small that the radiation is restricted to zero, or ﬁrst order with 2Λ sin θB = λ,
(7.64)
where θB is given the subscript B because this equation is also the Bragg condition met in Xray diﬀraction. In order for any diﬀraction to occur the width of the optical beam, w, should span at least one acoustic wavelength: Λ < w/ cos θ ≈ w because θ is small. Now Λ = V /F,
(7.65)
where V and F are respectively the velocity and frequency of the acoustic waves. Therefore the requirement on Λ can be reexpreesed as a frequency limit which acoustic waves must exceed in order to produce any useful acoustooptic eﬀect F ≥ V /w.
(7.66)
If the width W of the acoustic wave is suﬃciently narrow then the angular separation between the diﬀractive orders, λ /Λ, becomes smaller than the angular spread of the acoustic beam Λ/W . As a result diﬀraction of the optical beam in many orders can occur. This is called the Raman–Nath regime, but is not of any further interest to us here.
186 Fourier optics
In ﬂint glass with a light beam of 1 mm width F ≥ 3 106 Hz. The light scattered from the acoustic wave is Doppler shifted from frequency f0 to frequency f0 + F . It is thus possible to make use of either the deﬂection or the frequency shift of light within a Bragg cell. These changes are determined by the acoustic wave frequency, which in turn is the frequency at which the piezoelectric crystal is driven. The frequency shift is used in applications requiring heterodyning of optical beams, while beam deﬂection is used in information processing. Bragg cell
Laser beam
7.6.1
Fourier lens
CCD array
Fig. 7.18 Bragg cell diﬀracting a laser beam. Diﬀerent microwave frequencies cause diﬀerent beam angular deﬂections.
The requirement on the relative angular spreads of the two beams is very different when the acoustic beam is used simply to Doppler shift the frequency of the optical beam. Then the need is to maximize the power transfer into the laser beam and so the angular spreads of the optical and acoustic beams are made equal.
Microwave spectrum analysis
Figure 7.18 shows the elements of a system for analysing microwave spectra, which is used to scan continuously for incoming radar signals. A laser beam is incident on the Bragg cell and the incoming microwave signal is applied to the piezoelectric transducer. Each microwave frequency produces a speciﬁc laser beam deﬂection and the deﬂected beams are projected by the Fourier lens onto a linear detector array such as a CCD. Using eqns. 7.64 and 7.65 we can relate the accessible range of microwave frequencies to the angular spread of the microwave beam ∆F = V ∆(1/Λ) = 2V cos θB ∆θB /λ ≈ 2V ∆θB /λ.
(7.67)
This is maximized when ∆θB ﬁlls the angular divergence of the acoustic beam. The number of distinguishable frequencies across this bandwidth Nr is twice the angular divergence ∆θB divided by the angular divergence of the optical beam, because the deﬂection of the optical beam is twice the Bragg angle. Thus Nr = 2∆θB /∆θlaser = 2∆θB w/λ.
(7.68)
Bandwidths of several gigahertz are obtained and of order a thousand distinguishable frequency steps. Of course the lens focal length, the CCD detector array pixel size and number of pixels need to be matched to this performance.
7.7
Holography
Holograms are familiar from the embossed images on credit cards and displays in various promotions. The ﬁrst holograms were made in 1948 by Gabor, who received a Nobel prize in Physics for this work. However it needed the introduction of the laser to make the technique of practical use in the optical domain. Prototype systems are manufactured which store data in holograms with fast access and readout. The principles and several applications of holography will be described in the following sections. The images seen through holograms have depth and when the observer’s viewpoint is changed the scene changes exactly as if the original
7.7
Laser beam in Film
Mirror
Reference beam ct
je Ob
am
be
Object
Hologram Reference beam
Real Virtual
object
object Fig. 7.19 The upper diagram shows the exposure to produce a hologram. The lower diagram shows the images produced when the hologram is illuminated with the reference beam again. The part of this reference beam that is not scattered by the hologram is omitted from the lower diagram.
scene were present. What is special about a hologram is that, unlike a normal photograph which records the light intensity falling on it, the hologram records the amplitude and phase of the light from the scene. Coherent illumination is necessary, and can readily be provided by a laser. The exposure of a hologram and the later image reconstruction are shown in ﬁgure 7.19. During the exposure the light from a laser is divided into two parts that remain coherent with one another. One beam, called the reference beam, falls directly on the ﬁlm that will become the hologram, while the other beam illuminates the object to be recorded. Light scattered from this object forms an object beam which also falls on the ﬁlm. The fringes on the ﬁlm produced by interference between the object beam and the reference beam contain phase information from the object beam, and this pattern is stored as exposed and unexposed
Holography 187
188 Fourier optics
regions on the ﬁlm. Processing preserves these fringes as dark and light areas.
Film
Afterwards the hologram is replaced in its initial position with respect to the reference beam, but now the object is removed. The observer sees a realistic three dimensional virtual image of the object exactly where the object would have been. In addition there is a second real image which has its surface features inverted, so that, for example, a nose would appear to project into rather than out of a face. This is known as a pseudoscopic image. The observer sees the virtual image through a window formed by the hologram; and if the hologram is broken in two, either piece alone gives the same view – through a window that is now smaller. This property shows that the information about all the image is held over all the hologram; which is very diﬀerent from how information is held on a normal photograph.
7.7.1
R
Fig. 7.20 The diagram shows a point object in a reference beam. The reference beam and the scattered light from a point object produce circular interference fringes on the ﬁlm.
Principles of holography
An elementary example of holography is shown in ﬁgure 7.20 where a point object is placed in an incoming plane laser beam at a distance R from the recording ﬁlm. The fringes produced by interference between the direct light and that scattered by the object are circular and the mth bright fringe has a radius ρm , such that R2 + ρ2m − R = mλ, then ρ2m /2R = mλ, and ρ2m = 2mRλ,
(7.69)
to an adequate approximation. This is exactly the same expression as that for the radii of the even order Fresnel zones given in eqn. 6.55. Therefore the pattern appearing in the developed hologram will be a zone plate. When this is illuminated by the reference beam the focusing property of the Fresnel zone plate discussed in Section 6.12.1 will produce a virtual image where the point object had been and a second real image a distance R in front of the hologram. For convenience these images are shown separately in ﬁgure 7.21. An extended object generates a more complex fringe pattern in the recording medium, each point on its surface that can see the hologram making its own contribution at each point on the hologram, which is the case we now consider. Suppose that the reference beam in ﬁgure 7.19 is Er = Er exp (iωt); then the object beam will be Eo (x) = Eo (x) exp (−ikx sin α) exp [ −iφ(x) ] exp (iωt), where x is the coordinate running up the page and α is the angle the object beam makes with the reference beam. Eo and Er are real. The phase, φ(x), and amplitude, Eo (x), are determined by the detailed shape
7.7
Holography 189
of the object. Using eqn. 5.12 the time averaged intensity of the light falling on the recording material at x is T (x) = Re{[ Eo (x) + Er ]∗ [ Eo (x) + Er ]}/2 = { Er2 + Eo2 + Er Eo exp [ −i(kx sin α + φ(x)) ] +Er Eo exp [ i(kx sin α + φ(x)) ] }/2.
(7.70) R
Ideally this intensity distribution is linearly reproduced in the transmission of the processed hologram and in this case the aim should be to choose the exposure time so that the response of the recording medium is linear for the full range of intensities across the hologram. If this is the case then when the hologram is illuminated by the reference beam (with the object removed) the transmitted light amplitude will be E(x) = Er T (x) = Er (Er2 + Eo2 ) exp (iωt) +Er2 Eo exp [ −i(kx sin α + φ(x)) ] exp (iωt) +Er2 Eo exp [ i(kx sin α + φ(x)) ] exp (iωt).
Hologram
(7.71)
Of these three terms the ﬁrst is a forward beam aligned along the reference beam. Apart from a constant the second term is the object beam and produces the virtual image where the object had been. The ﬁnal term describes a beam tilting downward at angle α to the reference beam and this produces the real image. Its amplitude is the complex conjugate of the object beam and results in a pseudoscopic image. The three emerging beams are also recognizable as the diﬀracted beams of order +1 (virtual), 0 (forward) and −1 (real).
7.7.2
Hologram preparation
The following are some general points that bear on the preparation of holograms in the laboratory. In order that the reference and object beams are coherent the path diﬀerences should be small compared to the coherence length of the laser, and the coherence area should extend well beyond the object being photographed. With beams inclined at an angle α it is evident that the fringe spacing is of order λ/ sin α, so that there is a need to keep the whole apparatus stable during the exposure to the level of parts of a wavelength. This is less of a problem in laboratories in which an intense pulsed laser is available, for example a ruby laser giving 20 ns long pulses at 694 nm. The emulsion needs to contain ﬁne enough silver bromide grains (under 10 µm diameter) to record the fringes, but because ﬁne grains require as much light to render them developable as coarse grains the ﬁlm is relatively slow. In the object beam there are large local ﬂuctuations of intensity due to interference eﬀects and so a suitable arrangement is to have the reference beam about three times more intense than the object beam. When looking at a hologram in one’s hand many large circular features catch
R
Fig. 7.21 The diagrams show the reconstruction of a hologram of a point object in the reference beam. The reference beam beyond the hologram is omitted for clarity. In the upper diagram the reconstruction of the real image is seen, and in the lower the reconstruction of the virtual image.
190 Fourier optics
When the hologram is in the near/far ﬁeld of the object it is known as a Fresnel/Fraunhofer hologram. If the object lies in the reference beam and this is incident perpendicular to the ﬁlm the hologram produced is known as an inline hologram. Other alignments give what are called oﬀaxis holograms, which are the simplest to produce and avoid the confusion of superposed images seen with an inline hologram. If the reference and object beams arrive from opposite sides of the ﬁlm this gives a reﬂection hologram.
the eye. These are due to dust and other imperfections; the interference fringes holding the image information are too small to be detected by eye. Phase holograms can be produced by treating the exposed ﬁlm so as to rehalogenate the silver grains. These renewed silver halide grains then migrate to the nearby unexposed silver halide leaving regions of depleted optical density. Phase diﬀraction gratings have a considerable advantage over amplitude gratings because the fraction of light which they diﬀract into each ﬁrst order image can be as large as 33%, as against 6% for amplitude gratings; phase holograms share this advantage. Modern materials such as photopolymers and photoresists are widely used to produce phase holograms. When the former are exposed to light the monomers polymerize and this produces a useful change in refractive index; when the latter are exposed to light they become soft and can be etched away with solvents leaving the image in relief. Surfaces with any degree of roughness seen in laser light show many small bright speckles that change with any movement of the viewer. Speckles are the result of interference between light diﬀracted from nearby parts of the surface and hence their angular size is simply the angular resolution of the camera or the eye doing the viewing. On the ﬁlm plane this is 1.22λf /D, f and D being the respective focal length and entry pupil diameter. Speckles will appear on holograms, and one way to reduce their impact is to take multiple exposures with the beam incident on the object being displaced a minute amount between exposures.
7.7.3
Motion and vibration analysis
Holograms can be exposed in such a way as to reveal patterns of motion of musical instruments and living tissue. One approach is to freeze the motion by pulsing the laser source on for a few nanoseconds, and then to repeat the pulse after a short delay. The interference patterns from the two exposures will match over surfaces that have not moved or have moved a whole number of wavelengths during the delay. These regions will appear bright in the reconstructed image and form contours of equal, known movement. An alternative is to take a long exposure which will give a uniform intensity at moving points, while any nodes will appear very bright. Holograms formed by double exposures are widely used to render visible in three dimensions the ﬂow of ﬂuids, as for example in a wind tunnel. This contrasts with the projected images obtained with Schlieren photography or with the Mach–Zehnder interferometery described in Section 5.7.3. In the case of a wind tunnel the ﬁrst exposure would be made with the air at rest, and the second with the air ﬂowing. When the scene is reconstructed from the hologram the contours of increased and reduced density appear as bright regions in three dimensions.
7.7
7.7.4
Holography 191
Thick holograms
The fringes in a hologram in each small local area are inclined at equal angles to the reference and object beams as shown in ﬁgure 7.22 with spacing Λ = λ/[ 2 sin (α/2 ],
Object beam
Fringes
where α is the angle between these beams. This separation is of order one wavelength so that if the recording medium is tens of microns thick each local region has many parallel fringes which act as Bragg planes. They are equivalent to the planes of high density in the acoustooptic modulator described in Section 7.6. There is constructive interference between these layers provided that the angles of incidence and diﬀraction, θ, satisfy λ = 2Λ sin θ, where Λ is the layer spacing. This condition holds when the reconstruction beam points in exactly the direction of the reference beam, and Bragg scattering occurs in the direction of the beam forming the virtual image. This of course strengthens the virtual image at the expense of the real, pseodoscopic image. It has however an even more important and useful negative eﬀect. When the beam used in reconstruction is incident at some angle different from the reference beam the Bragg condition is no longer satisﬁed and no image is seen. As a consequence the way is open to record multiple images in a thick hologram; recorded and reconstructed with the reference beam rotated to a new angle for each image. This makes it feasible to produce a display hologram that changes as one walks past; a new hologram is revealed at each step the viewer takes. These holograms are usually phase holograms in order that the brightest possible virtual image is obtained. Photopolymer ﬁlms used to make such thick holograms are typically 100 µm thick.
Reference beam
α /2 α /2
Reconstruction beam θ Λ θ
Fig. 7.22 The upper panel shows the formation of interference fringes in a thick emulsion by the object and reference beams. The lower panel shows Bragg scattering from the interference fringes after chemical development.
Thick reﬂection holograms produced by two reference beams incident normally on the recording material from opposite directions will produce planar interference fringes parallel to the surface and spaced at intervals of λ/2. This structure will reﬂect almost all the light at wavelengths close to λ and very little at other wavelengths: it constitutes what is called a notch ﬁlter. Such ﬁlters are useful for making safety glasses which reduce the intensity of a laser beam by many orders of magnitude while not reducing the ambient light at other wavelengths. By using three laser beams simultaneously and panchromatic ﬁlm6 a colour hologram can be produced. The choices of the 476.5 nm Ar+ ion laser line, the 532 nm Nd:YAG line and the He:Ne 632.8 nm line give good rendering of colour. Viewing would seem to require that the same three lasers are used: each wavelength has produced an interference pattern and this will generate an image of that colour when the corresponding laser illuminates the hologram. However there is crosstalk because each
6
Responsive to wavelengths across the visible spectrum.
192 Fourier optics
Hologram Spatial filter
Object
RGB lasers Viewing direction
Image
White light
Fig. 7.23 The upper panel shows Denisyuk’s method for recording a colour hologram. Viewing in white light is shown in the lower diagram.
reconstruction beam is also diﬀracted by the fringes produced by the other lasers. Crosstalk will be suppressed if the recording medium is suﬃciently thick, because a reconstruction beam of colour Y will only satisfy the Bragg condition for the fringes made with colour Y and for no others. As a bonus it is now possible to use a white light source in the reconstruction phase. Only those components in the white light beam whose wavelengths are extremely close to one of the laser wavelengths can satisfy the Bragg condition and these alone contribute to the image. Together these three Bragg selected wavelength bands form a full colour image. All other wavelengths in the reconstructing white light beam pass onward in the reference beam direction. The Bragg diﬀraction requirement for obtaining an image from a thick hologram eﬀectively insulates the three colour images from each other. Figure 7.23 shows a simple layout invented by Denisyuk for producing and viewing full colour holograms. The three laser beams are spatially ﬁltered by a pinhole to give pure spherical wavefronts. Each reference wave passes through the recording material with some portion being reﬂected back from the object in an object beam. Interference fringes will lie roughly parallel to the surface of the recording medium because the reference and object beam enter the recording layer from opposite directions. Several layers of nodes will therefore be formed through the depth of the recording medium so that the Bragg isolation of each colour is acheived. The viewing of the Denisyuk hologram in white light is also shown in ﬁgure 7.23.
7.8
Object plane
Transform plane
Image plane
The techniques of spatial ﬁltering and holography underlie methods that are used for processing information. The applications include the automatic matching of ﬁngerprints and large scale data storage and retrieval. Examples of the latter are the recently launched Gbyte WORM (write once read many times) holographic memories. These techniques use coherent light and rely on the properties of Fourier transforms. The remaining sections of this chapter will be used to introduce a few representative developments in optical information processing.
7.8.1 f
f
f
f
Fig. 7.24 4f architecture for coherent optical processing.
Optical information processing
The 4f architecture
Figure 7.24 shows one common arrangement for optical processing with coherent light, called the 4f geometry. Coherent plane waves arrive from the left, and might be provided by a laser plus beam expander with spatial ﬁltering. The components shown – object plane, the ﬁrst transform lens, the transform plane, a second transform lens and the image plane – are spaced at intervals of a focal length. For simplicity the focal lengths are taken to be equal, and the discussion is restricted to one transverse dimension. As an example of optical processing we consider how to compare a sequence of patterns, fi (x), with a ﬁxed pattern g(x),
7.8
Optical information processing 193
where x is the dimension perpendicular to the optical axis. The Fourier transform of g(x) is generated in the form of a hologram as explained below and this, G(k), is placed in the transform plane. Then in turn the patterns fi (x) are placed in the object plane. The light amplitude transmitted through the transform plane will therefore be the product Fi (k)G(k). This is Fourier transformed by the second lens so that at the image plane the amplitude is h(x) = FT[ Fi (k) G(k) ] = fi (x) ⊗ g(x),
(7.72)
where the convolution theorem has been used. Thus if the patterns match so that fi (x) = g(x), h(x) = g(x )g(x − x )dx. (7.73) Usually g(x) is some fairly random pattern, so that h(x) will be large in the forward direction, when x = 0, and small elsewhere. On the other hand, whenever fi (x) is diﬀerent from g(x), h(x) will have a uniform distribution. A pattern match is therefore signalled by a bright spot appearing on the optical axis in the image plane. A defect of this simple system is that with an inline arrangement the unscattered reference beam travels forward making it diﬃcult to distinguish when a match is achieved. What is required is to displace the signal indicating a match between target and test patterns to a location well away from the optical axis. Figure 7.25 shows the layout used to produce the hologram of the target pattern G(k) which has this desireable property. Part of the broad planar laser beam passes through a transparency that carries the target pattern and is then focused onto the recording ﬁlm. The remainder of the laser beam is deﬂected onto the same sensitive area. Its amplitude at a point x on the recording plane is Er exp (−ikx sin α), where α is the deﬂection and k is the wave number of the laser light. If g(x0 ) is the target pattern at x0 in the object plane, then its diﬀraction pattern (Fourier transform) on the ﬁlm plane at coordinate x is G(kx/f ) = g(x0 ) exp [ −ikx0 (x/f ) ] dx0 , where the sine of the angle in which the light is diﬀracted is approximated by x/f . This arrangement produces a Fourier transform hologram in which the Fourier transform of the object interferes with the reference beam. The particular layout shown in ﬁgure 7.25 makes up a VanderLugt ﬁlter, named after its originator. The eﬀect of using a reference beam tilted at an angle α is that the bright spot indicating a match in the comparison is displaced a corresponding distance f sin α below the optical axis on the image plane in the detector arrangement shown in ﬁgure 7.24. There are practical diﬃculties in getting a positive match when the scale and orientation of the patterns being compared are not the same.
Deflection α
Recording plane
Laser
beam
Object
Lens
f
Fig. 7.25 Setup for producing a VanderLugt ﬁlter.
194 Fourier optics
Both the target and test patterns can be presented electronically using a spatial light modulator. This is usually a liquid crystal display (LCD) of the sort to be described in Chapter 10. Eﬀectively the LCD forms a screen a few centimetres across having a million or so pixels, each of which can be independently set to transmit or absorb light. Referring to ﬁgure 7.24 the LCD screen placed in the transform plane carries the ﬁxed pattern G(k) while the screen placed in the object plane can be switched from one pattern to another, fi (x), in tens of microseconds, making possible comparisons at high rates.
7.8.2
Data storage and retrieval
The ability to store many distinguishable holograms each written with the reference beam incident at a diﬀerent angle implies the possibility of high capacity data storage on holograms. Data densities of up to 100 bits/µm2 have been achieved, which compares well with around 10 bits/µm2 on DVDs but is still less than the 160 bits/µm2 (100 Gbits per square inch) obtained with current longitudinal recording on magnetic disks. The parallelism of data storage on holograms makes for high access rates: all the information on an individual hologram or page is available simultaneously. Given an access time of 1 ms to a particular 1Mbit hologram implies a data rate of 1 Gbs−1 , which is many times faster than the sequential read out rate from other media.
Exercises (7.1) Calculate the Fourier expansion for a repetitive sawtooth wave, f (x), having a repeat distance λ and for which f (x) = x when −λ/2 < x < λ/2 ?
range? What angular spread of microwave beam is required to cover the bandwidth? The velocity of ultrasound in lithium niobate is 6600 ms−1 .
(7.2) What is the fringe count between successive maxima of visibility on an Michelson interferogram taken with a sodium source ﬁltered to pass only the two D lines?
(7.5) A Cd state emits a 643.8 nm line which has a lifetime of 4.1 10−10 s. What is the line width in frequency and in wavelength?
(7.3) (a) Show that convolution is commutative, that is that f (x) ⊗ g(x) = g(x) ⊗ f (x). (b) In ﬁgure 7.2 imagine that one delta function δ(x − xm ) is replaced by −3 δ(x − xm ). What diﬀerence does this make to the convolution? (7.4) A microwave beam with frequencies ranging from 40 to 60 MHz drives a lithium niobate acoustooptic Bragg cell. The optical beam is from a HeNe laser of wavelength 632.8 nm and for this wavelength lithium niobate has a refractive index 2.2. What is the microwave wavelength? What width of laser beam is required to permit 100 frequencies to be distinguished across microwave input frequency
(7.6) A thin transmission hologram is recorded with a HeNe laser of wavelength 632.8 nm, the object and reference beams being inclined at ±30◦ to the normal to the surface respectively. What is the average fringe spacing in the emulsion? (7.7) A camera is used to photograph the virtual image in the setup described in the previous question. As large a depth of ﬁeld as possible is required, however the speckle size increases as the lens is stopped down. What would be a suitable aperture stop to keep the speckle size on ﬁlm under 10 µm? (7.8) (a) Calculate the degree of ﬁrst order coherence of a source emitting a single spectral line with
7.8 Lorentzian shape given by eqn. 7.37. (b) What would be the visibility of fringes seen with a Michelson interferometer using this source?
Optical information processing 195
is pulsed on and oﬀ with a cycle time of τp . If the pulses are square between zero and unity what is the frequency distribution of the radiation?
(7.9) (a) Calculate the degree of ﬁrst order coherence of a source emitting a single spectral line with Gaus (7.11) Show that for both the Lorentzian and Gaussian sian shape given by eqn. 7.40. (b) What would power distributions the value at resonance is apbe the visibility of fringes seen with a Michelson proximately the inverse of the FWHM. interferometer using this source? (7.12) If F (k) is the Fourier transform of f (x), what are (7.10) A monochromatic source of angular frequency ω0 the Fourier transforms of df /dx and d2 f /dx2 ?
This page intentionally left blank
Astronomical telescopes 8.1
8
Introduction
The major telescopes now being used for astronomical observations in the visible and near visible regions of the spectrum are reﬂecting telescopes with primary mirrors of diameters ranging up to 10 m. These huge apertures make it possible to catch enough light to record distant sources emitting when the universe was less than 10% of its present age. Mirrors rather than lenses are the practical optical elements at this scale. Firstly it is essential to support such large pieces of glass over their whole area in order to prevent the physical sagging which would otherwise alter their optical properties signiﬁcantly; and this can only be done for mirrors. Secondly mirrors are free of chromatic aberration, lose little light through absorption in their surfaces and function at wavelengths at which lenses would absorb strongly. In the following section the main design features of astronomical telescopes will be described. Details of some representative large telescopes are listed in table 8.1. Atmospheric extinction of the radiation from astronomical sources is primarily due to water, carbon dioxide, ozone and oxygen molecules. Each species has absorption bands and they all scatter radiation at other wavelengths. The molecular scattering of light1 is proportional to λ−4 . Absorption below 300 nm wavelength is mainly due to ozone and is almost complete. Starting at 300 nm, the optical window, with extinction of 10% or so in the visible, extends into the infrared with increasingly frequent absorption bands, some almost opaque, to around 15 µm. Most molecular densities fall oﬀ exponentially with height, and at a height of 4 km a telescope is above 95% of the atmospheric water vapour. However the major part of the ozone is above 20 km. The twinkling of the stars is an indication of atmospheric turbulence, and this turbulence reduces the resolving power of any telescope on the Earth’s surface. Typically for telescopes located at 3000 m high sites in dry areas the best resolution achievable (the seeing) at 500 nm wavelength is around 0.4 arcsec, which is what could be achieved with a telescope above the atmosphere having an aperture diameter of only 25 cm; but the larger telescope retains its advantage in gathering more light. Adaptive optics is the term used to describe the ways and means by which the image distortions due to atmospheric turbulence are corrected in real time. These techniques for recovering the full potential
1
This is called Rayleigh scattering and is discussed in Chapter 12
198 Astronomical telescopes
Table 8.1 A table of the parameters for some representative large telescopes. Telescope
Design
Diameter/segments
Mounting
Location/height
Keck I,II Hobby–Eberle VLT14
Ritchey–Chretien Spherical primary Ritchey–Chretien
10 m/36 9.2 m/91 8 m/1
Altaz Fixed elevation Altaz
Hawaii/4208 m Texas/2072 m Chile/2635 m
of very large telescopes are described in the third part of the chapter. Adaptive optics has scientiﬁc applications elsewhere: in controlling the wavefronts of laser beams used in microscopy, in laserinduced fusion and in studies on human vision. Individual telescopes are being used together in arrays in order to further extend the resolution in the visible and near visible part of the spectrum. Light from the telescopes is brought together to interfere and the fringe patterns, interferograms, used to give information on the source. At Mauna Kea the two 10 m Keck telescopes and four 1.8 m telescopes will be used as an interferometer with a largest separation of 135 m – giving a potential resolution at 1 µm wavelength of approximately 7.4 10−9 rad or 1.5 mas (milliarcsec). Interferometry with multiple apertures is discussed in the fourth part of the chapter. r
z Focus
Several Michelson interferometers have been built with kilometre long arms to detect the tiny change in relative length between their arms which is expected when a gravitational wave strikes the Earth. Only rare, catastrophic events, such as supernova explosions in our galaxy are likely to give detectable signals. This very diﬀerent application of interferometry in astrophysics is the topic of the ﬁfth segment of the chapter. A ﬁnal short segment of the chapter is used to describe the observations of gravitational lensing and includes a simple account of the origin of this general relativistic eﬀect.
8.2 fp
Fig. 8.1 Paraboloid mirror.
Telescope design
Telescopes need to have a large entry pupil in order to collect the maximum light and so to detect weak and distant astronomical sources; and also to resolve the details of the structure of sources. An unresolved star is one whose angular diameter is much smaller than the angular resolution of the telescope viewing it. The entry pupil of a telescope is the primary mirror, the one which ﬁrst intercepts the light. Now the angular resolution due to diﬀraction alone, ∆θ, was calculated in Section 6.7, ∆θ = 1.22λ/D,
(8.1)
for light of wavelength λ at a circular aperture of diameter D. In the case of the Hubble Space Telescope, the HST, with a 2.4 m diameter primary mirror the resolution limit imposed by diﬀraction is 2.54 10−7 rad
8.2
Telescope design 199
or 52.4 mas. The aberrations in a mirror telescope are made comparably small by using mirrors whose shapes are paraboloids and hyperboloids of revolution. Paraboloid mirrors have the useful property that all incident rays parallel to the axis pass, after reﬂection, through the geometric focus as shown in ﬁgure 8.1. A paraboloid has the equation r 2 = 4fp z where z is the axial distance from the pole of the paraboloid and r is the distance oﬀ axis. fp is the equivalent paraxial focal length of the mirror. However the aberration of a paraboloidal mirror rises rapidly when the target moves oﬀ axis: the angular size of comatic ﬂare projected on the sky is α ≈ θD 2 /(4fp2 ),
(8.2)
where θ is the angle oﬀ axis. Fortunately hyperboloid shaped mirrors also give a good axial focus and have oﬀaxis aberrations that are opposite in sign to those of paraboloidal mirrors. The combination of one mirror of each shape has the potential for producing an image for which the oﬀaxis aberrations largely cancel. Figure 8.2 shows how all the rays passing through one geometric focus of a hyperboloid mirror will, after reﬂection, pass through the second focus. Using the coordinates shown the equation of the hyperboloid is r 2 = (z 2 − a2 )(e2 − 1) where e is called the eccentricity. The linear magniﬁcation is the ratio of the image to object distance m = (e + 1)/(e − 1). If the hyperboloid mirror is placed so that its ﬁrst geometric focus coincides with the focus of the paraboloid, then the ﬁnal image will lie at its second focal point. The size of the image of a star of angular diameter θ produced by the two mirror combination would be D∗ = mfp θ.
r
Focus
Focus
a
(8.3)
As noted above, the oﬀaxis aberrations are very much reduced if the curvatures of the mirrors are chosen suitably. The standard layout for the primary and secondary mirrors is shown in ﬁgure 8.3 with the ﬁnal image being projected through a hole in the primary. The classical Cassegrain telescope shown in the upper panel has a convex secondary mirror, while the Gregorian telescope has a concave secondary mirror. A cross shaped frame of thin rods, called a spider supports the secondary from the telescope tube. Thus light travelling toward the primary is Fresnel diﬀracted by the secondary and its spider, and if the image of a star is overexposed it acquires thin cruciform arms. In both these telescope designs the primary is the aperture stop and also the entrance pupil, so that it is the diameter of the primary mirror which determines the telescope’s potential light gathering power and angular resolution. The Cassegrain is more compact than the Gregorian for the same focal length, which gives a decisive advantage when the telescope weighs tons and is several metres in length. In addition the Cassegrain secondary mirror and the hole in the primary can be made smaller. The
z
a(e+1)
a(e1)
Fig. 8.2 Hyperboloid mirror. The dotted curve shows where the other half of the hyperboloid surface would lie.
200 Astronomical telescopes
Cassegrain
Primary
Final
focus
focus
Gregorian
Final focus
Fig. 8.3 Cassegrain and Gregorian telescopes.
ﬁnal image surface is curved toward the secondary mirror in the classical Cassegrain and its variants. In one popular variant on the Cassegrain telescope the shapes of both mirrors have been altered slightly to make an aplanatic (free of coma as well as spherical aberration) telescope; the primary mirror also becomes hyperboloid in shape, but with an eccentricity only slightly diﬀerent from that of a paraboloid mirror. This Ritchey–Chretien design was used for the HST, Keck and many other large telescopes. The principal residual aberration is then the astigmatism with angular size in the sky, given in arc seconds Azimuth
α = 0.5mθ2 /(57.3f /#),
(8.4)
where θ is the angle oﬀ axis in arcminutes, m is the magniﬁcation produced by the secondary and f /# is the focal ratio for the complete telescope. Altitude
Fig. 8.4 Telescope in altaz mounting.
When observing a star or galaxy an astronomer needs to be able to hold the image at a ﬁxed point in the ﬁeld of view as the Earth rotates. Two mountings suitable for this purpose are widely used to support telescopes and are shown in ﬁgures 8.4 and 8.5. The ﬁrst diagram illustrates the altaz mounting in which one rotation axis is vertical and the other horizontal. In the alternative equatorial mounting one axis points to the
8.2
pole star and the other is horizontal. The latter mounting has the advantage that once a target is in view the only motion required thereafter to retain the target in view is rotation around the polar axis. In the case of the altaz mounting the tracking speed is variable and even when the target is tracked precisely its image rotates in the ﬁeld of view. This latter feature requires correction using additional optical elements if a long exposure – which is often the case – is required. All these weaknesses of the altaz mounting are less signiﬁcant now that computer control can be complex, reliable and cheap. When the telescope weighs many tons, having a layout in which the centre of gravity can lie directly over the bearings is a decisive advantage for the altaz mounting. It leads to a lighter, smaller frame and reduces the size of dome required to house the telescope. Multimetre telescopes mirrors are extremely diﬃcult to cast, and those of over 5 m diameter are often segmented.2
Example 8.1 The HST is a Ritchey–Chretien telescope, with mirrors having outer diameters of 2.4 and 0.8 m. Their equivalent radii of curvature, on axis, are 11.04 and 1.36 m, so their paraxial focal lengths are +5.52 and −0.68 m respectively; the mirrors being placed 4.91 m apart. Thus the focus of the primary mirror, the prime focus, is 0.61 m beyond the secondary mirror. If the distance of the ﬁnal focus from the secondary is v, then applying the mirror eqn. 3.5 gives
Telescope design 201
2
The twin 10 m Keck telescope primary mirrors are made up of 36 hexagonal segments cast from a ceramic material, Zerodur, of low thermal expansion coeﬃcient (∼10−7 K−1 ). Each segment is 76 cm thick, 1.8 m across and weighs 0.4 t. A single Keck instrument plus supports weighs 300 t and occupies an eightstorey high spherical shaped dome.
Polar
1/v = 1/0.61 − 1/0.68, whence v = +6.41 m, putting the ﬁnal focus 1.5 m behind the primary mirror. The overall focal length f is given by eqn. 3.39,
Telescope
1/f = 1/fp + 1/fs − d/(fp fs ), where fp,s are the mirror focal lengths and d their separation and this yields a focal length 57.2 m. Then one arcsec in the sky projects onto 279 µm in the image plane: the plate scale is said to be 279 µm/arcsec. At this scale a pixel in the CCD arrays, which are typically 15 µm across, covers 0.05 arcsec. The diﬀraction limit at 500 nm wavelength given by eqn. 8.1 is 0.04 arcsec, so the pixel size is well matched to the attainable resolution. The long overall focal length relative to the short optical path within the telescope shows that the Ritchey–Chretien and classical Cassegrain telescopes are in fact mirror telephoto systems. Initially the HST showed unexpected spherical aberration because the primary had been shaped to an incorrect proﬁle, departing at the periphery by 2 µm from the correct proﬁle. This was rectiﬁed by adding a pair of mirrors. A ﬁeld mirror, M1 , placed a little ahead of the focal plane imaged the entry pupil onto a second mirror, M2 , so that points on M2 were in onetoone correspondence with points on the primary mirror. M2 was shaped so as to cancel the aberration of the primary exactly.
Declination
Fig. 8.5 mounting.
Telescope
in
equatorial
202 Astronomical telescopes
8.2.1
Auxiliary equipment
The full range of spectrometers described in Chapters 6 and 7 are deployed in studying emission spectra of celestial sources in the visible and near visible parts of the spectrum. For high resolution work echelle gratings are often used. Combinations of crossed prisms and gratings permit the simultaneous examination of a wide spectral range. Where appropriate, optical ﬁbres are used to transfer light from individual star images to corresponding individual points along the entry slit of a spectrometer. One end of each ﬁbre peers through a precisely located hole in a plate covering the image plane, while the other end is placed at its chosen location along the spectrometer slit. In this way several hundred stellar spectra can be recorded simultaneously. f
f coll
θ
f cam
α
Principal Field plane lens
Collimator
Camera
Fig. 8.6 Image reducer for matching image area to detector area.
A very simple auxiliary component is the ﬁeld lens used to produce a ﬂat image plane to match the planar surface of detectors. This is positioned close to the secondary focus. More complex optics provides magniﬁcation or demagniﬁcation to match the image size or resolution to that of detectors. A focal reducer is shown in ﬁgure 8.6. This consists of a ﬁeld lens near the telescope focus, a collimator and the detector camera. Rays from a star lying on the optical axis are shown as solid lines. The ﬁeld lens images the exit pupil of the telescope to just ﬁll the collimator aperture which is the entry pupil of the camera/collimator combination. If fcol and fcam are their focal lengths the image size is changed by a factor fcam /fcol . The ﬁeld lens’ area should match the ﬁeld of view of the telescope, f θ, where f is the telescope’s focal length, and θ its angular ﬁeld of view in the sky. In order to maintain the etendue along the optical chain we require Dcol α = Dθ, where Dcol is the collimator diameter, D the primary diameter and α is the angular ﬁeld of view of the collimator. Additionally we see in ﬁgure 8.6 that f θ = fcol α. Combining these two requirements determines the collimator diameter Dcol = Dfcol /f. The region of parallel beam between the collimator and camera in the focal reducer provides a natural location to insert dispersive elements such as prisms, gratings or Fabry–Perot etalons, or for simple ﬁlters.
Nasmyth
focus
Fig. 8.7 Nasmyth focus along the altitude axis of a Cassegrain telescope in an altaz mounting.
Light auxiliary equipment can be mounted at the Cassegrain focus behind the primary mirror. Figure 8.7 shows an alternative in which a plane mirror diverts the light along the altitude axis to what is called the Nasmyth focus. In the case of the 10 m Keck telescopes heavy equipment at the Nasmyth focus is carried around on a horizontal frame which follows the telescope’s azimuthal rotation. This avoids locating heavy gear at the Cassegrain focus where it would be carried around in three dimensions. Another alternative is the Coude focus, which requires further mirrors to put the image in a stationary position. This necessitates a long focus, rotating mirrors and a narrow ﬁeld of view.
8.3
Schmidt camera 203
Light from stars is dispersed in travelling through the Earth’s atmosphere, and when a star is not directly overhead (not at the zenith) the image becomes a coloured strip. This dispersion is removed by putting Risley prisms in the optical path. An individual Risley prism is made from two thin prisms glued face to face so that their deﬂections cancel. According to eqn. 2.28 a single prism of narrow angle α and refractive index n would provide a deﬂection (n − 1)α and a dispersion α(dn/dλ). In a Risley prism the types of glass used for the component thin prisms are chosen so that (n1 − 1)α1 = (n2 − 1)α2 , while at the same time there is a net dispersion. The compensator for atmospheric dispersion is a pair of Risley prisms in series. When viewing directly overhead they are rotated to be in opposition and their dispersions cancel; for viewing in other directions the angle between the pairs is changed so that their net dispersion cancels that of the atmosphere.
8.3
Spherical mirror
Schmidt camera
A radically diﬀerent telescope design that gives a much wider ﬁeld of view than the Ritchey–Chretien is the Schmidt camera. In 1929 Schmidt came to appreciate that if the aperture stop of a concave spherical mirror is placed at the centre of curvature then all directions of incidence through the centre of this stop to the mirror are radial and are equally good optical axes. Consequently with this layout the coma and astigmatism are eliminated. What remains is the spherical aberration, which Schmidt was able to reduce drastically by placing an aspheric glass corrector plate across the stop aperture. Figure 8.8 shows the principal components of Schmidt’s design. The corrector plate cancels out spherical aberration by being convex at the centre (focusing) while it is concave at the periphery (defocusing). This surface ﬁguring can only remove spherical aberration at a single wavelength so there is residual spherical aberration at other wavelengths. That remaining aberration can be further reduced if the corrector plate is made into an achromatic doublet like those discussed in Section 3.7.7. Two drawbacks of the Schmidt design are that the image surface is highly curved and that the image is diﬃcult to access. A typical large aperture Schmidt is the UK Schmidt of 1.24 m diameter, focal length 3.07 m located at the Siding Spring observatory in Australia. It has a ﬁeld of view of 6.6◦ over which the resolution is about 1 arcsec and has a plate scale of 67 arcsec/mm. Originally it was used to produce an atlas of the southern sky and is now also used for measuring the spectra of sources. Schmidts of much larger aperture are impractical because a larger thin corrector would sag signiﬁcantly, and because the overall length, being twice the focal length, makes mounting diﬃcult. A common variation on Schmidt’s design, due to Maksutov, uses a thin meniscus lens to correct the spherical aberration.
Image surface R Corrector
Fig. 8.8 Schmidt camera.
204 Astronomical telescopes
Stellar luminosities The apparent luminosity, Wm−2 , of a star is the actual power density arriving at the Earth. The measured absolute luminosity, L, of a star is the total radiated power in watts, which can be calculated from the apparent luminosity if the distance to the star, D, is known L = 4πD2 . Apparent magnitude, m, is a scale devised by Pogson in the nineteenth century to quantify the visual scale of intensity that went back in some form to classical Greece. Values of 1 and 6 correspond respectively to the brightest stars (excluding the Sun) and faintest stars visible to the eye. The apparent luminosity corresponding to an apparent magnitude zero is explicitly deﬁned to be 2.52 10−8 Wm−2 in the V–band, that is using a broadband ﬁlter centred at 510 nm to capture all visible wavelengths. Then, because the eye responds logarithmically rather than linearly to light intensity the relationship between apparent magnitude and apparent luminosity is m = −2.5 log10 [ /(2.52 10−8W m−2 )].
(8.5)
Finally absolute magnitude is deﬁned as the apparent magnitude which a star would have at a distance of 10 parsec from the Earth, i.e. at a distance of 3.086 1017 m. The Sun itself has absolute (Vband) magnitude 4.72 and an apparent magnitude −26.85. Very large telescopes can identify stars with apparent magnitudes as high (weak) as +25, for which the apparent luminosity is 2.52 10−18 Wm−2 .
8.4
Atmospheric turbulence
Observatories are sited on remote mountains which enjoy clear skies and low levels of humanity’s light pollution. However, unlike the HST, they cannot escape the eﬀects of turbulence in the atmosphere. The energy from turbulent motion is dissipated as local heating in the atmosphere and this results in important variations locally in the refractive index of the atmosphere. These variations cause a star’s image to scintillate, meaning to move and to change in intensity at a rapid rate. These movements are suﬃciently small and rapid that all the eye registers is a twinkling. An incoming wavefront from a star is planar before entering the atmosphere but its surface becomes crumpled by the local variations in the refractive index of the atmosphere. The area over which the wavefront which arrives at the Earth is relatively ﬂat (but not necessarily parallel to the incident wave above the atmosphere) is called the Fried parameter r0 . More precisely: over a distance r0 the phase variation of the wavefront has a root mean square value of 1 rad. Diﬀraction at this equivalent aperture gives an angular resolution (point spread function) of approximately λ/r0 . At a mountain observatory site with good seeing r0 can be around 25 cm at a wavelength of 500 nm and then the resolution is
8.5
Adaptive optics 205
0.4 arcsec. The shape of the wavefronts arriving at a telescope changes on a timescale τ0 which can be 50 ms in conditions of good seeing at 500 nm. It follows that the resolving power of telescopes with mirrors of any larger diameter than r0 is degraded to that of a mirror of diameter r0 , although their light gathering capacity is unaﬀected. The Fried parameter and τ0 both vary strongly with wavelength r0 ∝ λ6/5 , so that the extent of the point spread function, λ/r0 , is proportional to λ−0.2 . This gives a slow improvement in seeing as the wavelength increases. The eﬀect of this turbulence on the image seen depends on the size of the telescope’s aperture. When an unresolved star is viewed through a telescope whose diameter d is less than the Fried parameter the image has an Airy disk of angular diameter λ/d. This image moves jerkily at intervals of average duration τ0 with angular displacements of order λ/r0 . On the other hand if the telescope has diameter D much larger than the Fried parameter the image consists of speckles of angular size λ/D. These speckles, numbering roughly (D/r0 )2 , continually move, fade, coalesce and reform with the same characteristic timescale τ0 . A long exposure produces a blurred image of angular size λ/r0 , and with Strehl ratio3 (r0 /D)2 .
8.5
Adaptive optics
The techniques described as adaptive optics are used to overcome the eﬀects of atmospheric turbulence and to recover the potential resolving power of large diameter telescopes. The ﬁrst step is to sense the shape of the wavefront arriving from an unresolved star which is suﬃciently close in direction to the target for their images to suﬀer essentially the same distortion; the reference star is called a guide star. This information on the wavefront shape is then used to deform a ﬂexible mirror placed in the optics of the telescope in such a way that after reﬂection the wavefronts recover their planar shape. The control sequence has to respond on a timescale of milliseconds in order to compensate the changing distortion faithfully and in real time. Figure 8.9 shows schematically the adaptive optics which could be placed at the Nasmyth focus of a large telescope. After reﬂection from the deformable mirror the light encounters a beam splitter, with part being directed to a wavefront sensor and part going to the detectors that are recording data, which are called the science instruments. Electronic signals from the wavefront sensor are used as input to a processor that controls actuators which change the shape of the deformable mirror. The commonest wavefront sensor is the Shack–Hartman sensor shown in ﬁgure 8.10. A planar, square array of identical lenslets focuses the
3
See Section 7.2.1
206 Astronomical telescopes
Source and guide star
Telescope Science instrument
Deformable Beam splitter
mirror
Wavefront sensor Control signals Computer
Fig. 8.9 Adaptive optics.
Wave fronts
Lenslets
CCD array
Fig. 8.10 Shack–Hartman wavefront sensor. In the upper panel an undistorted plane wave is incident and in the lower panel the incident wave is distorted. The pattern of image points on the detector array is shown on the right in each case.
incoming light from the guide star onto a CCD array having for example four pixels per lenslet. In the upper panel of the ﬁgure an undistorted plane wave is incident and the image on the CCD is a square array of dots, each dot lying on the optical axis of its lenslet. What is shown in the lower panel is the result when the incident wave is distorted. Each image spot is now displaced by a distance and in a direction determined by the local orientation of the wavefront across its particular lenslet. If the wavefont has a tilt of ∆θ and the lenslet focal length is f , then the displacement is simply f ∆θ. A section through a distorted wavefront appears in ﬁgure 8.11 where the vertical lines separate the regions seen by individual lenslets. Once the direction and magnitude of tilt over each cell is known it is evident that the whole wavefront can be reconstructed with a precision set by the lenslet diameter. Note for future reference that the wavefront sensor will not detect an overall delay or advance of the wavefront due to a change in the refractive index common to the whole area of the telescope aperture. This piston component of the distortion becomes signiﬁcant when light from two or more telescopes is brought together to interfere. The most important part of wavefront correction is to remove the overall tip or tilt of the wavefront because this accounts for just under 90% of the image distortion. This correction can be performed using a
8.5
rigid plane mirror that can be rotated about either of two orthogonal axes in its own plane as shown in ﬁgure 8.12. If this tip/tilt correction is insuﬃcient a second, this time deformable mirror can be used to remove the remaining distortion. Deformable mirrors are available with up to several thousand actuators over the surface. One type of deformable mirror has a thin glass or ceramic skin which is bonded to actuators mounted on a ﬂat rigid plate. The actuators are generally piezoelectric rods which expand or contract under an applied voltage and their movement ﬂexes the mirror surface. The total correction sharpens the image of an unresolved star to the extent that in the case of Keck II the Strehl ratio improves from less than 0.01 to 0.35.
Adaptive optics 207
Wavefront
Fig. 8.11 Wavefront segments reconstructed by the Shack–Hartman wavefront sensor.
If the eﬀective height of the turbulence in the atmosphere is H above ground level then the isoplanatic angle within which the wavefront distortion is uniform is given by
θ0 ∼ r0 /H.
(8.6)
H is typically 5 km so that with a Fried parameter of 15 cm the isoplanatic angle is only a few arc seconds. Unfortunately the isoplanatic regions around the available guide stars cover only a small fraction of the sky. A less satisfactory alternative is to use a narrow laser beam to generate an artiﬁcial guide star near to the target. The most eﬀective method is to excite ﬂuorescence at a wavelength 589 nm in sodium atoms concentrated in a layer lying around 90 km above the Earth. A diﬀerent scheme is to back scatter a laser beam oﬀ the atmospheric molecules, a process that produces a guide star at 10 to 20 km above the Earth. Laser guide stars have an inherent drawback. On its way upward the laser beam is deﬂected by the turbulence and on its return it undergoes an almost equal and opposite deﬂection. The artiﬁcial guide star remains stationary and cannot give any information on the tip/tilt component of the atmospheric distortion. However any faint star lying within the isoplanatic angle of the target can be used to provide the pointing information that determines the tip/tilt correction, despite this star being too faint to help beyond that. Extreme care is taken to reduce variations in telescope performance arising from factors local to the telescope and its protective dome. Local convection, diﬀerential heating of the telescope components and radiative cooling of the mirrors during a night’s observation are eliminated as far as possible. Measures such as cooling the dome interior to nighttime temperatures during the day, insulation of the dome from work areas below it, forcing a slight downdraft of air through the slit during viewing, and using baﬄes to reduce wind movement all contribute to maintaining a stable environment.
Fig. 8.12 Mirror used for tip/tilt correction.
208 Astronomical telescopes
8.6
Michelson’s stellar interferometer
The ﬁrst successful measurement of a star’s angular diameter was carried out by Michelson in 1921 using the stellar interferometer sketched in ﬁgure 8.13, making a type of measurement foreseen some 60 years earlier by Fizeau. Light from the star, Betelgeuse, was received by a
Baseline B S’
S D d
Fig. 8.13 Michelson’s stellar interferometer. s and s are unit vectors in the directions of the sources. B is the baseline vector length and D the vector separation of the inner mirrors.
4
This was the 2.5 m Hooker telescope located at a height of 1742 m on Mt Wilson in California. Betegeuse is the red star in the right shoulder of Orion, αOrionis. At a distance of 430 lightyears (4.1 1018 m) it is the nearest red supergiant star. The current accepted value for its angular diameter is 0.054 arcsec, making its geometric diameter about 650 times larger than that of the Sun.
pair of plane mirrors mounted on a long bar attached to the frame of the telescope4 and reﬂected into the telescope by two further mirrors. These outer mirrors could be moved apart along the bar to a maximum separation of over 6 m. The resultant image at the telescope’s focus was therefore crossed by fringes due to interference between the light following the separate mirror paths. It is very important when thinking about astronomical interferometry to remember that the light from any point on a star’s surface is incoherent with the light from any other point on that star, or equally on any other star. Consequently the intensity pattern seen by Michelson was the sum of the intensities of the interference patterns due to light from each region of the star. The complex amplitude at the telescope’s image plane caused by light from one edge A of the star would be E = E0 [exp (−ikp1 ) + exp (−ikp2 )] exp (iωt),
8.6
Michelson’s stellar interferometer
209
with light of angular frequency ω and wave number k. p1 and p2 are the lengths of the two diﬀerent mirror paths and δ = p1 − p2 their diﬀerence. Thus the time averaged intensity (8.7)
∆ = k(δ − δ ) = k(s − s ) · B.
(8.8)
Now because s and s are unit vectors and the star’s angular size, ∆θ, is small we have ∆θ = s − s . Then with B aligned parallel to s − s eqn. 8.8 reduces to ∆ = kB∆θ. (8.9) In Michelson’s apparatus ∆ is varied by moving the outer mirrors apart symmetrically. When this happens B increases, ∆ increases and the sets of fringes due to A and A move apart. Eventually a separation of the outer mirrors is reached at which the phase diﬀerence ∆ is exactly π. Then the fringes due to A and A are exactly out of step and a uniform total intensity is produced across the star’s image. Figure 8.14 illustrates how the fringe visibility changes with increasing phase diﬀerence. In this ﬁgure the fringe envelope is the single slit diﬀraction pattern due to a mirror, of width d acting as an entry aperture. The spacing of the fringes within the envelope is that for a point source viewed by the inner mirrors. There are therefore D/d fringes across the envelope, and what was very important for Michelson, their location would not change when the outriggers were moved in or out. At the setting giving zero visibility, where ∆ = π, eqn. 8.9 gives ∆θ = π/kB = λ/(2B). Michelson used this result to deduce the angular size of Betelgeuse, obtaining a value 0.047 arcsec. The essential feature of the measurement is that the angular resolution has been boosted to λ/(2B) by using outriggers whose separation is B, whereas the telescope alone has a resolution of λ/D. After Michelson’s measurements the technique languished because longer outriggers proved very unstable and because of atmospheric turbulence. The diﬃculty lay in the need to hold the two path lengths equal to within the source coherence length, for otherwise interference is not possible. Beginning in the 1940s measurements of stellar diameters were made with the new technique of intensity interferometry which requires
I( θ ) / I(0)
0.8 0.6 0.4 0.2 0 1
0.5
0 d sin θ / λ
0.5
1
0.5
0 d sin θ / λ
0.5
1
0.5
0 d sin θ / λ
0.5
1
1 I( θ ) / I(0)
Let the baseline, that is the vector separation of the two outer mirrors, be B; let D be the separation of the inner mirrors; and let the unit vector in the direction of A be s. Then δ = s · (B − D) at the geometric image point. The corresponding path diﬀerence for light arriving at the same image point from the opposite edge A of the star is δ = s · B − s · D. Thus there is a phase diﬀerence between the sets of fringes produced separately by the sources A and A
1
0.8 0.6 0.4 0.2 0 1
1 I( θ ) / I(0)
I = EE ∗ /2 = E02 [1 + cos (kδ) ].
0.8 0.6 0.4 0.2 0 1
Fig. 8.14 Fringes with decreasing visibilities: 0.81, 0.33 and 0.14.
210 Astronomical telescopes
less strict path equality. This technique, developed by Hanbury Brown, will be discussed in Chapter 17. In more recent times the development of lasers, fast electronics and computers provided tools with which astronomers could carry stellar amplitude interferometry much further. Long baselines can be measured with a precision of tens of nanometres and altered suﬃciently fast to compensate continuously for the piston component of atmospheric distortion. Interference patterns (interferograms) can now be recorded using light from separate telescopes or from an array of apertures placed in front of a single large telescope; and where Michelson only made a single measurement of the baseline length at which the fringe visibility went to zero, nowadays the visibility itself is measured for as many baseline lengths and orientations as possible. From this new information the images can be reconstructed of complex distant sources which might be too small for an individual telescope to resolve. This ﬁeld of aperture synthesis interferometry will be described in the following sections.
8.7
For simplicity we assume that the source intensity and the detector eﬃciency are both constant over the range (λ0 ±∆λ/2) and zero outside this range ∆λ is small compared to λ0 so that δλ = λ − λ0 is also small and hence the approximation can be made that
Modern interferometers
Whatever the interferometer the diﬀerence between the lengths of the two optical paths has to be kept to less than the coherence length if interference is to occur. This poses a problem if the light from two separated telescopes is brought to interference as illustrated in ﬁgure 8.15. Each telescope collects light from a star, and this is focused into a parallel beam and guided by mirrors so that the beams from both telescopes are brought together at a common detector. The geometric paths from a star not lying directly overhead diﬀer by much more than a coherence length. In order to cancel out this diﬀerence delay lines are incorporated in the optical paths from each telescopes. The mobile mirrors in the delay lines are carried on trolleys moving along rails. The lengths of the complete optical paths are monitored continuously with laserbased heterodyne Michelson interferometers of the type described in Section 14.6.3.
Detectors, perhaps placed behind wavelength ﬁlters, respond to a range of wavelengths, which leads to modiﬁcations in the treatment given = cos [ (2πδ/λ0 ) (1 − δλ/λ0 ) ] 2 in the last section. Equation 8.7 for a single point object becomes = cos (2πδ/λ0 ) cos (2πδ δλ/λ0 ). λ0 +∆λ/2 The integral I= E02 [ 1 + cos (kδ) ] dλ, (8.10) cos (kδ) = cos [ 2πδ/(λ0 + δλ) ]
λ0 −∆λ/2
cos (kδ) dλ
= cos (2πδ/λ0 )
cos (2πδ δλ/λ20 ) dλ
= (λ20 /πδ) cos (k0 δ) sin (πδ ∆λ/λ20 ) = cos (k0 δ)sinc(πδ/Lc ).
where the integral is taken over the range of wavelength detected. The integral reduces to I = E02 ∆λ [ 1 + sinc(πδ/Lc ) cos (k0 δ) ] ,
(8.11)
where Lc = λ20 /∆λ is the coherence length of the radiation: see eqn. 5.25. Thus the length of the fringe train is of order λ0 /∆λ fringes, which amounts to 10 fringes if the bandwidth is 10%. At the centre of
8.7
Baseline B
s
Afocal telescope Beam combiner
Delay lines
Fig. 8.15 Telescope interferometer with delay lines. The telescopes have auxiliary lenses to produce a parallel beam, making them afocal.
the pattern the path diﬀerence is close to zero so that all wavelengths remain in phase giving an easily recognizable white light fringe; while at the edges of the pattern the fringes are coloured. Atmospheric turbulence poses other diﬃculties for interferometry with telescopes because it causes the optical path lengths from source to each telescope to change rapidly and in an uncorrelated manner. One approach is to use telescopes or mirrors that have diameters of order of the Fried parameter, r0 , so that the wavefront distortion is at least uniform across each telescope and then to equip each telescope with tip/tilt correction using adaptive optics. However this leaves uncorrected the piston component of the distortion, which can be of order tens of microns in the visible spectrum and which is changing on a time scale of τ0 . The fringes wander by corresponding amounts. Michelson in his experiments found that although the fringes moved about due to turbulence they did not change their proﬁles, which is vital for any measurement. With his relatively compact interferometer he was able to hold the fringes in view by having a glass block mounted in one of the two optical paths. This he would tilt by hand to hold the fringes in view. In modern systems the fringes are tracked electronically and kept in view by altering the optical delay lines appropriately. Equality of the two total optical paths from the star is maintained to a precision of several nanometres for interferometers with baselines of order 100 m. Most observations have been made with interferometers using detectors sensitive to infrared radiation because τ0 , r0 and the isoplanatic angle all improve (increase) with in
Modern interferometers 211
212 Astronomical telescopes
creasing wavelength.
Telescopes
Beam splitter Detector
Detector + Amplifier
Fig. 8.16 Pupil plane interferometer. The full lines are light paths and the broken lines carry electrical signals.
A stellar interferometer in which, as in ﬁgure 8.13, the fringes fall on a CCD detector array in the image plane is termed an image plane interferometer. Figure 8.16 shows a second variant, called a pupil plane interferometer, in which the collimated beams impinge on a beam splitter and the light from each exit face falls on a single detector. These plane wavefronts interfere leaving the whole ﬁeld of view at each detector uniformly illuminated: when one is dark the other is light – as we now show. Suppose A and A are the complex amplitudes of the incoming beams, with A = A , then the outputs falling on the two detectors are A1 = A + iA , and A2 = iA + A , (8.12) where the factor i is due to the phase diﬀerence between the reﬂected and transmitted beams from a beam splitter. Now if A = −iA then A1 = 0 and A2 = 2A ; thus when one exit pupil is dark the other is bright, and vice versa. Subsequently the electronic signals from the two detectors are subtracted to give a resultant ∆I = I1 − I2 = A1 A∗1 − A2 A∗2 = 2m(A∗ A ).
(8.13)
In the pupil plane interferometer the fringes are scanned across the detectors by dithering the position of a mirror in one optical delay line so that the path diﬀerence changes by many wavelengths. A single detector can be used rather than the detector array needed for image plane interferometry.
8.8 Source
R
s0
s
Telescope
y O x
Telescope B
Fig. 8.17 Coordinate system for aperture synthesis.
Aperture synthesis
There is much more information contained in the fringe patterns than the single value of the angular radius of a star. By combining data from interferometers using baselines of diﬀerent lengths and orientations it is possible to obtain images of celestial objects. The analysis techniques which are in use were for the most part developed earlier to extract images from analagous radiotelescope measurements. In this section the basic analysis will described, with later sections covering important experimental and analysis details.5 Figure 8.17 shows an extended astrophysical source viewed by an interferometer. Two orthogonal axes, Ox and Oy, are indicated at the source and these are repeated at the telescopes, with the third orthogonal axis being directed along the unit vector s0 pointing from the centre of the baseline to a reference point on the source, R, called the phase 5 This section has been adapted from slides of Professor Haniﬀ’s talk ‘Optical Interferometry – A Gentle Introduction’ at the 2003 Michelson Interferometry Summer School at the California Institute of Technology, Pasadena, CA.
8.8
centre. Let the source brightness be I(s) where s is the unit vector pointing from the centre of the baseline to any point on the source. Then the intensity at the detector is, apart from constants, P (s0 , B) = I(s) [ 1 + cos (kδ) ] dA, (8.14) where dA is an element of area of the source centred in a direction with the unit vector s, and where k points along s0 . The diﬀerence in length between the two optical paths from the elements of the source at s to the detector is δ = s · B + p, (8.15) where p is the diﬀerence between the optical paths from telescope entrance pupil to detector – including the delay lines. Now the detector responds to a range of wavelengths, usually restricted by a ﬁlter, for example the Kband from 2 to 2.4 µm. The overall path diﬀerence between the two paths must be held less than the coherence length which in this case is ∼10 µm. The choice is made to set the delay line length diﬀerence so the path diﬀerence is zero for the phase centre s0 · B + p = 0.
(8.16)
δ = [ s − s0 ] · B = ∆s · B.
(8.17)
Thus Now the angle that ∆s subtends on the sky is ∆s/s0 , which is simply ∆s because s0 is a unit vector. ∆s is small and can therefore be resolved into component angles in the ORx plane and m in the ORy plane with ∆s2 = 2 +m2 . In order to simplify the notation kB is similarly resolved into components u and v along the Ox and Oy axes respectively. Then the relative phase between light arriving along the two paths becomes kδ = u + mv + φ,
(8.18)
where φ is any small phase introduced by deliberately altering the difference between the lengths of the delay lines from the condition of zero path diﬀerence at the phase centre. Thus the intensity at the detector becomes P (u, v; φ) = I( , m) [ 1 + cos ( u + mv + φ) ] d dm = P0 + cos φ I( , m) cos ( u + mv)d dm − sin φ I( , m) sin ( u + mv)d dm, (8.19) where P0 = I( , m)d dm is the total intensity. The other two integrals in this equation are components of the Fourier transform V (u, v) = I( , m) exp [ i( u + mv) ]d dm. (8.20)
Aperture synthesis 213
214 Astronomical telescopes
This transform is special: it converts the source brightness angular distribution into a distribution in u and v, that is into an intensity distribution in spatial frequency along the Ox and Oy axes respectively. Rewriting the detected intensity in terms of V (u, v) gives P (u, v; φ) = P0 {1 + Re [ V (u, v) exp (−iφ) ] }.
(8.21)
In the case of the pupil plane interferometer Re [ V (u, v) exp (−iφ) ] is the modulation observed when the delay line length is dithered. The very same modulation is seen across the image plane of the image plane interferometer. If the detected power is measured separately for φ set to zero in one case and φ set to π/2 in the other, the respective power values will be P0 (1 + ReV ) and P0 (1 + ImV ). This gives enough information to be able extract the phase and magnitude of V . The maximum and minimum of intensity occur when Re [ V (u, v) exp (−iφ) ] = ±V (u, v) and hence the visibility of fringes (Pmax − Pmin ) = V (u, v) (Pmax + Pmin ) which shows that V (u, v) is simply the visibility deﬁned in eqn. 5.14. Therefore V (u, v) is given the name complex visibility. It is important to keep ﬁrmly in mind that the visibility amplitude is not the amplitude of the fringes but rather the amplitude of their contrast. Putting eqn. 8.20 into words: the complex visibility of the fringe pattern is the Fourier transform of the source brightness distribution.6 We have seen that the visibility V (u, v) is the Fourier transform of the source brightness I( , m) at a single point in the the (u,v) plane. What is needed to reconstruct the source distribution fully is a series of measurements of the visibility for diﬀerent baseline lengths and orientations, giving a range of values of kB and hence of (u,v). Then if the distribution of these measurements is dense enough in the (u,v) space the following approximation can be made I( , m) = V (u, v) exp i( u + mv) du dv ≈ V (u, v) exp i( u + mv)∆u∆v, (8.22) where the sum is taken over all the measurements and where ∆u, ∆v are the spacings between the measurements. Evidently both the phase and magnitude of the visibility have to be well determined if the image reconstruction is to be reliable. 6 This statement is one version of the van Cittert–Zernike theorem. See for example M. Born and E. Wolf, Principles of Optics, seventh edition, published by Cambridge University Press, (1999).
8.8
Aperture synthesis 215
For clarity in presentation of the analysis of stellar interferometers it has been assumed, thus far, that the starlight is monochromatic. In practice, in order that the fringes are bright enough to give reliable measurements, the spread of wavelengths used can be 10% of the mean wavelength. Thus the fringes of the diﬀerent colours will only be superposed around the location where the optical path diﬀerence is zero, which is located at image of the phase centre on the target. Only here will the fringe contrast accurately reﬂect the true visibility. It is essential therefore to make measurements on the visibility amplitude close to the central white fringe. Also the displacement of the white fringe from the image of the phase centre determines the phase of the visibility: a displacement of f fringe widths gives a phase of 2πf . Note that the number of clear fringes is much fewer than the number, B/d, expected with a monochromatic source: for a bandwidth of 10% there will be roughly 10 such fringes. When the source is centrosymmetric so that I( , m) is symmetric around = 0 and m = 0, examination of eqn. 8.20 shows that the complex visibility is always real. In this simple case all that is required are measurements of the visibility amplitude. In addition when the source is centrosymmetric the measurements only need to be made for one baseline orientation to ﬁx the source intensity distribution. A simple example is a circular source of uniform intensity for which the calculation of Section 6.7 can be reused. With the notation of ﬁgure 8.18 an element of the source at (ρ, α) covers a solid angle d dm = sin ρ dρdα = ρ dα dρ to a good approximation, and B · ∆s = Bρ cos α. Then θ 2π V = ∆I exp (ikBρ cos α) ρ dρ dα, 0
where ∆I is the intensity per unit solid angle of the source and θ is its angular radius. Thus θ V = 2π ∆I J0 (kBρ) ρ dρ 0
= 2π θ2 ∆I J1 (kBθ)/(kBθ) = 2I J1 (kBθ)/(kBθ),
(8.23)
1
Source
0.8 Visibility
0
α
0.6 0.4 0.2
ρ Telescope
0
where I is the total intensity of the source. Figure 8.18 also shows this distribution for a source of 20 mas angular diameter observed at a wavelength of 2.2 µm. When the source is more complex in shape the measurement of the phase of the visibility is also required. The eﬀect of the piston component of the atmospheric distortion has been mentioned above. If the interferometer consists of a pair of telescopes there is the lag/lead in phase between them produced by changes
0 10 20 30 40 50 60 70 Baseline in metres Fig. 8.18 Visibility versus baseline for a 20 mas source observed at a wavelength of 2.2 µm.
216 Astronomical telescopes
in refractive index across the whole area of either telescope. Electronic systems are used to track these fringe movements and provide compensating delay line correction. However it is not possible to separate the displacement of the white fringe due to the visibility phase from the much larger eﬀect due to the path length changes in the atmosphere. Thus each determination of the phase of the visibility contains an unknown error, larger than 2π. The phase can be recovered if a suitable centrosymmetric reference star lies within the isoplanatic region around the target by relating the positions of the target fringe to the reference star’s image. In other cases phase recovery requires the techniques described in the sections following this.
8.9
Aperture arrays
The apertures used to collect the interfering beams can be mirrors themselves in altaz or equatorial mountings which return the light to a ﬁxed telescope, as in the COAST array at Cambridge in the UK, or they can be individual telescopes. In the case of COAST the mirrors are moved to various stations along three tracks forming a Yshape of about 65 m extent. In this way baselines with a variety of lengths and orientations can be readily obtained. When the telescopes are stationary the rotation of the Earth provides the means to change the orientation of the baselines.
Fig. 8.19 Mask used for aperture synthesis on the Keck I telescope projected onto the primary mirror. The hexagons are the 36 segments from which the 10 m diameter mirror is made.
An alternative approach in interferometer imaging has been widely used. A plate with many identical circular holes across its surface is placed at the exit pupil plane of a large telescope. The individual holes function as separate telescopes and interference between all these pairs of telescopes is then simultaneously present in the image plane. If the orientations and lengths of the baselines formed by the pairs of apertures are all diﬀerent then fringe patterns for an each aperture pair have a unique and known fringe spacing and orientation. This permits the fringe patterns to be disentangled using Fourier analysis. With the aperture diameters being smaller than r0 and the exposures being shorter than τ0 a uniform distortion is frozen over each aperture. Frequent exposures are made to increase the overall signal relative to noise from electronics and scattered light. Figure 8.19 shows the aperture pattern used by Tuthill and colleagues7 with the segmented Keck I telescope, projected back onto the primary mirror. The 21 apertures are placed so as to avoid any of the 210 baselines being the same as any other baseline, which makes these 210 fringe patterns distinguishable from one another because each has a known unique fringe orientation and fringe spacing. As noted earlier, it is essential to measure the phase accurately as well as the amplitude of the complex visibility in order to be able to 7 P.G. Tuthill, J.D. Monnier, W.C. Danchi, E.H. Wishnow and C.A. Haniﬀ: Publications of the Astronomical Society of the Paciﬁc 112, 555 (2000).
8.10
reconstruct the source distribution. Of these the phase measurement is the most troublesome because the path diﬀerence is continually changing by tens of microns thanks to the piston component of the atmospheric distortion. If there is a bright unresolved star within the isoplanatic cone around the target of interest its white fringe can be used to give a phase reference. When no reference is available the following less complete solution is adopted. Visibilities are measured for each baseline provided by a triangle of apertures. The measured phases of the visibility function are then φ12 = φ012 + φ1 − φ2 , φ23 = φ023 + φ2 − φ3 , φ31 = φ031 + φ3 − φ1 . Here φ0ij is the undistorted phase for the pair of apertures (i,j). The measured phase, φij , is this true phase altered by the distortions at the two apertures involved: φi and φj . Adding these three equations gives the phase closure relation φ12 + φ23 + φ31 = φ012 + φ023 + φ031 .
(8.24)
This sum recovers the sum of the three actual phases. When there are many apertures the phase closure relations from all independent sets of three apertures provide almost as many equations as there are phases. Only a small amount of external information is then required to fully reconstruct the image. Measurements on each short exposure taken will yield diﬀerent values of the visibility phases but give, within errors, the same closure phases. On the other hand the magnitude of the visibility will be the same within experimental errors from one exposure to the next. It is important to appreciate that, in general, the visibility phases and the image can be simultaneously reconstructed from the closure phases and visibility amplitudes.
8.10
Image recovery
Once the visibilities are available from enough baselines an attempt can be made to recover the image using eqn. 8.22. However, as is now shown, this image is imprinted with the distribution of the baselines in (u,v) space. Suppose W (u, v) is the Fourier transform of the true source distribution I( , m). Then the image recovered is R( , m) = W (u, v)g(u, v) exp [ i( u + mv) ]dudv, (8.25) where g(u, v) is unity at each point a measurement was made and zero elsewhere. Thus R( , m) = I( , m) ⊗ G( , m), (8.26)
where G( , m) =
g(u, v) exp [ i( u + mv) ]dudv
Image recovery 217
218 Astronomical telescopes
is the Fourier transform of the distribution of the baselines. G( , m) is the diﬀraction pattern produced on the (l, m) plane by the sampling in the (u, v) plane. It has complicated lobes which smear the actual source distribution into the reconstructed image R( , m). Various algorithms have been invented to try to extract the actual source distribution I( , m) from R( , m). A commonly used example called CLEAN will be described here.8 The procedure is to pick out the brightest cell in the image located, let us suppose, at ( 1 , m1 ) with intensity R1 . Then a distribution γG( − 1 , m − m1 )R1 is subtracted from the image brightness distribution where γ is a constant of around 0.5 called the loop gain. This step removes what was the brightest region of the source taking account of the known smearing of the aperture pattern. This step is then repeated to remove the next brightest cell, and
CIT 6 (Dec97)
WR 104 (Jun98)
100 Milliarcseconds
50 0 N
50 100
E 100
50 0 50 Milliarcseconds
100
100
50 0 50 Milliarcseconds
100
Fig. 8.20 Images of an evolved carbon star CIT6 and a Wolf–Rayet star WR104, obtained with aperture synthesis at the Keck I telescope. This ﬁgure originally appeared in the Publications of the Astronomical Society of the Paciﬁc (P.G Tuthill, J.D Monnier, W.C Danchi, E.H Wishnow and C.A Haniﬀ; PASP: 112, 555 (2000)) Copyright 2000, the Astronomical Society of the Paciﬁc; reproduced with the permission of the editors, and by courtesy of Professor Tuthill.
so on. Eventually the residual image after these successive subtractions will be of nearuniform low intensity containing only background light and detector noise. Finally in order to recover the best estimate of the actual image the ‘sources’ at ( i , mi ) with intensities γR( i , mi ) are convoluted with the point spread function (PSF) of the full telescope aperture. In practice γ needs to be tuned for each set of observations. Images obtained by Tuthill and colleagues from the Keck I telescope with a 15 hole mask using a related analysis technique are shown in ﬁgure 8.20.
8 J.A.
Hogbom, Astronomy and Astrophysics Supplement 15, 417 (1947).
8.11
Comparisons with radioastronomy 219
It is perhaps surprising that the masking technique is so successful in producing a diﬀraction limited images after throwing away 90% or more of the incident light. However by restricting each aperture to the size of the Fried parameter and by making short exposures the temporal and spatial variation of the atmospheric distortion is eliminated from each exposure, and as a result fringes with good signal to noise are obtained. From such fringes obtained with multiple apertures the analysis described here can simultaneously eliminate the atmospheric distortion and reconstruct the image. Exposures with the full aperture simply superpose the images from all the area elements of the lens with their diﬀerent and time varying distortions.9
8.11
Comparisons with radioastronomy
The resolution obtained with an interferometer of baseline length B at wavelength λ is λ/B. By chance the best resolution recently achieved with optical interferometers is very similar to that of current radiotelescopes. The diﬀerence in wavelengths (∼1 µm and ∼10 cm) is compensated by the diﬀerence in baseline lengths (∼100 m and ∼104 km). However radio interferometry and optical interferometry diﬀer critically in the way radio and light waves are detected. Detectors of light respond to the intensity of the radiation, which is proportional to the electric ﬁeld squared EE ∗ . On the other hand detectors of radio waves are electric circuits in which the current is proportional to the electric ﬁeld. The radio signal from one telescope can therefore be detected, ampliﬁed and even recorded without losing any of its phase content. This ampliﬁed signal (with a universal time marker) is then transmitted from its parent radiotelescope site to a distant location where it can interfere with another telescope’s signal. In the contrasting case of optical interferometers coherent ampliﬁcation with a phase reference is feasible but would degrade the signal to noise appreciably. The light from both telescopes must be transmitted directly along optical paths and brought to interfere. In addition the lengths of these paths must be measured to a fraction of the much smaller optical wavelength. The development of the technology to measure distances of order 100 m to a precisions of 10 nm, and to move mirrors to a comparable precision at speeds of ms−1 is relatively recent. The lasers required in the measuring process have themselves to be stable to an equivalent precision, namely better than one part in 1010 . Radio interferometry was implemented earlier from a simpler technological base and has accumulated vastly more data. The problems of phase and image recovery in the presence of atmospheric distortion are common to both types of interferometry. Consequently many analysis techniques, such as phase closure, which were originally developed for the interpretation of radio 9 I am indebted to Professors Tuthill and Haniﬀ for clariﬁcation on a number of issues related to masking.
220 Astronomical telescopes
data have been reused in analysing optical data.
8.12 0, τ
τ /4
τ /2
3 τ /4
Fig. 8.21 Distortion of a circle of test masses produced by gravitational waves travelling perpendicular to the diagram.
Gravitational wave detectors
Einstein, on the basis of his general theory of relativity, predicted the existence of gravitational waves. Any nonsymmetric accelerating mass will emit gravitational radiation, but a perfectly spherical star collapsing radially would not radiate. The only experimental evidence is as yet indirect: this comes from measurements of the slow changes in the orbital period of a binary pulsar PSR1913+16. The period is changing at just the rate expected if the system is radiating energy in the form of gravitational waves, and for the measurement of this eﬀect Hulse and Taylor were awarded the 1993 Nobel Prize in Physics. Gravitational waves travel at the speed of light and distort spacetime as they travel though it. The simplest gravitational wave excites quadrapole oscillations of spacetime, and this motion is illustrated in ﬁgure 8.21. A circle of free masses, initially at rest, is shown at time intervals of one quarter period as a plane gravitation wave passes. In one half cycle the ring is stretched in one direction and squashed at right angles; in the next half cycle the distortion is reversed. If the arms of a Michelson interferometer are aligned along the dotted lines then the relative lengths of the arms would change at the frequency of the gravitational wave, and the interference fringes would oscillate to and fro in synchronism. Of course if the wavefronts are not parallel to the plane of the interferometer there would be a reduced eﬀect. The eﬀects of gravitational waves are all weak because the gravitational coupling is itself weak compared to electromagnetism: for example two protons repel one another electrostatically with a force 1036 times stronger than their gravitational attraction. The amplitude of a gravitational wave is expressed as the strain or change in length per unit length of the fabric of spacetime. A supernova explosion occuring at the edge of our galaxy is expected to give a burst of gravitational waves arriving at the Earth lasting for a few milliseconds, with a frequency around 1 kHz, and with a strain of around 10−18 . A pair of merging neutron stars would give a chirp of radiation in which the frequency would rise from 40 Hz to 1 kHz over several seconds with a strain of order 10 −21 . A strain of 10−18 amounts to 10−6 nm in 1 km, and the measurement of such a small quantity is a formidable challenge. This is the ultimate test of interferometry where displacements equal to 10−6 of a fringe must be detected. Nonetheless the current rate of improvement in the sensitivity of huge Michelsontype inteferometers makes it likely that the direct detection of gravitational waves will occur within a decade. It might appear that this precision is unattainable because the displacements are much smaller than the diameter of an atom. However an em wave reﬂected from a surface is reﬂected from all the atoms on
8.12
Gravitational wave detectors 221
that surface, so it is the average position of the atoms that determines the phase of the reﬂected wave. A sketch of the principal features of the current large gravitational wave detectors is shown in ﬁgure 8.22. The volume of the optical paths is enclosed and evacuated in order to eliminate refractive index variations and convection currents. The laser prefered is a stabilized Nd:YAG laser producing a beam of about 10 W at wavelength 1064 nm. This feeds the beam splitter of a Michelson interferometer whose arms are as long as feasible in order to maximize the displacement when a gravitational wave arrives. In the case of the larger LIGO detectors in the USA each arm consists of a 4 km long Fabry–Perot cavity resonant with the laser wavelength, while the GEO detector in Germany has arms 600 m long. Throughout the interferometer the light beam is in the TEM00 mode with its simple Gaussian proﬁle. This makes it possible to use small diameter mirrors despite the great distances involved. The mirrors and beam splitter are freely suspended so that the mirrors act as the test masses for the gravitational wave. They have reﬂection coeﬃcients close to unity so that the light makes many round trips within the cavity. This feature eﬀectively increases the cavity length by a similar factor and hence too the sensitivity. By suitably positioning the beam splitter it is arranged that the returning beams from the cavities interfere in such a way that all the light emerges toward the laser. Toward the photodiode the beams interfere destructively, which is called the dark fringe condition. This arrangement has two advantages. Firstly it is far easier to detect a small oscillation in light level at the photodiode if its quiescent state is dark, than if it were carrying a large current whose ﬂuctuations can mask the signal. Secondly the light emerging toward the laser is reﬂected back into the Michelson by a recycling mirror in front of the laser; this power recycling technique boosts the light level in the cavities by an order of magnitude. Eﬃcient recycling requires that the recycling mirror and the Michelson form another resonant cavity. The diﬀerence between the travel times for N passes in the two arms of an interferometer when the gravitational strain is equal to +S and −S along the arms is ∆t = 4N LS/c, where L is the rest length of each arm. The phase diﬀerence between light emerging after N passes along the arms is ∆φs = 2πc∆t/λ = 8N πLS/λ = 4πScτs /λ
(8.27)
where τs is the storage time. In obtaining this simple result two eﬀects were neglected. Firstly the frequency fg of the gravitational wave may be high enough that the period of one oscillation is shorter that the storage time. In this case there can be a reversal of the strains during the time the light is in the arms which reduces the phase diﬀerence: in the limit of very high frequency gravitational waves there would be no
Cavity Recycle mirror Laser
Cavity
Beam splitter Detector
Fig. 8.22 Layout of a Michelson interferometer for gravitational wave detection.
222 Astronomical telescopes
phase diﬀerence. Secondly, all the light has been assumed to stay in the cavities for the full storage time, whereas it continually arrives and leaves through the entry mirror. τs must be redeﬁned as the time it takes the intensity of radiation in the cavity to fall by a factor e when the laser is turned oﬀ. When both eﬀects are taken into account the phase diﬀerence becomes ∆φ = ∆φs / 1 + (2πfg τs )2 . (8.28) At very high frequencies such that fg τs 1 this reduces to ∆φ = 2cS/(λfg ), and this is called the storage time limit. Current experiments use kilometre long cavities with ﬁnesses of order a few hundred, which means the storage times approach one millisecond. This implies that detectors are only likely to be sensitive to gravitational waves of frequencies up to a few kHz.
Reflected intensity
8.12.1

Frequency off resonance
+
Fig. 8.23 The sign of the change in the reﬂected intensity from a Fabry–Perot cavity when the the cavity length is increased reverses at resonance. This is equally true when changes are made instead in the laser frequency.
Lasercavity locking
It has emerged in the last section that the operation of the interferometer requires that several cavities should resonate with the laser wavelength. The crucial process of searching for and maintaining this condition is called locking the cavity and the laser. In order to understand how this is achieved we ﬁrst go back to ﬁgure 5.31, which shows the intensity transmitted through a Fabry–Perot etalon. As a result the reﬂected intensity must have the shape around resonance that is shown in ﬁgure 8.23. In this ﬁgure the variation of reﬂected intensity is plotted against the frequency of the light or equivalently the cavity length. Above resonance an increase in frequency causes a rise in the reﬂected intensity while below resonance an increase in frequency causes a fall in intensity. This diﬀerence is the basis of the Pound–Drever method for locking the cavity and laser. It is an illustration of how optics and electronics can be integrated to achieve sophisticated and delicate control systems. Figure 8.24 shows a simpliﬁed outline of the equipment used. Light from the laser passes through an optoelectronic modulator across which a voltage is applied from a radiofrequency oscillator at tens of MHz. The eﬀect of applying an oscillatory voltage across the modulator is to cause a small amplitude oscillation in its refractive index. Consequently the light emerging from the modulator has a synchronous oscillating phase. Optical modulators are discussed in more detail in Chapter 11. The electric ﬁeld of light which has passed through such a modulator is E = E0 exp [ i(ωt + β sin (Ωt))],
(8.29)
where ω is the angular frequency of the incident light, Ω is the angular frequency of the applied radio frequency voltage, and β is the very small
8.12
Gravitational wave detectors 223
amplitude of the induced oscillation in phase. This is known as phase modulation. With β small eqn. 8.29 can be approximated as follows E = E0 exp (iωt) [ 1 + iβ sin (Ωt) ] = E0 exp (iωt){1 + (β/2) exp (iΩt) − (β/2) exp (−iΩt)}. (8.30) Thus the beam emerging from the modulator now contains three diﬀerent frequencies. The waves at angular frequency ω are called the carrier waves; and the waves at angular frequencies ω ± Ω are called the upper and lower sidebands. Now suppose the wave described by eqn. 8.30 is reﬂected from a cavity which is near resonance for the carrier. The sidebands will be well oﬀ resonance so their reﬂection coeﬃcients are close to unity and change little with the wavelength. Thus the amplitude of the electric ﬁeld of the reﬂected wave is Er = E0 exp (iωt){rc + iβ sin (Ωt)/2},
(8.31)
OE modulator Laser
I=
= A + B cos (Ωt) +
(E02 β/2)m[ rc
−
rc∗
] sin (Ωt),
(8.33)
The reﬂection coeﬃcient for a cavity is available in eqn. 5.51 rc = r [ exp iδ − 1 ] / [ 1 − r2 exp iδ ], where r is the reﬂection coeﬃcient at either mirror of the etalon and δ is the phase change accumulated by light in going to and fro once in the cavity. Close to resonance δ = 2nπ + with being small so that the reﬂection coeﬃcient of the cavity can be approximated as follows rc = ir/(1 − r2 ) = −rc∗ . Substituting this value for rc into eqn. 8.33 gives V = 2r/(1 − r2 ).
Photodiode
(8.32)
where A and B need not be calculated in detail. As shown in ﬁgure 8.24 the reﬂected light falls on a photodiode detector which produces an electrical signal proportional to the light intensity. This signal is taken to the mixer where it is multiplied by a signal K sin (Ωt) coming directly from the same radiofrequency oscillator that provides the beam modulation; ﬁnally the mixer output is time averaged. As a result only the term in eqn. 8.32 which has the sin (Ωt) variation contributes to this time average. Apart from a multiplicative constant, the ﬁnal output is V = m[ rc − rc∗ ].
Mirror Actuator Cavity
where rc is the reﬂection coeﬃcient of the cavity for the carrier at angular frequency ω. Hence the reﬂected intensity is Er Er∗
Mirror
(8.34)
This output has the desired property that it changes sign with at resonance and can be used as a control signal to reach resonance. This signal drives an actuator to move one of the cavity mirrors (as shown in 8.24), or is used to tune the laser frequency.
RF Oscillator
~
Mixer
Fig. 8.24 Pound–Drever stabilization scheme. This locks the Fabry–Perot cavity length to an integral multiple of half wavelengths of a laser. The solid lines show the light paths while the broken lines carry electronic signals.
224 Astronomical telescopes
8.12.2
Noise sources
Three sources of noise restrict the sensitivity of the interferometers used in gravitational wave searches: seismic, thermal and shot noise. How important these are and how the detectors are designed to reduce the overall noise are now described. There is some variation between the importance of diﬀerent noise sources, depending on the detector designs Seismic noise encompasses ground motion due to Earth tremors, to the wind blowing over the surface and to human activity. This last source can involve movements of tens of microns due to vehicles moving at hundreds of metres distance. The intensity of seismic noise increases as the frequency falls and dominates over other sources below approximately 10 Hz. The mirrors and beam splitter are all suspended from spring loaded supports by pairs of thin wires like pendulums. The supports absorb vertical disturbances and pendulum bobs are insensitive to high frequency horizontal movements of their supports. Fine control of a mirror’s position and damping is made magnetically. Thermal noise is due to the minute thermal vibrations of all the mechanical components in the optical chain. It dominates other noise sources over the midfrequency range around 100 Hz. Each mechanical object has its own natural frequency of vibration and at these frequencies the thermal vibrations can stimulate motion that amounts to a signiﬁcant noise as far as gravitational wave detection is concerned. The pendulums formed by the mirrors and their suspensions swing at around one hertz and the wires vibrate like violin strings at a few hundred hertz. These are sharp resonances and only aﬀect a narrow frequency range. The mirrors themselves can vibrate and are made of fused quartz for which the resonances are also sharp. However at frequencies away from the sharp resonances the mirrors when locked are highly stable.
10
This is a quantum eﬀect caused by the variation of the number of photons at any one time. The topic is discussed in Chapters 14 and 15.
The ﬁnal noise source of importance arises from the natural ﬂuctuations in the intensity of the laser beam.10 These ﬂuctuations can both mask true signals and also aﬀect the stabilization and they form the dominant noise source at high frequencies. In summary the noise is equivalent to an uncertainty in the arm length ∆L = h ¯ cλ/(4πτ P ), (8.35) where ¯h is Planck’s constant/2π, P is the optical power in the cavity and τ the duration of the measurement. Over a 4 km path and with a laser of 1 µm wavelength the detectable strain is √ S = 3.96 10−19/ P τ (8.36) with P in watts and τ in ms. Evidently P must be made as large as possible. This is partly achieved by locking the cavities on resonance with the laser so that the energy in the cavities is large than the laser beam intensity by a factor roughly equal to the cavity ﬁnesse. In addition the
8.13
Seismic 19 Log10 [ S(f)/Hz ]
power recycling increases the light stored in the Fabry–Perot cavities by a further large factor. The aim is to store powers of up to 100 kW in the future. Mirror surfaces have therefore to be smooth enough to avoid scattering or absorbing energy from the beams. The mirrors through which the light enters the cavities have to be highly homogeneous to reduce internal scattering and the consequent heating which changes the refractive index of the glass. The power cannot be increased indeﬁnitely because eventually practical problems arise through the heating of optical components.
21
Thermal
Shot
23 25
The gain in sensitivity brought about by increasing the light stored in the Fabry–Perot cavities has two equivalent interpretations. From one viewpoint we can see that the longer the storage time is, the longer will be the path over which the light travels during the passage of a gravitational wave and hence the higher the sensitivity. From another perspective the increase in light stored reduces the fractional ﬂuctuation in light intensity and this also increases sensitivity.
Gravitational imaging 225
1
10
100
Frequency in Hz
1000
Fig. 8.25 Noise sources of modern kilometre long gravitational wave detectors.
The noise levels achievable with current detectors are notionally indicated in ﬁgure 8.25. Gravitational signals need to produce strains larger than the noise levels shown there in order to be detectable. The value of S is determined by multiplying the ordinate from the graph by the square root of the bandwidth of the signal. In broad terms the gravitational wave detectors are optimized for detection of waves in the audio frequency range 10–5000 Hz.
8.13
Gravitational imaging
The ﬁrst critical test for the general theory of relativity was provided by observations to ﬁnd out whether light passing near to the Sun would be deﬂected by the distortion of spacetime due to its mass. A simple analogy is provided by rolling a ball bearing across a ﬂat, taut, horizontal sheet of rubber. Its path would normally follow a straight line. On the other hand if a heavy weight is placed on the sheet the sheet is bowed down in that region so that the ball bearing would be deﬂected from a straight line path. In 1919 the deviation of light from Mercury as it approached extinction behind the Sun was measured and found to agree with the predicted 1.750 arcsec. Quantatively a ray of light passing at a minimum distance r from a mass M is deﬂected through an angle α = 4GM/(c2 r) = 2Rs /r,
(8.37)
called the Einstein angle. G is the gravitational constant (6.67 10 −11 kg−1 m3 s−2 ) and Rs (2GM/c2 ) is the Schwarzschild radius for a mass M . In the case of the Sun (with mass 1.99 1036 kg and radius 6.69 105 km) this Schwarzschild radius is only 1.4 km. When source and mass are axially symmetric around a line through the observer and all are separated from one another by very large distances the observer may see a circular Einstein ring image of the source extending all round the deﬂecting
Mercury’s image was only visible because there was, by chance, an eclipse at the same time it passed into the Sun’s ‘shadow’. A comparison was made of the position of Mercury against the star ﬁeld on photographs recorded during the eclipse and at another time when the Sun’s image was far oﬀ. The measurement was unfortunately only accurate to around 0.5 arcsec. Radiotelescopes can follow a planet’s radio image very close to the Sun without the help of an eclipse, and they conﬁrm the general relativistic prediction within an experimental error of 0.01 arcsec.
226 Astronomical telescopes
Fig. 8.26 Partial Einstein rings from several sources at varying distances produced by the galaxy Abell 2218. The image was made with Hubble Space Telescope; courtesy Professor R.Ellis (California Institute of Technology), Professor W. Crouch (University of New South Wales) and the NASA Space Telescope.
mass. Sections of Einstein rings can be seen in ﬁgure 8.26 recorded by the HST. The deﬂecting mass is the galaxy Abell 2218 and the arcs are images of sources at various large distances beyond this galaxy. The image does not need to be in colour because the gravitational deﬂection is achromatic! Other eﬀects are also observed when there is good alignment of source, the deﬂecting mass and the Earth. A few quasars have been detected whose high redshifts are suﬃciently large that they must be far enough from the Earth that they ought to be undetectably faint. What happens in such cases is that an intervening galaxy is focusing a cone of rays toward the Earth and in so doing making the quasar appear brighter than it really is. This same focusing eﬀect is relied on to assist in the search for brown dwarf stars. Such stars may contribute to the dark matter in our galaxy, and being dark they are not easy to detect. The technique to detect them is, paradoxically, to monitor continuously the intensities of a large number of stars in a small satellite galaxy of our own galaxy called the Large Magellanic Cloud (LMC). Whenever a brown dwarf in our galaxy passes in front of a star in the LMC the focusing eﬀect can cause the apparent brightness of the star in the LMC to rise by a big factor and then to fall back to its original value after the ‘eclipse’. Similar microlensing events involving a brief change in intensity can be mimicked by astrophysical processes that cause the star in the LMC to heat up temporarily. However such processes aﬀect the distribution of the star’s energy output across the spectrum. This alternative cause of star brightening can therefore be excluded if the star shows an identical and simultaneous rise and fall in intensity for blue and red light separately. Only a few brown dwarf stars have so far been detected by this method so that it is unlikely that collectively they are a major contributor to dark matter.
8.13
Gravitational imaging 227
Exercises (8.1) An unresolved star of apparent magnitude +1.6 is (8.7) According to the Sparrow criterion practically all observed with telescope whose primary has diamethe information on the detailed features of a source ter 2.4 m. The image falls on a detector which has can be retrieved if the detector pixel size is half a response of 1 AW−1 of luminous intensity. What the resolution limit of the telescope. Hence relate charge accumulates on the detector in 1 ms? the optimal f /# of the telescope to the pixel size, p, and the mean wavelength, λ, of the radiation (8.2) Suppose, in conditions of seeing where the Fried detected. Can you suggest any reason for caution parameter is 20 cm, that the distortion is entirely in interpreting the features detected in the image tip/tilt. Calculate the rms variation in the separawhen the samplings across the image are spaced tion between the images of a laser guide star and a more closely than this limit (oversampling)? real guide star in the image plane of a telescope of focal length 30 m. The two stars can be assumed (8.8) A classical Cassegrain telescope has a primary mirto lie in the same isoplanatic cone. ror with focal length 8 m and a secondary for which (8.3) Two unresolved stars of equal intensity have ana is 4 m and e is 1.2. Calculate the separation of gular separation 0.003 arcsec. What separation is the mirrors, the location of the ﬁnal focus and the required for a pair of two telescopes used as an inplate scale. terferometer in order to resolve the stars in light of wavelength 500 nm? In what way is the orientation (8.9) What is the angular ﬁeld of view over which a paraboloidal mirror of 2.4 m diameter and foof the baseline important? cal length 5 m produces an image with less than (8.4) Show that if an astronomical source is centrosym0.5 arcsec distortion? What is the value of the cormetric the complex visibility is always real. responding ﬁeld of view for the Hubble Space Tele(8.5) What power level is required in a 4 km long cavity scope design? in the arm of a gravitational wave interferometer working at an optical wavelength of 1 µm in order (8.10) The Fabry–Perot cavities in a gravitational wave to make it feasible to detect gravitational waves detector have confocal mirrors a distance L apart with strain 10−21 and frequency 1 kHz? Estimate and the radiation stored has wavelength λ. What the storage time if the cavities have a ﬁnesse of 500. is the minimum width, w, of the TEM00 mode at the mirrors? If the mirrors are 4 km apart and the (8.6) Calculate the complex visibility of two equal inlight has wavelength 500 nm what is the value of tensity unresolved sources an angular distance 2θ the corresponding width parameter w0 ? apart.
This page intentionally left blank
Classical electromagnetic theory 9.1
Introduction
In 1864 Maxwell presented a uniﬁed theory of electric and magnetic ﬁelds, which is now encapsulated into four fundamental equations that take his name. These, together with Lorentz’s formula for the force on any charge in an electromagnetic ﬁeld, are the basic elements of classical (that is nonquantum mechanical) electromagnetic theory. In this chapter Maxwell’s equations will ﬁrst be introduced and used to infer, as Maxwell did, that em waves exist and that light is simply one form of electromagnetic radiation. Next, the energy content and energy ﬂow in electromagnetic waves will be discussed. The expression for the energy ﬂow in an electromagnetic ﬁeld will be deduced by applying the law of conservation of energy to the electromagnetic ﬁeld. This completes the formal basis of the classical theory of electromagnetism and provides the starting point for discussing the behaviour of light passing through matter. That discussion takes up the two following chapters and covers polarization, absorption and dispersion eﬀects. Many optically useful materials are dielectrics, that is poor conductors of electricity, and are also only weakly magnetic. The emphasis in this chapter will be on homogeneous and isotropic dielectrics when the ﬁelds in the electromagnetic radiation are suﬃciently small that the response of matter is linear, and the dielectric is stationary. The label used here for such materials in such ﬁelds is HIL. In Chapter 9 the discussion will be extended to anisotropic dielectrics, materials with which light can be manipulated through its polarization. The high electromagnetic ﬁelds in laser beams produce very striking nonlinear eﬀects in materials and these will be discussed after introducing lasers in Chapter 14. Propagation of electromagnetic radiation in metals, insofar as it aﬀects optics, is dealt with in Chapter 11. When the integral versions of Maxwell’s equations are applied at an interface between diﬀerent media they impose simple relationships between the ﬁelds on either side of the surface. In the central portion of this chapter the consequences of these boundary conditions are followed through for electromagnetic waves impinging on interfaces between dielectrics. This analysis not only proves the laws of reﬂection and re
9
230 Classical electromagnetic theory
fraction from ﬁrst principles but also determines the amplitudes and intensities of the reﬂected and refracted waves. The reﬂected and refracted amplitudes are found to depend strongly on the polarization of the incoming waves: if for example unpolarized light is incident on a dielectric/dielectric interface at Brewster’s angle the reﬂected light becomes plane polarized. Another product of the analysis presented here is an explanation of wave behaviour in total internal reﬂection and frustrated total internal reﬂection, a topic broached in Chapter 2. Antireﬂection coatings are essential for the optics of modern cameras because the number of lenses employed is such that most of the light would otherwise be lost in reﬂections. On the other hand, laser safety goggles are required to reﬂect essentially all the incident radiation over a restricted wavelength range. An account is given below of the use of thin multiple layers of dielectrics to form interference ﬁlters which either selectively reﬂect or transmit a range of wavelengths. The last portion of the chapter is used to discuss modes of the electromagnetic ﬁeld and their propagation in simple waveguides.
9.2
Maxwell’s equations
The electric and magnetic ﬁelds, E and B, can be determined directly by measuring the Lorentz force on a charge q travelling with velocity v, namely F = q(E + v ∧ B), (9.1)

and hence they are reasonably regarded as the physical em ﬁelds. The eﬀects these ﬁelds have on matter will now be outlined in order to put Maxwell’s equations in context.
+
Applied E field
Fig. 9.1 Electric ﬁeld lines around an electric dipole.
The action of an applied electric ﬁeld on an atom or molecule is to pull apart the positive charges (nuclei) and negative charges (electrons) so that an electric dipole, p = dq is produced, where q is the total charge of either sign and d is the vector separation of the centre of gravity of the positive charge from the centre of gravity of the negative charge. Electric ﬁelds within an atom are usually much larger than the applied ﬁelds so that the separation of the centres of the positive and negative charges produced by the applied electromagnetic ﬁeld is small in comparison to the size of the atom/molecule. Applying Coulomb’s law, the electric ﬁeld due to a proton or electron at a distance equal to an atomic diameter, 0.1 nm, is predicted to be of order 1011 Vm−1 . For comparison the electric ﬁeld within a continuous 1 kW laser beam is only around 106 Vm−1 , but can be large enough to disrupt atoms in very high power pulsed lasers. In some polar materials the molecules have an intrinsic electric dipole moment and these dipoles would usually point in random directions in the absence of any applied electric ﬁeld. In such polar materials an electric ﬁeld exerts a torque (p ∧ E) on each
9.2
Maxwell’s equations
231
molecular dipole. In liquids and gases, where these dipoles are free to rotate, this torque tends to align them parallel to the electric ﬁeld direction. The alignment of the molecular/atomic dipoles, whether in polar or nonpolar materials, caused by an applied electric ﬁeld is called polarization. Figure 9.1 shows the electric ﬁeld produced by a dipole which indicates that the electric ﬁeld produced by the aligned dipoles in a dielectric opposes the applied ﬁeld. An auxiliary ﬁeld D called the electric displacement is therefore introduced D = ε0 E + P.
(9.2)
P is the polarization of the material, deﬁned as the electric dipole moment per unit volume N p, N being the number density of the electric dipoles. The materials considered in this chapter are homogeneous, isotropic and at suﬃciently low intensity of the radiation so that P is linearly proportional to E, namely P = ε0 χE.
(9.3)
D = ε0 (1 + χ)E = ε0 εr E,
(9.4)
Thus where χ and εr are scalar constants for a given HIL material, and are respectively the electric susceptibility and the relative permittivity.1 An atomic/molecular susceptibility is also deﬁned α = p/(ε0 E) = P/(N ε0 E).
(9.5)
In general magnetic dipoles are induced in any material by an applied magnetic ﬁeld. There is thus an auxiliary magnetic ﬁeld deﬁned by H = B/µr µ0 ,
(9.6)
where µr is called the relative permeability of the material. Equations 9.4 and 9.6 are called the constitutive relations for a material. For the most part the eﬀects of the magnetic dipoles are unimportant at optical frequencies so we can generally take µr = 1 in all that follows. Maxwell’s equations encapsulate the understanding of electromagnetism that had been achieved by Faraday and other experimenters in the early part of the 19th century. The integral forms are D · dS = ρdV , (9.7) V S B · dS = 0, (9.8) S E · dl = − (∂B/∂t) · dS, (9.9) S H · dl = j · dS + (∂D/∂t) · dS, (9.10) S
S
1 The question of what D physically represents is not spelled out anywhere that I have looked. Feynman who possessed a very penetrating insight concluded that it is simply a useful tool.
232 Classical electromagnetic theory
where ρ and j are the free charge density per unit volume and free current density per unit area respectively. In the ﬁrst two equations the surface integrals are over a closed surface and the volume integrals are over the volume enclosed. In the last two equations the line integrals are over a closed path and the surface integrals over any surface spanning that closed loop. The ﬁrst of Maxwell’s equations is known as Gauss’ law of electrostatics and relates the ﬂux of the displacement current through a closed surface to the total charge inside the volume. It is equivalent to Coulomb’s law. The second equation is its magnetic counterpart, and the zero on the right hand side expresses the fact that free magnetic poles (monopoles) have never been detected. The last two equations are Faraday’s law and the Ampere–Maxwell law. Faraday’s law relates the electric potential around a closed loop to the rate of change of magnetic ﬂux through that loop, written here for the case that the loop is stationary. Analogous to this the Ampere–Maxwell law relates the magnetic ﬁeld integrated round a closed loop to the sum of the free and displacement currents through that loop. The diﬀerential forms of the equations are obtained by applying two theorems of vector calculus which hold for any vector Z. These are Gauss’s theorem and Stokes’ theorem, respectively Z · dS = ∇ · Z dV , (9.11) S V Z · dl = (∇ ∧ Z) · dS. (9.12) V
For example applying the latter to Faraday’s law gives (∇ ∧ E) · dS = − (∂B/∂t) · dS, S
which is true for any surface so we have ∇ ∧ E = −∂B/∂t, 2
Maxwell’s equations are consistent with the special theory of relativity. When the observer shifts to a diﬀerent inertial frame the Lorentz coordinate transformations convert Maxwell’s equations to an identical set of equations where the ﬁelds are now those measured in the new frame. It had seemed before Einstein’s discovery of the special theory an odd quirk that the corresponding Galilean transformations of Newtonian mechanics did not reproduce Maxwell’s equations in the new inertial frame. These matters are discussed in Chapter 11 of the third edition of J.D. Jackson’s Classical Electrodynamics published by John Wiley and Sons, New York (1998).
which relates ﬁelds at a single place. The diﬀerential forms of Maxwell’s equations2 are therefore ∇·D=ρ
(9.13)
∇·B=0 ∇ ∧ E = −∂B/∂t
(9.14) (9.15)
∇ ∧ H = j + ∂D/∂t.
(9.16)
In free space in the absence of any charges Maxwell’s equations reduce to ∇·E=0 ∇·B=0 ∇ ∧ E = −∂B/∂t ∇ ∧ B = µ0 ε0 ∂E/∂t.
(9.17) (9.18) (9.19) (9.20)
9.2
Maxwell’s equations
233
Materials, mainly metals, which contain large numbers of electrons that are not bound to individual atoms, but are free to carry current when an external electromagnetic ﬁeld is applied, are good electrical conductors, while those materials which contain very few free electrons are called dielectrics. Electrical conductivity determines the current density through the relation j = σE, (9.21) where the range of values that the conductivity σ in Ω−1 m−1 can take is huge: copper, an excellent conductor, has a conductivity of 6.45 107 Ω−1 m−1 while a dielectric such as glass has a conductivity 10−12 Ω−1 m−1 . In the right hand side of eqn. 9.16 the second term is the displacement current whose size will depend on the polarization of the material and, because of the time diﬀerential, on the frequency. There is such a big diﬀerence between the relative importance of the conduction and displacement currents in conductors on the one hand, and in dielectrics (and free space) on the other hand, that it is helpful to separate the discussion of these two classes. The discussion of the behaviour of waves in dielectrics will begin in this chapter and the discussion of the behaviour of em waves passing through conductors is postponed until Chapter 11.
Medium 1
D1 Charge density Q D2
Medium 2
Boundary conditions It is essential to be able to connect the values of the electromagnetic ﬁelds in one medium to those in another at any interface. Making the connection requires the use of the integral equations and not the diﬀerential forms because the latter refer to ﬁelds at the same place. Figure 9.2 shows projections of a plane surface between materials 1 and 2. The broken lines enclose a volume of pillbox shape which straddles the surface, the ﬂat faces being of unit area and lying a negligible distance from the interface. The actual interface is supposed to carry a surface charge of Q Cm−2 . Then applying Gauss’ law to this volume gives D2n − D1n = Q,
Plan view
Fig. 9.2 Integration pillbox volume straddling an interface between media.
(9.22) Medium 1
where D2n (D1n ) is the component of D perpendicular to the surface in material 2(1). Similarly Gauss’ magnetic law gives B2n − B1n = 0,
(9.23)
relating the normal components of the magnetic ﬁeld. Next consider the broken line closed path shown straddling the interface in ﬁgure 9.3. Here there is assumed to be a surface current density of js A m−1 travelling along the interface, perpendicular to, and into the plane of the diagram. The long arms are of unit length and they are a negligible distance from the interface. Applying the Maxwell–Ampere law to this path gives H1t − H2t = js ,
H1
(9.24)
Surface current σ
Medium 2
H2
Fig. 9.3 Integration loop straddling an interface between media.
234 Classical electromagnetic theory
where H1t (H2t ) is the tangential component of the magnetic ﬁeld H1 (H2 ) at the interface. Applying Faraday’s law to the same circuit gives E1t − E2t = 0.
(9.25)
Consequently when, as is usually the case for dielectrics, there is no surface charge or current, the tangential components of E and H, and the normal components of D and B are all continuous at such an interface.
9.3
The wave equation
Maxwell’s equations lead simply to wave equations for electric and magnetic ﬁelds. Consider the example of free space ﬁrst. Taking the curl of eqn. 9.19, and then using eqn. 9.20 gives ∇ ∧ (∇ ∧ E) = −∂/∂t(∇ ∧ B) = −µ0 ε0 ∂ 2 E/∂t2 .
(9.26)
The identity that is valid for any vector ﬁeld X, ∇ ∧ (∇ ∧ X) = ∇(∇ · X) − ∇2 X,
(9.27)
when applied to E gives ∇ ∧ (∇ ∧ E) = −∇2 E, because of eqn. 9.17. Then eqn. 9.26 can be rewritten ∇2 E = µ0 ε0 ∂ 2 E/∂t2 .
(9.28)
Starting from eqn. 9.20 a similar set of steps gives ∇2 B = µ0 ε0 ∂ 2 B/∂t2 .
(9.29)
These last two equations are wave equations; and it was shown in Chapter 1 that there are plane wave solutions. A suitable sinusoidal plane wave solution is E = E0 exp [i(ωt − k · r)], (9.30) whose real part is the actual electric ﬁeld, with E0 having Cartesian components (E0x ,E0y ,E0z ). One reason for using a complex form is that many mathematical manipulations will be simpler. Thus ∇ · E = ∂Ex /∂x + ∂Ey /∂y + ∂Ez /∂z = {E0x ∂/∂x + E0y ∂/∂y + E0z ∂/∂z}{exp [i(ωt − kx x − ky y − kz z)]} = {−ikxE0x − iky E0y − ikz E0z }{exp [i(ωt − kx x − ky y − kz z)]} = −ik · E. Similarly ∇ ∧ E = −ik ∧ E, ∇2 E = −k 2 E, while ∂E/∂t = iωE, ∂ 2 E/∂t2 = −ω 2 E.
9.3
Substituting the calculated diﬀerentials into eqn. 9.28 or eqn. 9.29 gives k 2 = µo ε 0 ω 2 . Consequently both the wave equations will be satisﬁed provided the wave velocity √ c = ω/k = 1/ µ0 ε0 . (9.31) It was pointed out in Chapter 1 that the right hand side of this equation was found to equal the velocity of light to within the experimental error. This equality established at a stroke that light is one form of electromagnetic radiation. According to the special theory of relativity the velocity of electromagnetic radiation in free space is constant irrespective of the motion of source or observer. The value of c was therefore ﬁxed in 1984, by convention, at the experimental value it had at that time c ≡ 299 792 458 m s−1 .
(9.32)
This leaves the units of length and time to be deﬁned in a way consistent with this requirement. Substituting the complex solution 9.30 into eqns. 9.17, 9.18 and 9.19 gives −ik · E = 0, −ik · B = 0, ˆ ∧ E/c, B = k ∧ E/ω = k
(9.33) (9.34) (9.35)
ˆ is the unit vector k/k. These results justify the statements where k made in Chapter 1 that E, B and k form a righthanded set of orthogonal vectors for electromagnetic waves travelling in free space: the waves are transversely polarized. The preceding analysis requires very few changes when it is extended to study electromagnetic waves in HIL dielectrics. These are materials with high electrical resistance so there are essentially no free charges or currents, but they can be polarized by an externally applied electric ﬁeld. Equations 9.13–9.16 with ρ = j = 0 and µr = 1.0 yield a wave equation for the electric ﬁeld ∇2 E = µ0 ε0 εr ∂ 2 E/∂t2 ,
(9.36)
with a parallel equation for the magnetic ﬁeld. The solutions are taken to be sinusoidal plane waves E = Re E0 exp [i(ωt − k · r)].
(9.37)
and the wave velocity is √ √ v = ω/k = 1/ µ0 ε0 εr = c/ εr .
(9.38)
The wave equation
235
236 Classical electromagnetic theory
The magnetic ﬁeld is ˆ ∧ E/v. B = k ∧ E/ω = k
(9.39)
Substituting the plane wave solution into the ﬁrst two of Maxwell’s equations shows that E, B and k again form an orthogonal set. We can calculate the refractive index, n, from the wave velocity n = c/v =
√ εr .
(9.40)
It is worth noting that the relative permittivity can change a great deal with the wavelength of electromagnetic waves. For example water has a relative permittivity of 80 at low frequencies, while at optical wavelengths the refractive index of water is 1.33. Water molecules have an intrinsic dipole moment, so evidently the time required to get these molecular dipoles to align with the applied ﬁeld is much longer than the period of light oscillations. Glass and other materials used in passive optical components are all HIL dielectrics. A characteristic impedance is deﬁned for any material as the ratio Z = E/H, which gives in the case of dielectrics Z = (µ0 /ε0 εr ) = µ0 c/n. 3
This quantity was previously mentioned in Chapter 1.
(9.41)
(9.42)
In the case of free space the characteristic impedance3 Z0 = 377 Ω. When the values of the ﬁelds in a plane wave are inserted into the expression for the Lorentz force eqn. 9.1, the electric and magnetic contributions are of magnitude qE and qEv/c respectively. Electron velocities in matter are much less than c so that it is generally adequate to ignore the magnetic force in a dielectric. The relative importance of conduction and displacement currents for a sinusoidal plane wave can be inferred by comparing j = σE and ∂D/∂t = −iωε0εr E. The ratio of their magnitudes is Rc/d = σ/ωε0 εr .
(9.43)
For glass the ratio, 0.113/εrω, is small at even low frequencies so it is appropriate to ignore the conduction current for dielectrics. However, in the case of copper this ratio becomes 0.728 1019/ω which remains large up to optical frequencies.
9.3.1
Energy storage and energy ﬂow
The energy stored in a capacitor at constant voltage and in an inductance carrying constant current provide simple examples of energy storage in electric and magnetic ﬁelds respectively. These cases will be considered
9.3
here and the general expression for the total energy stored in any electromagnetic ﬁeld will then be (plausibly) inferred. The energy stored in a capacitor of capacitance C at a voltage V is CV 2 /2. Assuming the capacitor has parallel plates, each of area A and a distance d apart, then ignoring edge eﬀects C = ε0 εr A/d. Also the electric ﬁeld E = V /d. Hence the energy stored in the capacitor is (ε0 εr A/d)(Ed)2 /2 = ε0 εr (Ad)E 2 /2 = E · D(Ad)/2, and the energy stored per unit volume is E · D/2. The energy stored in the magnetic ﬁeld within a solenoid of inductance L carrying a current I is LI 2 /2. Suppose the solenoid has area of crosssection A, length d and carries m turns per unit length. Then L = µ0 m2 Ad, while B = µ0 mI. Consequently the energy stored in the solenoid’s magnetic ﬁeld is, ignoring end eﬀects, [B 2 /µ0 ] (Ad)/2 = B · H(Ad)/2, so that the energy density in the magnetic ﬁeld is B · H/2. Thus the total energy density in an electromagnetic ﬁeld is U = (E · D + B · H)/2.
(9.44)
Energy ﬂow in electromagnetic radiation is a vector quantity N with units Wm−2 : in other words it is the power crossing unit surface area. Maxwell’s equations do not in themselves determine this ﬂow. Poynting realized that the extra ingredient was to apply conservation of energy to a volume of electromagnetic ﬁeld. Take a volume V with surface area S. The energy within V has three components whose changes must balance out to zero. Firstly there is the change in stored energy per unit time ∂/∂t( V U dV ). Secondly there is the outward ﬂow of energy per unit time through the whole surface S N · dS. This can be converted to a volume integral using Gauss divergence theorem, giving V ∇ · NdV . Thirdly there is the work done per unit time on whatever charges are enclosed in the volume V E · jdV , where j is the current density. In order that energy is conserved the total of these three contributions should be zero. Thus ∂/∂t U dV + ∇ · NdV + E · jdV = 0. V
V
V
This balance must be true for any volume so that ∂U/∂t + ∇ · N + E · j = 0.
(9.45)
We now need to use Maxwell’s equations for media. The scalar product of E with eqn. 9.16 gives E · j = E · (∇ ∧ H − ∂D/∂t).
(9.46)
The wave equation
237
238 Classical electromagnetic theory
Next we apply an identity valid for any pair of vectors to E and H E · (∇ ∧ H) = ∇ · (H ∧ E) + H · (∇ ∧ E) = ∇ · (H ∧ E) − H · ∂B/∂t, where eqn. 9.15 was used in replacing the second term on the right hand side. Substituting this result in eqn. 9.46 gives E · j = ∇ · (H ∧ E) − E · (∂D/∂t) − H · (∂B/∂t). Rearranging this equation with the help of eqn. 9.44 ∂U/∂t + ∇ · (E ∧ H) + E · j = 0.
(9.47)
Comparing eqns. 9.47 with 9.45 allows us to identify the energy ﬂow vector as N = E ∧ H, (9.48) which is known as the Poynting vector. In the case of a sinusoidal plane wave the actual (real) ﬁelds have Cartesian components Ex = E0 cos (ωt − kz) By = (E0 /v) cos (ωt − kz) which travel at velocity v and the magnitude of the Poynting vector is N = [E02 /(µ0 v)] cos2 (ωt − kz), Fig. 9.4 Electric ﬁeld lines: in the upper panel around a charge at rest and in the lower panel around a charge moving at constant velocity.
along the zdirection. The time average of the Poynting vector taken over many cycles of the electromagnetic wave is N = E02 /(2µ0 v) = E02 /2Z.
(9.49)
Similarly the energy density is 1 U = (ε0 εr E02 + E02 /µ0 v 2 ) cos2 (ωt − kz) 2 = (E02 /µ0 v 2 ) cos2 (ωt − kz),
(9.50)
with a time average U = E02 /(2µ0 v 2 ).
(9.51)
The time averaged energy density and energy ﬂow are necessarily very closely related; from eqns. 9.49 and 9.51 N = vU .
(9.52)
Electromagnetic ﬁelds also carry momentum as well as energy and it is the momentum of the radiation from the Sun that deﬂects comets’ tails so that they point radially away from the Sun. This momentum can be determined by considering the reﬂection of a plane electromagnetic wave at normal incidence from the ﬂat surface of a perfect conductor. The argument is only sketched here, but a rigorous proof can be found
9.4
Electromagnetic radiation 239
in Bleaney and Bleaney.4 The magnetic ﬁeld H is parallel to the surface in the dielectric but vanishes in the conductor so that the boundary condition eqn. 9.24 reduces to H = js . Using eqn. 9.1 the radiation pressure on the surface is ˆ P = js ∧ B = BH k,
(9.53)
ˆ is a unit vector perpendicular to the surface. Now the time where k average of the energy density in the incident waves is U = BH = ED. Thus the relation between the radiation pressure and the energy ﬂux in a plane electromagnetic wave is P = N/v.
9.4
c(t t0)
A B
(9.54)
Electromagnetic radiation
The electric ﬁeld of a plane electromagnetic wave, being transverse, is very diﬀerent from the electric ﬁelds met in electrostatics. The upper panel of ﬁgure 9.4 shows the radial electric ﬁeld of an isolated static electric charge. If the same charge is moving with constant velocity the ﬁeld lines are compressed in the direction of motion as shown in the lower panel; but they remain radial. What is needed to generate travelling waves with transverse ﬁelds is for the charge to accelerate. Figure 9.5 pictures a simple example: the electric ﬁeld lines are drawn for a charge that is at rest at A up till time t0 , it then moves to B and is again at rest there from time t1 . Before t0 the lines point back to A, and after t1 they point back to B. Between these times there are kinks in the ﬁeld lines which give the electric ﬁeld a transverse component. As time passes the kinks move continuously away from the charge with velocity c in free space. If the charge is instead made to oscillate between A and B the ﬁeld’s transverse component will alternate in direction giving waves with transverse ﬁelds oscillating at the frequency of the charge’s motion. Over an area of a wavefront whose dimensions are small compared to its distance from the source the wavefront approximates to a plane wave. The simplest source of radiation is an oscillating electric dipole. The motion is simple harmonic with the charges oscillating about their common centre. Referring to ﬁgure 9.1 the charges exchange positions every half cycle of the oscillation. As mentioned above an applied electromagnetic ﬁeld acting on matter induces and aligns atomic and molecular electric dipoles. An excited atom can be classically pictured as one which has absorbed electromagnetic radiation and become a dipole in which the electron is oscillating about the much heavier nucleus. Such 4 Chapter 8 of the ﬁfth edition of Electricity and Magnetism by B. and B. I. Bleaney, published by Oxford University Press (1983).
c(tt1)
Fig. 9.5 Electric ﬁeld lines from a charge at rest at A until time t0 ; it then moves rapidly and comes to rest at time t1 at B. The lines are drawn as they appear at a later time t.
Z r φ θ
θ Y φ X Fig. 9.6 Local spherical polar coordinates with the dipole axis as the polar axis θ = 0. The r, θ and φ local axes point in the directions of increasing r, θ and φ respectively.
240 Classical electromagnetic theory
atomic excitations have frequencies ranging from the near infrared to the ultraviolet. Molecular excitations, which involve the vibration and oscillation of the nuclei, have frequencies in the near to far infrared region of the spectrum. The radiation from the dipole carries oﬀ energy and so the oscillation dies away: the radiation damps the dipole motion. In this classical view the excited atom loses its excitation energy by radiating at its natural frequency. At points very close to an electric dipole the ﬁeld pattern is quite complex, while in the radiation zone at distances large compared to the wavelength the ﬁelds have a simple form. Using the local axes shown in ﬁgure 9.6 the ﬁeld components in the radiation zone are −ω 2 p0 sin θ cos (ωt − kr), 4πε0 c2 r Bφ = Eθ /c, Eθ =
(9.55) (9.56)
where p0 cos ωt is the oscillating dipole moment and ω/k = c. The power radiated is given by eqn. 9.48.The energy crossing an element dS of a
Fig. 9.7 A polar plot of the intensity distribution of radiation from a dipole in the radiation zone. The dipole direction is shown by the central arrow. The directions of the ﬁelds and the Poynting vector are also shown at a representative point. Lines of H are tangential to the dotted line circle.
sphere centred on the dipole per unit time is N · dS = E ∧ H · dS ∝ sin2 θdS.
(9.57)
This angular distribution of the radiated power is displayed in a polar diagram in ﬁgure 9.7, where the distance from the origin to the shaded
9.5
surface is proportional to sin2 θ. The intensity of radiation falls oﬀ toward the axis and is zero exactly along the axis, which agrees with what is seen in ﬁgure 9.5: that there is no transverse component of the electric ﬁeld along the axis through AB. Integrating over all directions gives the total power radiated: 2π π W = N · dS = Eθ Hφ r2 sin θdθdφ 0 0 2π π 4 2 2 3 = [(ω p0 )/(16π ε0 c )] cos2 (ωt − kr) dφ sin3 θdθ 0
0
= [(ω 4 p20 )/(6πε0 c3 )] cos2 (ωt − kr). Averaging over time gives W = ω 4 p20 /(12πε0 c3 )
(9.58)
per dipole. More complicated charge distributions than dipoles are possible in an atom, such as a quadrupole which consists of a pair of dipoles oppositely aligned. Quadrupole oscillations occur and of course radiate. However the radiation from quadrupoles and other multipoles is weak compared to dipole radiation when the radiating structure is much smaller than the wavelength of the radiation – which is the case for atoms radiating light. The dominance of dipole over other more complex radiation persists in quantum theory and will be discussed later. On a larger scale the antennae of radio stations are true classical electric dipoles: they are in the form of conducting wires in which alternating currents ﬂow.
9.5
Reﬂection and refraction
Electromagnetic theory is now applied to the behaviour of electromagnetic waves at interfaces between HIL dielectrics of diﬀerent refractive indices n1 and n2 . Figure 9.8 shows the wave vectors (k) and ﬁelds at a plane interface between two dielectrics when a plane sinusoidal electromagnetic wave is incident with its electric ﬁeld transverse to the plane of incidence. This is called a transverse electric (TE) wave, or alternatively an spolarized wave. The subscripts i, r and t refer to the incident, reﬂected and transmitted plane waves respectively. Both the materials are HIL so that the electric ﬁeld in the reﬂected and refracted waves will also be transverse to the plane of incidence. There are no free charges or currents. The alternative case when the magnetic ﬁelds are perpendicular to the plane of incidence is shown in ﬁgure 9.9 and bears the labels transverse magnetic (TM), or ppolarization. All the results obtained in the present section apply equally to p as well as to spolarized light. In the following section the case of spolarization is calculated in detail, while for the case of ppolarization only results are stated. The axes are oriented as shown in ﬁgure 9.8 with the surface being the plane z = 0, and with x = 0 being the plane of incidence. Along
Reﬂection and refraction
241
242 Classical electromagnetic theory
kr Ei
Er ki
Bi
Br
Dielectric 1
θi
θr Y
Dielectric 2
θt
Et Bt Z
kt
Fig. 9.8 Incident, reﬂected and transmitted wave vectors at a plane interface between two dielectrics for the case that the electric ﬁeld is perpendicular to the plane of incidence: spolarization.
kr
Bi
Br
ki Ei
Er Dielectric 1
θi θr
Dielectric 2
Y
θt
Z
Et
the yaxis, which lies in the surface, the plane waves reduce to the form E = E0 exp [i(ωt − ky y)], where (kx ,ky ,kz ) are the components of the wave vector along the axes. The boundary conditions derived above state that, provided there are no surface charges or currents, the components of E and H transverse to the surface and the components of D and B perpendicular to the surface are all continuous at the surface. This means, for example, that the transverse component of E just above the surface in the ﬁrst medium is identical to the transverse component of E just below the surface in the second medium. This continuity of the transverse component of the electric ﬁeld requires that E0i exp [i(ωi t − kiy y)] + E0r exp [i(ωr t − kry y)] = E0t exp [i(ωt t − kty y)].
(9.59)
If the ﬁelds are to match in this way at the surface for all times, it follows that ωr = ωi = ωt ,
Bt kt
Fig. 9.9 Incident, reﬂected and transmitted wave vectors at a plane interface between two dielectrics for the case that the electric ﬁeld lies in the plane of incidence: ppolarization.
which for simplicity we write ω. The above equation can be rewritten v1 kr = v1 ki = v2 kt ,
(9.60)
where v1 and v2 are the velocities of light in the two dielectrics. The ﬁrst equality simply tells that the reﬂected wave has the same wave
9.5
number and wavelength as the incident wave. The second equality can be rewritten again as n1 λi = n2 λt , (9.61) where n1 and n2 are the two refractive indices. Thus it is the wavelength that changes from material to material while the frequency remains the same. The ﬁeld values must equally match across the surface any point, that is at each and every value of y, hence kry = kiy = kty ,
(9.62)
which we write simply as ky . Expanding the results in eqn. 9.62 gives kr sin θr = ki sin θi , kt sin θt = ki sin θi .
(9.63)
Remembering that kr = ki the ﬁrst of these two equalities is just the law of reﬂection θr = θi . The second equality simpliﬁes to Snell’s law n1 sin θi = n2 sin θt . It is worth noting that these laws simply amount to the statement that the component of the wave vector parallel to the surface is unchanged by reﬂection or refraction. In the case of light in an optical ﬁbre ky is known as the propogation vector, the component of the wave vector along the ﬁbre axis. This analysis also provides an explanation of what happens in the process of total internal reﬂection (TIR) met earlier in Section 2.2. Then n1 > n2 and we choose an angle of incidence greater than the critical angle θc , so that sin θi > n2 /n1 . Thus eqn. 9.62 gives ky = ki sin θi = (n1 /n2 )kt sin θi > kt . It follows that the component of the refracted wave vector perpendicular to the surface, ktz , is imaginary, 2 2 ktz = kt2 − kty < 0.
(9.64)
Hence ktz = ±iκ where κ is real and positive. Here the negative sign must be taken because the positive sign gives a wave whose amplitude increases exponentially in the less dense medium, which is physically unreasonable. The transmitted wave is thus Et = E0t exp (−κz) exp [i(ωt − ky y)].
(9.65)
This is an evanescent wave: it travels parallel to the interface, while its amplitude and intensity fall oﬀ exponentially with the distance from the surface. Its intensity N = N0 exp (−2κz) (9.66) drops by a factor e over a distance 1/2κ. If the incident wave is spolarized then the magnetic ﬁeld B = k ∧ E/ω = −iκEx ey /ω − ky Ex ez /ω,
Reﬂection and refraction
243
244 Classical electromagnetic theory
which has a component along the ydirection and hence the evanescent wave is not a transverse wave. Similarly when the incident wave is ppolarized the electric ﬁeld of the evanescent wave has a component along the direction of travel. In optical ﬁbre an evanescent wave travels parallel to the ﬁbre axis within the optically less dense cladding which surrounds the optically denser core. We can investigate qualitatively the fraction of total power travelling in the evanescent wave in single mode ﬁbres. For values of θi above the critical angle κ2 = ky2 − kt2 = kt2 [(n1 /n2 )2 sin2 θi − 1],
(9.67)
and if θ is close to 90◦ κ2 ≈ kt2 (n21 − n22 )/n22 . Inserting into this equation the parameters for Corning SMF28 TM monomode optical ﬁbre, namely ncore = 1.4677, ncladding = 1.4624, at a wavelength λ = 1.310 µm, gives κ2 ≈ 0.3572.
Lens
Glass block
Fig. 9.10 Arrangement for observing the eﬀects of frustrated total internal reﬂection (FTIR).
Then κ is 0.6 µm−1 and the intensity falls by a factor e2 in a distance of around 1.67 µm. Now the core radius is only 4.1 µm, whence it follows that the crosssectional area of the cladding within this 1/e2 zone is π(5.772 − 4.12 ), that is 51.8 µm2 , which is comparable to the area of the core itself, 52.8 µm2 . Consequently a signiﬁcant fraction of the electromagnetic radiation through a monomode ﬁbre travels within the cladding. Figure 9.10 illustrates a simple way in which to observe the eﬀect of frustrated total internal reﬂection (FTIR). A planoconvex lens of radius of curvature 50 to 100 cm is placed on an optical ﬂat. Viewing as shown in the diagram at an angle such that the angle of incidence at the glass block/air interface is signiﬁcantly larger than the critical angle a dark patch will be visible at the point of contact. Light in the evanescent wave penetrates into the gap and is then reﬂected at the lens surface. Then because there is a phase diﬀerence of π between the reﬂections from air/glass and a glass/air interfaces the two reﬂections interfere destructively at the point of contact. Applying eqn. 9.67 for light of 500 nm incident at 45◦ on a glass/air interface, the penetration depth is only 80 nm. Thus the penetration depth is much less than the separation between the lens and the glass block at which the ﬁrst of Newton’s rings appear and therefore none will be seen. At Xray wavelengths the refractive index of metals is slightly less than unity, which means that Xrays incident in free space at close to grazing incidence on a metal are totally internally reﬂected. The design of the XMMNewton Xray telescope, launched into space in 1999, makes use
9.6
Fresnel’s equations 245
Focus
Paraboloidal mirror
Hyperboloidal mirror
Fig. 9.11 XMMNewton Xray telescope using TIR. The degree of focusing is exaggerated: the actual focal length is about ten times the combined length of the mirrors.
of this property. It consists of cylindrically symmetric thin alumininium mirrors with a 250 nm gold coating on their internal surfaces. An axial section of one mirror pair is pictured in ﬁgure 9.11; the ﬁrst mirror is a shallow paraboloid, and the second a shallow hyperboloid. Together they reﬂect and focus the Xrays from any source located close to the mirrors’ axis. One such mirror pair would not collect many Xrays so 58 coaxial mirror pairs of graded diameters are arranged concentrically around one another. The outermost mirror has diameter 70 cm, and the innermost 31 cm diameter; the total length of the mirror assembly is 60 cm and the focal length measured from the paraboloid/hyperboloid junction is 7.5 m.5
9.6
Fresnel’s equations
The analysis of reﬂection and refraction using electromagnetic theory is now continued in order to obtain the complex amplitudes of the reﬂected and refracted waves at the interface considered in ﬁgure 9.8. The exponential factors in equation 9.59 were shown to be equal so that the condition for continuity at the interface reduces to E0i + E0r = E0t .
(9.68)
Correspondingly, continuity of the tangential component of the magnetic ﬁeld across the interface in ﬁgure 9.8 for spolarization requires that H0i cos θi − H0r cos θi = H0t cos θt .
(9.69)
Making the substitution H = En/(µ0 c) from eqn. 9.42 this becomes n1 E0i cos θi − n1 E0r cos θi = n2 E0t cos θt .
(9.70)
Equations 9.68 and 9.70 can be solved simultaneously to give rs = E0r /E0i = (n1 cos θi − n2 cos θt )/(n1 cos θi + n2 cos θt ), (9.71) (9.72) ts = E0t /E0i = 2n1 cos θi /(n1 cos θi + n2 cos θt ).
5
In the following chapter the situation in which the refractive index can be less than unity will be discussed further. For the present take note that this does not imply that electromagnetic radiation or information can ever travel faster than c. Brieﬂy the reason is that the distance electromagnetic waves can penetrate the metal is negligible.
246 Classical electromagnetic theory
These results are called Fresnel’s equations and apply for spolarization. Fresnel’s equations for ppolarization can be obtained in a similar manner and turn out to be signiﬁcantly diﬀerent: rp = E0r /E0i = (n2 cos θi − n1 cos θt )/(n1 cos θt + n2 cos θi ), (9.73) (9.74) tp = E0t /E0i = 2n1 cos θi /(n1 cos θt + n2 cos θi ). Figure 9.12 shows the reﬂected and transmitted amplitudes for both types of polarization when light is incident in air on glass of refractive index 1.5.
1
tp
0.5
Coefficients
ts
0
rp
0.5
rs θB 1 0
20
40 60 Angle of incidence in degrees
80
Fig. 9.12 Variation of the reﬂection and transmission amplitude coeﬃcients for s/TE and p/TM polarized light as a function of the angle of incidence at a plane boundary between dielectrics in the case that light is incident in the optically less dense medium. The two refractive indices are 1.0 and 1.5.
Several comments can be made about the signs of the coeﬃcients given by Fresnel’s equations which illustrate the importance of associating a diagram of the ﬁeld vectors with Fresnel’s equations. First note that the sign of rp can be reversed if in ﬁgure 9.9 the direction chosen as positive for the reﬂected electric ﬁeld is reversed; that is the arrow labelled Er is reversed. The eﬀect of the choice of this direction becomes obvious at normal incidence, which is illustrated in ﬁgure 9.13. In the case of spolarization the Ei and Er are parallel, but point in opposite directions for ppolarization. This seems strange because at normal incidence the s and ppolarizations are indistinguishable. However if we
9.6
refer to ﬁgure 9.12 we ﬁnd that at normal incidence rp = −rs , so that the reﬂected electric ﬁelds do in fact point in the same direction for sand ppolarization at normal incidence.6 The relative intensities of the reﬂected and transmitted beams are determined by the ﬂuxes in the beams over unit area of the interface. The time averaged absolute ﬂuxes per unit surface area of the interface are given by F = N cos θ = E02 cos θ/2Z. (9.75)
6
Note that it doesn’t help to simply reverse the choice of the positive sense of the reﬂected electric ﬁeld: the annoying complications simply move to grazing incidence.
where Z is the impedance of the material, which in a dielectric is given by eqn. 9.42. The reﬂectance and transmittance are deﬁned as ratios of the ﬂux of radiation leaving a surface area divided by the ﬂux incident over the same surface area, thus Rp/s = (Fr /Fi )p/s = rp/s 2 ,
(9.76)
Tp/s = (Ft /Fi )p/s = tp/s 2 (Z1 cos θt /Z2 cos θi ).
(9.77)
ppolarization
If light is incident on a pile of glass plates at Brewster’s angle then each plate reﬂects a fraction of the spolarized light, while the light with ppolarization is transmitted. After passing through about ten plates the remaining spolarized light is very weak. Therefore the beam reﬂected at Brewster’s angle by a stack of ten plates is of course spolarized, and additionally the transmitted beam is ppolarized. At normal incidence the reﬂectance and transmittance are relatively easy to calculate because all the angles are π/2. For both polarizations R0 = (n1 − n2 )2 /(n1 + n2 )2 .
(9.79)
If there is little absorption the transmittance is T0 = 1 − R0 = 4n1 n2 /(n1 + n2 )2 .
(9.80)
A useful reference quantity is the reﬂectance for a glass/air or air/glass interface at normal incidence. This is 4% for glass of refractive index 1.5, and consequently a sheet of glass reﬂects close to 8% of light incident normally.
Er ki Y
spolarization
kr Ei
Er
ki
Y
Fig. 9.13 Electric vectors in the incident and reﬂected waves at normal incidence. Upper panel for ppolarization and lower panel for spolarization.
1 0.8 Reflectance
After using Snell’s law to replace θt , this equation reduces to sin2 θB = n22 /(n21 + n22 ). Then comparing this result with the identity sin2 θB = 1/(1 + cot2 θB ) gives tan θB = n2 /n1 . (9.78)
kr
Ei
Figure 9.14 shows the reﬂectances for the air/glass interface. At the angle of incidence, θB , at which the reﬂected amplitude for ppolarized waves changes sign its reﬂected intensity vanishes. This angle is called Brewster’s angle. From eqn. 9.73 it follows that n2 cos θB − n1 cos θt = 0.
Fresnel’s equations 247
0.6 0.4 0.2
s/TE θB
0 0
p/TM
20 40 60 80 Angle of incidence in degrees
Fig. 9.14 Reﬂectance for both s and ppolarization when light is incident in air on a plane glass surface. The glass has refractive index 1.5.
248 Classical electromagnetic theory
When sin θt is greater than unity (TIR) the reﬂection and transmission coeﬃcients become complex, which has interesting physical consequences. Then cos θt is purely imaginary so we set
1
rs
cos θt = −iχ,
θB
where χ is real and positive, and the negative sign is chosen to be consistent with eqn. 9.65. With the use of Snell’s law this gives (9.82) χ = (n1 /n2 )2 sin2 θi − 1.
Coefficients
0.5 0
rp
0.5 1 0
Phase shift in degrees
180 160 140 120 100 80 60 40 20 0 0
θC
20 40 60 80 Angle of incidence in degrees
θC
φp φs
θB 20 40 60 80 Angle of incidence in degrees
Fig. 9.15 The variation of the reﬂection amplitude coeﬃcient with the angle of incidence on a plane interface between two dielectrics for both s and ppolarization in the case that the light is incident in the optically denser dielectric. The refractive indices are chosen to be 1.5 and 1.0. In the upper panel the magnitudes of the coeﬃcients are shown and in the lower panel their phases relative to the incident radiation. θB is Brewster’s angle and θc is the critical angle.
(9.81)
Then substituting for cos θt in Fresnel’s equations rs = (n1 cos θi + in2 χ)/(n1 cos θi − in2 χ)
(9.83)
ts = 2n1 cos θi /(n1 cos θi − in2 χ) rp = (n2 cos θi + in1 χ)/(n2 cos θi − in1 χ)
(9.84) (9.85)
tp = 2n1 cos θi /(n2 cos θi − in1 χ).
(9.86)
Both reﬂection coeﬃcients have the form exp (iα)/ exp (−iα), that is exp (2iα), which indicates that the incident light is fully reﬂected with a phase shift of 2α. This result conﬁrms what was stated earlier: the transmitted waves only penetrate a short distance into the less optically dense material and there is no continuous ﬂow of energy perpendicular to the surface. In ﬁgure 9.15 the amplitude reﬂection coeﬃcients are shown in the case of a glass/air interface with n1 = 1.5. In the upper panel the magnitudes appear, and in the lower panel the phases. The reﬂected ppolarized wave (TM) amplitude, of course, disappears when the light is incident at Brewster’s angle. On comparing ﬁgures 9.12 and 9.15 it is apparent that the signs of rs and rp for glass/air at normal incidence are opposite to those for air/glass at normal incidence. This is signiﬁcant in experimental situations where, for example in observing Newton’s rings in reﬂection, there is glass touching glass in air. The two reﬂections, glass/air and air/glass, at the point of contact have opposite phase and interfere destructively. Reﬂection from dielectric metal surfaces involves absorption and then the Fresnel coeﬃcients are again complex. This case is considered in Chapter 11. Fresnel’s equations are consistent with Stokes’ relations obtained in Section 5.6 by applying time reversal invariance: eqns. 5.32 and 5.33 are valid in general. In the more restricted case that the coefﬁcients are real, which excludes absorptive surfaces and TIR, eqns. 5.34 and 5.35 also hold true. The Stokes’ relations apply equally when the interface is made up of one or more thin surface layers on for example a glass block. This fact was already taken advantage of in analysing the function of the Fabry–Perot etalon, and can be useful in situations where the components of the surface are not known or where it would be tedious to calculate directly using Fresnel’s laws.
9.7
9.7
249
Interference ﬁlters
In some modern optical applications surfaces which eﬀectively reﬂect very little light are essential, and in other applications it is equally important to have a reﬂectance as near 100% as possible. It was noted in Section 4.5.5 that the large number of surfaces met with in modern camera lenses are given coatings in order to reduce the reﬂectance to a fraction of a percent per surface. Safety goggles for those people working with lasers must reﬂect eﬀectively all the radiation around the laser wavelength and yet transmit enough light at other wavelengths to provide useful vision. High power lamps can be ﬁtted with cold mirrors that are designed to reﬂect visible light, yet transmit the accompanying longer wavelength infrared radiation. All these eﬀects are achieved by coating the surface of the optical elements with one or more layers of dielectric; the optical thickness of layers is often made exactly one quarter or exactly one half wavelength. Nowadays more than one hundred layers, alternating between two dielectrics, are routinely laid down on a single optical surface. The basic principle is that the individual reﬂections should interfere constructively/destructively to enhance/cancel their reﬂection amplitudes. For this reason these coatings are known as interference ﬁlters. Interference ﬁlters are used widely in optoelectronics for making mirrors that selectively reﬂect over a narrow wavelength interval, particularly in diode laser structures, and are then called dielectric Bragg reﬂectors.
9.7.1
Interference ﬁlters
k’
Analysis of multiple layers
pointing in the x−direction,
(9.87)
while the transverse ycomponent of the magnetic ﬁeld is given using eqn. 9.41 by ZH1 = (U − V ) cos θ pointing in the y−direction.
V exp[i(ω t + k ’.r)] θ
HV
In an individual layer, like that, of thickness d, shown in ﬁgure 9.16, the backward going wave is built up from waves reﬂected in all the later layers, which suggests that the analysis becomes very complicated with large numbers of layers. However things are made simpler once it is realized that the boundary conditions at a surface apply to the total light amplitudes arriving from all possible paths. We therefore concentrate on the total forward and backward going waves in any given layer. In the layer drawn these total waves have electric ﬁelds U exp[ i(ωt − k · r) ] and V exp[ i(ωt + k · r) ] respectively. For the present the waves are taken to be spolarized so that U and V are parallel and directed into the diagram; k and k are wave vectors in the incident and reﬂected directions each making an angle θ to the surface normal. The entry and exit surfaces of the layer, with respect to the incident beam, are given coordinates z = 0 and z = d respectively, and d = dez . Viewing the ﬁelds at time t = 0, the electric ﬁeld at the entry surface is transverse with value E1 = U + V
y
(9.88)
θ
HU
z k
U exp[i(ω t  k.r)]
d
Fig. 9.16 Forward and backward travelling total waves within one layer of a ﬁlter; the ﬁlter consisting of many layers of diﬀerent dielectrics. The electric ﬁelds point into the paper.
250 Classical electromagnetic theory
At the exit surface of the layer shown, with the same choice of directions, E2 = U exp (−iφ) + V exp iφ,
(9.89)
ZH2 = [ U exp (−iφ) − V exp iφ ] cos θ,
(9.90)
where φ = k · d = 2πnd cos θ/λ, n is the refractive index of the layer shown, and λ is the free space wavelength of the radiation. From eqns. 9.87 and 9.88 U = (E1 + ZH1 / cos θ)/2, V = (E1 − ZH1 / cos θ)/2. Substituting these expressions into eqns. 9.89 and 9.90 gives E2 = E1 cos φ − iZH1 sin φ/ cos θ, ZH2 / cos θ = −iE1 sin φ + ZH1 cos φ/ cos θ. Using Z = Z0 /n and putting u = n cos θ, these last two equations become E2 = E1 cos φ − iZ0 H1 sin φ/u,
(9.91)
Z0 H2 = −iuE1 sin φ + Z0 H1 cos φ,
(9.92)
which can be summarized in compact matrix form E1 E2 =M , Z 0 H2 Z 0 H1
where M=
cos φ −iu sin φ
−i sin φ/u cos φ
(9.93)
.
(9.94)
These expressions relate the ﬁeld values at the exit (right hand) plane to those at the entry (left hand) plane. At normal incidence and for a layer of optical thickness one quarter wavelength d = λ/4n. In this case the matrix reduces to 0 −i/n . (9.95) M= −in 0 If the waves have ppolarization the only change needed in the above analysis presented in this section is to change the deﬁnition of u to n/ cos θ. It is at this point that the same boundary conditions that were used to derive Fresnel’s equations are applied: the transverse components of the E and Hﬁelds at a surface in one material are identical to the transverse components at that surface in the adjacent material. Thus the matrix appropriate to a series of layers is simply M = M1 M2 .....Mm
(9.96)
where Mj is the transfer matrix for the jth layer and the material preceding/following the ﬁlter is given the label in/out. Thus the beam is
9.7
incident at an angle θin in a material of refractive index nin . Applying Snell’s law repeatedly we get for the jth layer sin θj = nin sin θin /nj , while φj = 2πnj dj cos θj /λ. (9.97) It is now straightforward to calculate the overall reﬂection and transmission coeﬃcients for a multilayer ﬁlter. Suppose that the incident electric ﬁeld in the absence of the ﬁlter is E0 and that the overall amplitude reﬂection coeﬃcient to be determined is r. With the ﬁlter in place the transverse ﬁelds at the entry surface in the incident medium obtained using eqns. 9.87 and 9.88 are Ein = E0 + rE0 , Z0 Hin = (E0 − rE0 )uin . Correspondingly at the exit surface of the ﬁlter immediately beyond the ﬁlter there is only a transmitted wave Eout = tE0 , Z0 Hout = tE0 uout . The reﬂection and transmission coeﬃcients are simply the magnitudes of the reﬂected and transmitted amplitudes divided by the incident amplitude in the absence of any ﬁlter. The overall transfer matrix is deﬁned by Eout Ein =M . Z0 Hout Z0 Hin Substituting for the ﬁelds in this equation gives t 1+r =M , tuout (1 − r)uin
(9.98)
from which the reﬂection and transmission coeﬃcients can be extracted directly. If there is a single layer of optical thickness one quarter wavelength thick and the light is at normal incidence, M is given by eqn. 9.95, and eqn. 9.98 reduces to t = (−i/n1 )(1 − r)uin tuout = −in1 (1 + r). Dividing the last equation by the previous one and rearranging gives (1 − r)nin nout = n21 (1 + r). The reﬂection will therefore be eliminated if n1 =
(9.99) √
nin nout .
In the case of a coating on glass with refractive index 1.5 in air the antireﬂecting coating would need to have a refractive index 1.224. To be
Interference ﬁlters
251
252 Classical electromagnetic theory
practical a dielectric for optical coatings must be easy to evaporate onto glass, durable and should not absorb moisture. Among the suitable dielectrics magnesium ﬂuoride has a refractive index 1.38 quite close to the value desired in a single antireﬂection coating. A quarter wave layer of magnesium ﬂuoride achieves an amplitude reﬂection coeﬃcient of −0.12 and a reﬂectance of 1.4% for light at normal incidence, compared to 4% in the absence of any coating. A much larger reduction in reﬂectance can be produced using two quarter wave thick layers of diﬀerent dielectrics. Taking the light to be at normal incidence again 0 −i/n2 0 −i/n1 M= −in2 0 −in1 0 −n1 /n2 0 = , (9.100) 0 −n2 /n1 which when substituted into eqn. 9.98 gives
1
Reflectance
0.8
t = −(n1 /n2 )(1 + r), nout t = −(n2 /n1 )(1 − r)nin .
0.6 0.4
Eliminating t from these equations
0.2 0
r = [(n2 /n1 )2 − nout /nin ][(n2 /n1 )2 + nout /nin ]. 0.6
0.8
1 1.2 λ / λ0
1.4
Fig. 9.17 The reﬂectance at normal incidence of an air/glass surface coated with a stack of sixteen pairs of magnesium ﬂuoride and titanium oxide layers. Each layer has an optical thickness of one quarter wavelength at wavelength λ0 . The reﬂectance is plotted against the wavelength divided by λ0 .
(9.101)
The requirement for r to be zero means that in the case of glass in air we need n2 = 1.224n1 . This requirement allows a ﬂexibility in the choice of coating materials that is not possible with a single layer. If the pair of layers is repeated s times the reﬂection coeﬃcient becomes r = [(n2 /n1 )2s − nout /nin ][(n2 /n1 )2s + nout /nin ],
(9.102)
It is thus possible by simultaneously making the ratio (n2 /n1 ) large and by using many layers to go to the other extreme and produce a coating with a reﬂection coeﬃcient extremely close to unity. Magnesium ﬂuoride and silicon dioxide with refractive indices 1.38 and 1.46 are suitable low refractive index materials, while titanium oxide and zinc sulphide are suitable materials with high refractive indices, 2.35 and 2.32, respectively. In most applications it is important that high (or low) reﬂectance is maintained over a range of wavelengths and angles of incidence, and this question of range is considered next. The example just introduced will be used for illustration. There are alternate layers of two dielectrics of equal optical thickness one quarter wavelength. Then the transfer matrix for a pair of layers at any wavelength and any angle of incidence
9.7
is
c −is/u −iu /s c 2 2 c − s u /u −ics(1/u + 1/u ) = , −ics(u + u ) c2 − s2 u/u
M=
c −is/u −iu/s c
(9.103)
where c = cos φ and s = sin φ. At the design wavelength (λ0 in free space) at normal incidence this matrix is a diagonal and r was seen to be real. r remains real over a range of wavelengths ±∆λ around λ0 , given by ∆λ/λ0 = sin−1 [ (u − u )/(u + u ) ]. (9.104) By adding suﬃcient of these large refractive index ratio layer pairs a very high reﬂectance can be achieved over almost all this wavelength interval. When the layers are alternately magnesium ﬂuoride and titanium oxide ∆λ is 0.165λ0 . The behaviour of r for multiple layers of magnesium ﬂuoride/titanium oxide can be calculated with the help of eqn. 9.103 and an example for a stack of 16 pairs of layers is shown in ﬁgure 9.17. Outside the wavelength range λ0 ± ∆λ the value of r is complex for a double layer, and as a result the reﬂectance in ﬁgure 9.17 oscillates as the wavelength is changed. The lobes outside the high reﬂectance band need to be suppressed in applications where a ﬁlter with a very sharp cutoﬀ is required. Examples of this requirement will be met later in telecommunications, where data is transmitted along a single optical ﬁbre using many laser beams of closely spaced wavelengths, maybe 0.4 nm apart. Each laser beam carries a separate stream of information from the others and has to be cleanly separated from them. The reﬂectance is apodized, that is to say the the lobes are removed, by making the refractive index variation across the ﬁlter follow the Gaussian modulation shown in the upper left hand panel of ﬁgure 9.18. The resultant reﬂectance as a function of wavelength is shown in the upper right hand panel. For comparison the lower panels show the refractive index variation and the reﬂectance variation without apodization. Interference ﬁlters can be designed to produce a number of other subtle eﬀects. For example a narrow transmission window can be produced within a broad wavelength range over which the transmission is eﬀectively zero. Two reﬂective ﬁlters of the type just discussed are separated by a layer of optical thickness one half wavelength of the optically denser dielectric: the structure is symbolically (HL)m HH (LH)m , where L/H signiﬁes a quarter wavelength of dielectric of low/high refractive index. This structure is a Fabry–Perot etalon with the two H layers forming the gap; its transmittance has very narrow width peaks at wavelengths such that mλH = 4nH d, where m is an integer and nH is the refractive index of an H layer. Thus the transmission maximum for m = 1 lies in the centre of the broad region of high reﬂectance. A ﬁlter with these
Interference ﬁlters
253
254 Classical electromagnetic theory
1
Reflectance
0.8 0.6 0.4 0.2 0
0.6
0.8
1 1.2 λ / λ0
1.4
0.6
0.8
1 1.2 λ / λ0
1.4
1
Reflectance
0.8 0.6 0.4 0.2 0
Fig. 9.18 Apodization of ﬁlters. The panels show the refractive index variation across the ﬁlter and the reﬂectance. In the upper panels the ﬁlter is apodized and in the lower the refractive index steps are constant.
characteristics is useful in picking out a narrow wavelength range while deleting all nearby spectral lines.
9.7.2
Beam splitters
Simple beam splitters are constructed by putting a multilayer coating on a glass plate, and the coating can be designed so that the reﬂected and transmitted beam intensities have a particular ratio, irrespective of whether the light has s or ppolarization. One method is to sandwich a multilayer coating between the two halves of a glass cube cut diagonally as shown in ﬁgure 9.19. Useful ratios between the percentage reﬂected and transmitted beam intensities, for example 50/50, can be selected through the choice of a suitable sandwich design. Generally the outer faces of the beam splitter cube require an antireﬂection coating to suppress unwanted reﬂections. Another possibility for splitting a beam is to deposit an iconel metal coating on a glass sheet, but this brings a bigger absorption loss. For example the reﬂected and transmitted intensities might both be 32%. Yet
9.8
Modes of the electromagnetic ﬁeld
255
another beam splitter type consists of a coated polymer pellicle only a few microns thick supported in a circular metal frame. A 50:50 split is typical with the light incident at 45◦ . As a compensation for its inherent fragility a pellicle beam splitter introduces no optical aberrations. However there is interference between the reﬂections from the surfaces, just as in a Fabry–Perot etalon. Consequently the fraction of light transmitted/reﬂected oscillates as the wavelength changes by typically 5%. The example analysed here will be for a symmetric dielectric coating between between glass prisms, where the angles of incidence at the coating are all 45◦ , and there is negligible absorption. In this case especially simple relationships hold between the reﬂected and transmitted amplitudes and their phases. Suppose beams are normally incident on the faces 1 and 2 with electric ﬁelds E1 and E2 respectively in ﬁgure 9.19. Then the symmetry of the interface requires that the emerging beams from the faces labelled 3 and 4 have electric ﬁelds respectively E3 = rE1 + tE2 , E4 = tE1 + rE2 ,
(9.105)
where r = r exp (iφr ) and t = t exp (iφt ) are the respective reﬂection and transmission amplitude coeﬃcients. Stokes’ relation, eqn. 5.32, applied to this symmetric interface yields r2 + t2 = 1, while Stokes’ relation, eqn. 5.33, gives rt∗ + tr∗ = 0. In order that these terms cancel they must have opposite phase so
3
(φr − φt ) = (φt − φr ) ± π, Beam in
thus ﬁnally φr − φt  = π/2.
(9.106)
Choosing φt = 0, φr = π/2, we have for this symmetric 50:50 beam splitter √ √ t = 1/ 2, r = i/ 2. (9.107) This illustrates the general property for symmetric beam splitters that the reﬂected and transmitted waves are out of phase by π/2.
9.8
Modes of the electromagnetic ﬁeld
Mechanical systems have a number of degrees of freedom which deﬁne the number of independent types of motion that a system may undergo, and which we call modes. For example a set of n point masses has 3n degrees of freedom and each mode is simply the motion of one mass
1
4
2
Beam in
Fig. 9.19 Beam splitter using a multilayer coating between two 45◦ prisms.
256 Classical electromagnetic theory
in one of the three orthogonal directions. Within electromagnetic wave theory we can identify corresponding modes of the electromagnetic ﬁeld which are analogous to mechanical modes. Modes of the electromagnetic ﬁeld are solution of Maxwell’s equations subject to whatever boundary conditions are imposed by optical elements. For instance, just outside a perfectly conducting mirror the tangential electric ﬁeld would be zero. When the space is unbounded suitable modes with simple mathematical properties are sinusoidal plane waves, which can have any wavelength, any direction and any transverse polarization. Sinusoidal plane waves have simple mathematical properties and as was demonstrated in Chapter 7 any wavepacket with plane wavefronts can always be duplicated by a superposition of sinusoidal plane waves. These component waves will be grouped in wavelength around the mean wavelength of the wavepacket. When there are boundary conditions there will no longer be modes at all wavelengths and travelling in all possible directions. Instead their distribution in wavelength and angle becomes discrete. The most interesting and useful optical arrangements are those for which the boundary conditions have some symmetry, such as the cylindrical symmetry of optical ﬁbre or a Fabry–Perot cavity having circular mirrors. In the latter case the modes have wavelengths such that there are nodes of the electric ﬁeld distribution at the mirror surfaces. If the cavity has small, well separated mirrors the paraxial approximation can be made. Waves are restricted to have the form E(r, t) = A(r) exp [i(ωt − kz)]
(9.108)
where A varies slowly along the beam (z) direction, so that ∂A/∂z
k. Then the wave equation 9.36 simpliﬁes to the paraxial Helmholtz equation (∂ 2 /∂x2 + ∂ 2 /∂y 2 )A − 2ik ∂A/∂z = 0. (9.109) The solutions are Gauss–Hermite functions7 of which the simplest is the TEM00 mode with a Gaussian proﬁle and cylindrical symmetry, described in Section 6.15. The other modes have electric ﬁeld distributions which have broader, more complex shapes, with one or more radial and azimuthal nodes across the mirror planes. The Gaussian mode has the most compact distribution of energy around the optical axis, so that spillage around the mirrors at each reﬂection preferentially depletes the other modes. All waves trapped in a Fabry–Perot cavity can be resolved into a linear superposition of these cavity modes. More generally when there are two modes with the same transverse form but diﬀerent wavelengths they are referred to as diﬀerent longitudinal modes. Correspondingly modes of diﬀerent transverse form are referred to as diﬀerent transverse modes: the Gauss–Hermite modes are good examples. A 7 See for example the ﬁfth edition of Optical Electronics in Modern Communications by Amnon Yariv, and published by Oxford University Press (1997).
9.8
Modes of the electromagnetic ﬁeld
257
wavepacket from the source will almost certainly not all enter a cavity, some being lost around its edges and some reﬂected from it. That part of the wavepacket that is captured is a new wavepacket made up of cavity modes whose wavelengths match those of the incident waves from the source. Two important physical properties of modes can be usefully expressed in mathematical language. The ﬁrst property is that each mode is orthogonal to all the other modes, meaning that it cannot be decomposed into a superposition of the other modes. Put another way they do not overlap: an integral over all space of the product of any pair of modes is always zero. The second property is that any waveform that is consistent with the boundary conditions, that is to say one which describes light conﬁned within an optical structure, can always be replicated by a superposition of the modes for that structure. In a sense the modes are like unit vectors in a space in which each vector corresponds to a possible waveform consistent with the boundary conditions. These properties were explicitly proved in presenting Fourier analysis in Chapter 7 for sinusoidal waves, which are modes for electromagnetic waves in free unobstructed space. If the material through which the wavepacket travels has some dispersion, then modes of diﬀerent wavelength will travel at diﬀerent speeds and hence the shape of the wavepacket changes. Individual modes travel without any change of shape; that is to say they propogate freely. We shall ﬁnd later that optical ﬁbre for long haul useage is constructed speciﬁcally so that only one mode propogates freely along it.
9.8.1
Mode counting
The modes of classical mechanical systems have equal timeaveraged energies when they are in a state of thermal equilibrium. Thus it came about in the late 19th century that scientists anticipated that em modes should, by analogy, also have equal energies in thermal equilibrium. The black body radiation spectrum was therefore conﬁdently calculated – giving a prediction which diverged to inﬁnity at short wavelengths! Experiment revealed instead that the spectrum peaked at a wavelength characteristic of the temperature of the body. The resolution of this diﬃculty led to the development of quantum mechanics and a profound change in the understanding of em radiation. The topic of black body radiation will therefore open the third section of this book where quantum phenomena are introduced. In preparation we evaluate the number of modes available in unit volume of free space using sinusoidal waves as the modes. A simple approach is enough: the modes are counted inside a cubical box with perfectly conducting walls, shown in ﬁgure 9.20. This seems a very
y x
z
Fig. 9.20 Cubical box used for counting modes.
258 Classical electromagnetic theory
restrictive simpliﬁcation, but it turns out that the result is quite independent of the shape or surface of the box and hence is universally valid. The tangential component of the electric ﬁeld must vanish at the walls which lie at x = 0 and L, y = 0 and L and z = 0 and L. The electric ﬁelds of the modes are standing waves Ex = Ax cos (kx x) sin (ky y) sin (kz z), Ey = Ay sin (kx x) cos (ky y) sin (kz z), Ez = Az sin (kx x) sin (ky y) cos (kz z), where Ax,y,z are amplitudes, which must depend on the polarization. In terms of unit vectors oriented parallel to the edges, a wave vector is deﬁned k = kx ex + ky ey + kz ez . The tangential components vanish automatically at the planes through the origin. These tangential components also vanish at the other faces provided that nz Radius k L /π
kx L = nx π, ky L = ny π, kz L = nz π,
(9.110)
with nx , ny and nz all being positive integers. Hence n2x + n2y + n2z = k 2 L2 /π 2 .
ny
nx
Fig. 9.21 The space displaying the electromagnetic modes in a reference volume. The labelling is explained in the text.
(9.111)
The numbers nx , ny and nz are now used as the coordinates along orthogonal axes which deﬁne a new threedimensional space. The possible choices of nx , ny and nz deﬁne points which form the grid shown in ﬁgure 9.21 and are contained within the positive octant of a sphere of radius kL/π. The number of modes is simply the volume of the octant k 3 L3 /6π 2 , so that the number per unit volume of physical space is k 3 /6π 2 . Now consider the values of the amplitudes Ax,y,z . In free space ∂Ex /∂x + ∂Ey /∂y + ∂Ez /∂z = 0,
(9.112)
Ax kx + Ay ky + Az kz = 0.
(9.113)
so that Thus only two of the amplitudes can be chosen freely, or put another way there are two independent polarizations. Then the total number of modes is N = k 3 /3π 2 . Finally the density of modes around wave number k is ρk (k)dk = dN = k 2 dk/π 2 .
(9.114)
This can be reexpressed in terms of frequency using the equality ρf (f )df = ρk (k)dk. Thus ρf (f )df = 8πf 2 df /c3 ,
(9.115)
9.9
Planar waveguides
259
or in terms of the angular frequency ρω (ω)dω = ω 2 dω/(π 2 c3 ).
(9.116)
If the medium is material of refractive index n, rather than free space, this result changes to ρω (ω)dω = n3 ω 2 dω/(π 2 c3 ).
(9.117)
Once the density of modes is known a useful relationship between the mode count and the etendue of an optical system can be inferred. Because the modes are uniformly distributed in direction the fraction directed within the solid angle, Ω, deﬁned by the beam at the entrance pupil is Ω/4π. If A is the pupil area then the light crossing this area in one second originates within a distance c/n from the pupil. The corresponding volume from which the light comes is therefore Ac/n, where n is the refractive index of the material preceding the pupil. Thus the number of modes which pass through the pupil per unit time is N = ρ(ω)dω(Ω/4π)(Ac/n) = 2n2 ΩAdf /λ2 ,
(9.118)
where λ is the free space wavelength. Now using eqn. 4.11 we know that n2 ΩA is simply the etendue of the optical system, so we have N = 2T df /λ2 .
(9.119)
Now it was proved earlier, see eqn. 5.31, that the etendue into a coherence area is λ2 . Therefore it is plausible to take the factor T /λ2 to count the number of modes per unit frequency interval, while the remaining factor two simply counts the number of independent polarization states. The overall product is the number of modes with diﬀerent combinations of transverse proﬁles, frequency and polarization. Further insight into this result is obtained later by applying the uncertainty relation in Section 12.10.1.
9.9
Planar waveguides
Total internal reﬂection is used to trap electromagnetic waves in planar structures as well as in optical ﬁbres. The analysis here of these simpler waveguides will provides a link to the analysis of optical ﬁbres treated in Chapter 16. Figure 9.22 shows the wave vectors of a plane sinusoidal wave reﬂected between plane parallel conducting mirrors that lie a distance a apart. The incoming plane wave has its electric ﬁeld in the xdirection, perpendicular to the diagram and parallel to the surfaces; while the wave vector direction makes an angle θ with the zdirection. The reﬂected wave is directed at −θ. The incident electric ﬁeld is Ex = E0 exp [i(ωt + ky sin θ − kz cos θ)]
x
Ex k1
θ
θi
z k2
a
y Fig. 9.22 A planar waveguide formed by plane parallel conducting surfaces. Wave vectors are drawn for the component plane waves making up the total wave that propogates within the gap. The electric ﬁeld amplitude for one mode is drawn on the left.
260 Classical electromagnetic theory
and the reﬂected wave which cancels this at the upper surface, y = 0, is Ex = −E0 exp [i(ωt − ky sin θ − kz cos θ)]. Thus the total electric ﬁeld is Ex = E0 exp [i(ωt + ky sin θ − kz cos θ)] −E0 exp [i(ωt − ky sin θ − kz cos θ)] = 2i E0 exp [i(ωt − kz cos θ)] sin (ky sin θ). This ﬁeld will also vanish at the lower surface provided ky a = ka sin θ = mπ,
(9.120)
where m is an integer. On the left of the ﬁgure the electric ﬁeld amplitude is plotted for the case that m = 2. Waves satisfying eqn. 9.120 propogate without alteration and are the modes of the waveguide. The mth mode travels along the zdirection with wave vector kz2 = k 2 − (mπ/a)2
(9.121)
and with group velocity c cos θ. Although the electric ﬁeld is transverse to the wave’s direction of travel there is a component of magnetic ﬁeld along the zdirection: such modes are therefore called TE modes. TM modes with a component of electric ﬁeld along the direction of travel can be constructed in an analogous manner. Waves whose wavelengths are greater that 2a cannot satisfy eqn. 9.120 and are reﬂected back at the entry to the waveguide. In general an incident wave will therefore be made up of components which can propagate freely and other components which get reﬂected.
Fig. 9.23 Crosssections of twodimensional waveguides. The material of higher refractive index is shaded.
Similar modes occur when the radiation is conﬁned by TIR within a layer of glass between two layers of glass of lower refractive index. In this case the requirement that a wave travels unaltered is now more complicated. At each reﬂection the wave penetrates into the lower refractive index medium as an evanescent wave and this leads to a phase delay. The reﬂection coeﬃcient for a TE mode is given by eqn. 9.83 and the phase shift is 2αs , where tan αs = n2 χ/(n1 cos θi ).
(9.122)
Then using eqn. 9.82 to replace χ tan αs = (n21 sin2 θi − n22 )/n1 cos θi = (n21 cos2 θ − n22 )/n1 sin θ, where we have used the fact that the angle θ is the complement of the angle of incidence θi in ﬁgure 9.22. In place of eqn. 9.120 we therefore have ky a = mπ + 2αsa + 2αsb , (9.123)
9.9
Planar waveguides
261
where 2αsa and 2αsb are the phase shifts for reﬂections at the two surfaces bounding the guide layer. A wave not satisfying the boundary condition given by eqn. 9.123 would leak out into the external layers and thus be attenuated as it travelled. In terms of rays the phase shift at total internal reﬂection indicates that the reﬂected ray leaves the interface at a point displaced along the surface from where the incident ray struck it. This Goos–Haenchen shift has been directly observed and it occurs because the ray penetrates a little into the less dense medium as an evanescent ray. Crosssections of waveguides with boundaries in two dimensions are shown in ﬁgure 9.23. In each case the external materials have lower refractive index, and in some cases one waveguide surface is bounded by air. At near infrared wavelengths the refractive indices of silicon and silica are respectively 3.5 and 1.46. Therefore such radiation is easily guided along silicon waveguides within silica, though with considerable absorption. A mode must satisfy a duplicate of eqn. 9.123, so we have for a guide of crosssection ax by ay embedded as shown in the left hand lower panel
np
θp
n0
kx ax = mπ + 4αs , ky ay = mπ + 4αs .
θ1
n1
(9.124) Waveguides positioned as in the right hand lower panel can be suﬃciently close so that the evanescent wave of one overlaps the other waveguide. There will therefore be coupling of the waves within the two waveguides, a property whose use in optoelectronics is described in Chapter 16.
n2
np
θp
n0
9.9.1
The prism coupler
Light can be transferred eﬃciently into a planar waveguide by the use of prism couplers as shown in ﬁgure 9.24. The prism is separated by a thin air gap from a planar waveguide of thickness d, whose refractive index is greater than that of the substrate on which it rests. Quantities relating to the prism, air, waveguide and substrate are labelled p, 0, 1 and 2 respectively. Monochromatic light is incident on the base of the prism at an angle greater than the critical angle for the glass/air interface. With an air gap of around half a wavelength the evanescent wave penetrates into the planar waveguide below. The angle of incidence in the prism determines the angle of refraction in the waveguide: using Snell’s law sin θ1 = (np /n1 ) sin θp , so that the transverse component of the wave vector in the waveguide is ky = k1 cos θ1 = k1 1 − (np /n1 ) sin2 θp = (2π/λ) n21 − n2p sin2 θp ,
(9.125)
θ1
n1 n2
Fig. 9.24 Two prism coupler arrangements for launching light into a planar waveguide.
262 Classical electromagnetic theory
where λ is the free space wavelength. This wave entering the waveguide will be a mode provided that it satisﬁes eqn. 9.123
Reflectance
ky d = mπ + 2α10 + 2α12 ,
Wavelength or Incidence angle
Fig. 9.25 Variation of the reﬂection coeﬃcient of a prism coupler with either wavelength or angle of incidence θp . The sharp dips occur when the light is coupled into a mode of the waveguide.
(9.126)
where m is an integer, while 2α10 and 2α12 are the phase shifts for reﬂections at the waveguide/air and waveguide/substrate interfaces respectively. Either the angle of incidence or the wavelength can be varied until both eqns. 9.125 and 9.126 are satisﬁed, and then light is coupled into the waveguide and travels along the waveguide. When this happens the intensity of the light reﬂected from the prism base shows a marked drop as illustrated in ﬁgure 9.25. Whenever the reﬂection coeﬃcient dips in this way it is sure that the light is coupled into the waveguide. The layouts shown in ﬁgure 9.24 are used to measure the properties of thin layers of dielectric that make up optoelectronic components and electronic chips. If the layer is thick enough to support two or more modes at a given wavelength then there is suﬃcient information available from measuring the angles of incidence θp at which coupling into the dielectric layer occurs to extract both the thickness and refractive index of this layer using eqn. 9.126. If the layer is so thin that only one mode is carried, the refractive index is needed in order to calculate the thickness. The technique is complementary to the ellipsometry described in Chapter 11.
Exercises (9.1) Titanium oxide has refractive index 2.35. Calculate the velocity of light in this material and its characteristic impedance. What does the wavelength of light from a laser of wavelength 633 nm change to when it enters the oxide? (9.2) An electric dipole 10−9 C m oscillates at a frequency 10 GHz. Calculate the electric and magnetic ﬁelds at 1 km distance in directions making angles of 90◦ , 45◦ and 0◦ with the dipole axis. What is the time average energy ﬂux in these three directions? (9.3) Calculate the critical angle for a glass/air interface the glass having refractive index 1.5 at a wavelength 633 nm. What is the depth in air over which the light intensity falls by a factor e when light is incident at 42◦ on a glass/air interface, the glass having refractive index 1.5? Repeat the calculation for an angle of incidence 41.82◦ . (9.4) Show that if θB and θB are the Brewster angles for light incident in opposite directions on a plane in
terface between dielectrics, their sum is precisely 90◦ . (9.5) Starting from Fresnel’s equations prove the Stokes’ relations tt = 1 − r 2 and r = −r. r and t are the amplitude reﬂection coeﬃcients and r and t the corresponding coeﬃcients when the same radiation follows the reverse path. (9.6) Calculate the amplitude reﬂection and transmission coeﬃcients using Fresnel’s equations for light incident at 30◦ on an air/glass surface. (9.7) Light is incident at Brewster’s angle on a pile of n thin glass plates. The reﬂectance for the air/glass interface at Brewster’s angle is R for ppolarized light. Calculate the transmittance of the complete stack for ppolarized light. You may neglect multiple reﬂections in this example. (9.8) Express the laws of reﬂection and refraction in vector form. You can use ki , kr and kt to represent
9.9
Planar waveguides
263
a Fabry–Perot cavity made of such mirrors? The the incident, reﬂected and transmitted wave vecrefractive index of magnesium ﬂuoride is 1.38 and tors, and n to represent a vector normal to the surthat of titanium oxide is 2.35. face. In each case the magnitude k = 2π/λ where λ is the wavelength in the material. (9.10) Suppose the refractive index of the XMMNewton (9.9) Calculate the reﬂectance at normal incidence of mirror gold surfaces is 0.9999 at Xray wavelengths. glass of refractive index 1.5 in air which is coated At what angles of incidence will the Xrays be towith one double layer of magnesium ﬂuoride and tally internally reﬂected in air? titanium oxide, each layer being of optical thickness one quarter wavelength. The titanium oxide (9.11) Is the relationship implicit in eqn. 9.105 that r = r inconsistent with Fresnel’s Laws which reis in contact with the glass. Repeat the calculaquire r = −r? tion for four double layers. What is the ﬁnesse of
This page intentionally left blank
Polarization 10.1
10
Introduction
The ability to produce and manipulate light in one state of polarization underlies the function of many modern devices such as DVD reader/writers and liquid crystal displays (LCDs). This technology requires the use of materials that have an anisotropic response to light, which means using nonHIL materials. These include dichroic materials which absorb light in one state of polarization more that in the orthogonal state and so provide polarization ﬁlters.1 Birefringent crystals on the other hand have a refractive index that varies with the orientation of the electric ﬁeld of the light with respect to the crystal’s axes. This property provides the means, for example, to rotate the plane of polarization of light. Similar anisotropic behaviour can be induced in certain materials when an external electric or magnetic ﬁeld is applied. These ﬁeld induced eﬀects are of especial interest because they can be used to switch and to modulate light electronically at the high rates needed for optical communications. A ﬁrst step in this chapter will be to establish the connection between states of circular polarization, introduced brieﬂy in Chapter 1, and states of plane polarization. Dichroic and birefringent materials and their uses will be described in the next section. The Jones matrices, which provide a compact way of characterizing coherent states of polarization, are also introduced and explained. After this the optical eﬀects of applied electric and magnetic ﬁelds and their applications are discussed. Birefringent liquid crystals are the essential element in a considerable fraction of consumer product displays; their properties and use will form the ﬁnal topic in the present chapter.
10.2
States of polarization
Monochromatic electromagnetic radiation travelling in the zdirection in an HIL material can be in two independent states of plane polarization. One of these could have its electric ﬁeld pointing along the xdirection, which in complex form is E1 = E0 ex exp i(ωt − kz), B1 = (E0 /v) ey exp i(ωt − kz),
(10.1)
1
The word dichroic is sometimes also used with a diﬀerent meaning: to describe a material which splits light into beams of diﬀerent colours.
266 Polarization
where v = ω/k is the wave velocity; the other would have ﬁelds orthogonal to these E2 = E0 ey exp i(ωt − kz),
y
B2 = −(E0 /v) ex exp i(ωt − kz).
The time average of the intensity of the light waves formed by superposing these polarization states contains no interference terms
E ωt z
(10.2)
x
Fig. 10.1 Electric ﬁeld of a left circularly polarized wave as seen looking toward the oncoming light. It rotates anticlockwise. The electric ﬁeld of a right circularly polarized wave rotates clockwise.
I = ε0 c(E1 + E2 ) · (E1 + E2 )∗ /2 = ε0 c(E1 E1∗ + E2 E2∗ )/2. Thus these two polarization states can be used as basis states, in a manner analogous to the use of orthogonal vectors, to build any other plane polarized state. For example, by superposing them in the proportions cos θ and sin θ we get light travelling in the zdirection with its plane of polarization inclined at an angle θ to the xdirection. Of course one basis state cannot be formed from the other. The choice of the basis states is not unique: they could be chosen to have their electric ﬁelds at angles θ and θ + π/2 to the xdirection. Adding the chosen basis states with a phase diﬀerence of π/2 produces an electromagnetic wave with new properties. The electric ﬁeld is √ E = (E0 / 2) { ex exp [i(ωt − kz)] + ey exp [i(ωt − kz − π/2)]}, (10.3) of which the real part, the actual electric ﬁeld, is √ E = (E0 / 2) [ ex cos (ωt − kz) + ey sin (ωt − kz)],
(10.4)
and this is shown at the location z = 0 in ﬁgure 10.1. A common convention is employed here when describing states of polarization: the viewer is assumed to be looking towards the oncoming light. Then in the present case the electric vector rotates anticlockwise around the zaxis with angular velocity ω, which is known as a state of left circular polarization. There is a second state of circular polarization in which the electric ﬁeld vector rotates clockwise around the zaxis. This, right circularly polarized state, is obtained by adding the plane polarized states with the ycomponent leading by π/2, √ Er = (E0 / 2) { ex exp [i(ωt − kz)] + ey exp [i(ωt − kz + π/2)]}, (10.5) and the resulting real part is √ Er = (E0 / 2) [ ex cos (ωt − kz) − ey sin (ωt − kz)].
(10.6)
When unequal contributions of the orthogonal plane polarized states are added with an arbitrary phase lag φ, the resultant electric ﬁeld is an elliptically polarized state. The real electric ﬁeld has components Ex = Ea cos χ; Ey = Eb cos (χ − φ) = Eb (cos χ cos φ + sin χ sin φ) (10.7)
10.2
where χ is ωt − kz. Rearranging these equations gives sin χ = ± 1 − (Ex /Ea )2 , and Ey /Eb − (Ex /Ea ) cos φ = sin χ sin φ = ± sin φ 1 − (Ex /Ea )2 .
States of polarization 267 y α
z
x
Squaring this last line gives (Ey /Eb )2 + (Ex /Ea )2 − 2(Ex Ey /Ea Eb ) cos φ = sin2 φ,
(10.8)
which is the equation of an ellipse tilted with respect to the xaxis at an angle α, which depends on φ. The dependence of α on φ will be determined next.
y α
z
x
Let Eu and Ev be the components of the electric ﬁeld refered to axes aligned on the ellipse’s major and minor axes. Ex = Eu cos α − Ev sin α, Ey = Eu sin α + Ev cos α.
y
(10.9)
Then when eqn. 10.8 is expanded in terms of Eu and Ev the coeﬃcient of Eu Ev will vanish; thus
α
z
x
2 cos α sin α(1/Eb2 − 1/Ea2 ) + 2(sin2 α − cos2 α) cos φ/(Ea Eb ) = 0, whence the tilt angle is given by tan 2α =
2Ea Eb cos φ/(Ea2
y
−
Eb2 ).
(10.10)
If Ea /Eb  > 1.0 α is less than 90◦ , otherwise it is between 90◦ and 180◦. In ﬁgure 10.2 the broken lines indicate the path followed by the end point of the electric ﬁeld vector in the xOy plane for various choices of α. The phase lag, φ, increases from zero in the topmost panel to π/2 in the lowest panel. Notice that the largest excursions in Ex and Ey are respectively ±Ea and ±Eb whatever the phase lag. The product E · E∗r vanishes and hence the time averaged intensity of any superposition of left and rightcircularly polarized waves I = ε0 c(aE + bEr ) · (aE + bEr )∗ /2 = ε0 c [ aa∗ E E ∗ + bb∗ Er Er∗ ]/2
(10.11)
contains no interference terms. Therefore the right and leftcircularly polarized states are orthogonal to one another. They make an equally good pair of basis polarization states, meaning that any plane, circular or elliptically polarized state can be reproduced by a linear superposition of right and left circularly polarized states. For example √ √ E1 = (1/ 2)(E + Er ); E2 = (1/ 2)(E − Er ), (10.12) reproduces the states of plane polarization that were introduced at the start of this section.
z
α=0 x
Fig. 10.2 The paths of the end point of the electric ﬁeld vector for elliptically polarized waves. The panels show from the top downward cases of increasing phase lag of the ycomponent ﬁeld with respect to the xcomponent; starting from zero and ending with a lag of π/2.
268 Polarization
When a wave is reﬂected at a perfect mirror the electric ﬁeld of the reﬂected wave cancels that of the incident wave at the surface of the mirror. Hence a linearly polarized wave incident normally is reﬂected with its polarization unchanged. When circularly polarized light is incident normally the electric ﬁeld of the reﬂected wave rotates in the same sense as the electric ﬁeld of the incident wave in order that they cancel at the surface. Now because the reﬂected wave is travelling in the opposite direction to the incident wave, it has the opposite circular polarization: right (left) circular polarization if the incident wave has left (right) circular polarization. In Chapter 5 it was explained how light from actual sources consists of ﬁnite length wavepackets which have random phases. It is necessary to add that the polarizations of the wavepackets from sources are random. For example light from lamps met in the home contains wavepackets whose planes of polarization are uniformly distributed around the ray direction. The interference eﬀects discussed in earlier chapters are only observed if the waves are not only coherent but also to have the same polarization. If the beams being superposed have orthogonal polarizations then, as we have just seen, there is no interference. Interference is totally suppressed in Young’s two slit experiment if polarizers which have orthogonal polarization, are placed one over each slit. We now discuss such polarizers.
10.3
Dichroism and Malus’ law
Materials which absorb light of one plane of polarization more strongly than light of the orthogonal polarization are called dichroic. The most well known example is that of Polaroid sheet. Its basic constituent is a plastic material, polyvinyl alcohol, which is ﬁrst stretched into a continuous sheet so that the long polymer molecules are lined up parallel in the stretch direction. Then iodine ions are deposited from vapour and these attach themselves along the length of the polymer molecules turning the latter into conductors. When light with its electric ﬁeld parallel to the stretch direction falls on the Polaroid sheet a current ﬂows along these conductive paths and energy is absorbed from the light beam. The process is so eﬃcient that the beam is almost totally extinguished by a sheet only a few hundred microns thick. Light with its electric ﬁeld transverse to the stretch axis cannot produce a current because there is no conductive path and hence light with this polarization is not absorbed. Commercial Polaroid sheets are available whose transmittance across the whole visible spectrum for one sense of polarization exceeds 50%, while being less than 10−4 for the orthogonal polarization. More robust dichroic ﬁlters made of glass are needed because high power laser beams would simply melt a sheet of polymer. In one type of glass ﬁlter the surface is covered with equally spaced, parallel metallic strips whose
10.4
Birefringence 269
pitch is less than one wavelength. Alternatively the volume of the glass is loaded with parallel aligned, nanometre sized silver particles. Figure 10.3 shows a beam of initially unpolarized light incident passing through a pair of polarizing sheets whose transmission axes are inclined at an angle θ. Assuming that there is full transmission for one sense of polarization and full absorption for the orthogonal sense, then the electric ﬁeld of light emerging from the ﬁrst sheet, the polarizer, is Ep = E0 cos (ωt − kz) [ ex cos θ + ey sin θ ],
Hence the time averaged intensity (10.13)
a result known as Malus’s law, where I(0) is the intensity incident on the analyser. Thus the intensity transmitted by a polarizer when the light incident is unpolarized would be onehalf the incident intensity. With crossed polarizer and analyser, that is with θ = π/2, no light emerges from the analyser. A practical way to determine of the quality of a polarizer is to direct polarized light onto one face and then to measure the minimum and maximum transmitted intensities as the polarizer is rotated in its own plane about the beam as axis. The ratio of these intensities is called the extinction ratio and is 10−3 or less for simple sheet polarizers. If a beam contains a mix of plane polarized light and unpolarized light the fraction of the polarized light can be determined by continuously measuring the transmittance through a polarizer as this is rotated through an angle of 90◦ about the beam axis. At maximum all the polarized and half the unpolarized light is transmitted, while at the minimum simply half the unpolarized light is transmitted. Thus the fraction of plane polarized light, the degree of polarization is
10.4
x
Polarizer θ
x
Unpolarized light incident
Ea = E0 cos (ωt − kz)ex cos θ.
P = (Imax − Imin )/(Imax + Imin ).
Transmission axis
Transmission axis
while that emerging from the second sheet, the analyser, is
I(θ) = [ ε0 c/2 ] E02 cos2 θ = I(0) cos2 θ,
Analyzer
(10.14)
Birefringence
The physical structure of many crystalline materials is anisotropic so that the ease with which an electric ﬁeld can displace the electron clouds within atoms depends on the direction of the electric ﬁeld relative to the crystal’s axes. As a result the relative permittivity and refractive index also depend on the orientation of the electric ﬁeld: this eﬀect is
Fig. 10.3 Transmission of unpolarized incident light through a pair of Polaroids with their transmission axes making an angle θ. The arrows with solid heads indicate the electric vectors at each step.
270 Polarization
Table 10.1 The refractive indices of several uniaxial crystals, and the wavelength range over which their transmittance is high.
Calcite Quartz LNO LNO YVO4
Wavelength
no
ne
High transmittance
589 nm 589 nm 633 nm 1300 nm 633 nm
1.658 1.544 2.286 2.220 1.993
1.486 1.553 2.202 2.146 2.215
350–4000 nm 200–2300 nm 400–5000 nm 400–5000 nm 400–4000 nm
called birefringence. There can be no such eﬀect when electromagnetic waves travel in either amorphous materials, in most liquids or in crystals such as cubic crystals which exhibit a high degree of internal symmetry. Among liquids the class of liquid crystals show birefringence and these provide the essential component of liquid crystal displays in TV and PC monitors; they are discussed in the ﬁnal part of this chapter.
te
lci
Ca tic
op is
ax
Extraordinary ray
Ordinary ray
Fig. 10.4 Separation of ordinary and extraordinary rays in passing through a sheet of calcite.
A birefringent crystal belonging to the uniaxial class is symmetric under rotations around its optic axis. Calcite (CaCO3 ), crystalline silica (quartz, SiO2 ) and lithium niobate (LiNbO3 ) are commonly met uniaxial crystals. In the case of calcite the carbonate groups (CO3 ) form parallel planes and the optic axis lies perpendicular to these planes. It is much easier to move electrons within rather than perpendicular to these planes and this is the origin of the birefringence. Figure 10.4 illustrates what happens when a narrow beam (ray) of light is incident perpendicular to a sheet of calcite whose optic axis lies in the plane of the diagram. The ray splits: one component obeys Snell’s law, which in this case means it travels undeviated through the sheet; the other component is deviated on entering the calcite. When the calcite sheet is rotated about the incident ray the deviated ray rotates at the same rate. Anyone looking through such a calcite sheet sees not one, but two images of whatever lies behind the calcite. The undeviated image is formed by light with its electric ﬁeld vector in the plane perpendicular to the optic axis, and this is called ordinary polarization. Light producing the other, rotating image has what is called extraordinary polarization: its electric vector points perpendicular to that of the light with ordinary polarization. This ability to split light according to its polarization using birefringent materials provides many useful tools in research and industrial applications. In calcite the electric ﬁeld of light with ordinary polarization is acting in the plane in which the electrons are easy to move and as a result the velocity of the ordinary rays is less than that of the extraordinary ray. Materials like calcite, in which the extraordinary rays travel faster than the ordinary rays are called negative uniaxial materials. Conversely in positive uniaxial materials the ordinary rays travel faster. Table 10.1 lists the refractive indices of several uniaxial crystals and the wavelength ranges over which their transmittance is high. As will be explained below, the extraordinary index varies with the direction of the ray. What
10.4
Birefringence 271
is given in the table, written ne , is the value of this index when the extraordinary ray travels perpendicular to the optic axis of the crystal and the electric ﬁeld points along the optic axis. The ordinary refractive index is the same, no , whatever the ray direction.
10.4.1
Analysis of birefringence
In order to take account of the anisotropy of birefringent materials we shall retrace parts of the analysis of electromagnetic waves given in the previous chapter. Recall that Maxwell’s equations were used to give an equation valid for either the electric or magnetic ﬁeld: the wave equation, eqn. 9.28. Then a plane wave solution was attempted. This was found to satisfy the wave equation provided that the wave velocity satisﬁed the relation √ v ≡ ω/k = 1/ µ0 ε0 µr εr , (10.15) which in the case of a nonmagnetic dielectric amounts to having a refrac√ tive index n = εr . Here, we consider a sinusoidal plane wave travelling in a uniaxial crystal whose ﬁelds in complex form are E = E0 exp [i(ωt − k · r)], B = B0 exp [i(ωt − k · r)]. Then Maxwell’s equations, 9.13, 9.14, 9.15 and 9.16, in the absence of free charge, reduce to k · D = 0; k ∧ E = ωµ0 H;
k · H = 0; k ∧ H = −ωD.
(10.16)
These equations demonstrate that k, E and D are all perpendicular to H, and therefore coplanar, just as for isotropic materials. Taking the vector product of k with the third of the four equations gives k ∧ (k ∧ E) = ωµ0 k ∧ H = −ω 2 µ0 D.
(10.17)
The relative permittivity is no longer a scalar quantity because the electric polarization of the material depends on the direction of the electric ﬁeld with respect to the crystal axes. Fortunately if the material’s absorption is negligible, which is often the case, the constitutive relation becomes relatively simple when the coordinate axes coincide with the crystal’s principal axes Dx εx 0 0 Ex Dy = ε0 0 εy 0 Ey . (10.18) 0 0 εz Dz Ez Put more succinctly D = ε0 ε · E, where the diagonal elements of the matrix ε are εx , etc. In uniaxial crystals there is symmetry around the optic axis, which is taken here to be the zaxis. Then εy = εx = ε1 . Replacing D in eqn. 10.17 gives k ∧ (k ∧ E) = −(ω 2 /c2 )ε · E.
In some crystalline materials the three relative permittivities appearing in eqn. 10.18 are all diﬀerent. Such crystals are known as biaxial and have two optic axes. Their optical properties are more complex than those of uniaxial crystals and will not be discussed here in any detail. A full account of biaxial materials is given in Polarization of Light by S. Huard, published by John Wiley and Sons, New York (1990). Crystals of cubic symmetry are all isotropic; crystals with tetragonal, trigonal and hexagonal symmetry are all uniaxial; crystals with orthorhombic, monoclinic and triclinic symmetry are all biaxial.
272 Polarization
Using the identity of eqn. 9.27 this becomes (k · E)k − k 2 E + (ω 2 /c2 )ε · E = 0.
(10.19)
This is the desired wave equation whose solution will yield the wave velocities, and the refractive indices, of the ordinary and extraordinary waves. Without losing generality, k can be taken to lie in the xOz plane inclined at an angle θ to the optic axis. Then eqn. 10.19 can be expanded to read 2 2 (ω /c )ε1 − kz2 Ex 0 kx kz Ey = 0. 0 (ω 2 /c2 )ε1 − k 2 0 2 2 2 kx kz 0 (ω /c )ε3 − kx Ez (10.20) There is one simple solution with the electric ﬁeld pointing in the ydirection, which requires (ω 2 /c2 )ε1 − k 2 = 0, that is
√ v ≡ ω/k = c/ ε1 .
For this orientation of the electric ﬁeld, in the xOy plane, the constitutive relation, eqn. 10.18, collapses to D = ε0 ε1 E. Thus D is parallel to E just as for isotropic materials. Light with this alignment of the electric √ ﬁeld has ordinary polarization with refractive index no = ε1 independent of the direction of k. Light with ordinary polarization obeys Snell’s law whenever it enters or leaves the birefringent material. The other independent solution involves the remaining two coupled equations (ﬁrst and third lines) of eqn. 10.20 and has extraordinary polarization. In order for these equations to be consistent [ (ω/c)2 ε1 − kz2 ][ (ω/c)2 ε3 − kx2 ] − kx2 kz2 = 0.
(10.21)
After some manipulation this can be reexpressed as a requirement on the refractive index n(θ) of the extraordinary wave, when the wave vector makes an angle θ with the zaxis: n(θ) = kc/ω.
(10.22)
This refractive index has to satisfy 1/n2 (θ) = cos2 θ/n2o + sin2 θ/n2e ,
(10.23)
√ √ with ne = ε3 , and as before no = ε1 . This result shows that the refractive index of the extraordinary wave varies with the angle the wave vector makes to the crystal’s optic axis. At one extreme, when the wave travels along the optic axis the refractive index of the extraordinary wave is the same as that of the ordinary wave. At the other extreme when the wave vector is perpendicular to the optic axis the refractive index
10.4
is ne (given in Table 10.1). Substituting this value of n back into eqn. 10.20 gives the ﬁeld vectors which, apart from a constant factor, are Ee = −ex n2e cos θ + ez n2o sin θ, De = ε0 n2o n2e (−ex cos θ + ez sin θ).
(10.24) (10.25)
These ﬁelds point in diﬀerent directions. Both are perpendicular to the electric ﬁeld of the ordinary wave, which conﬁrms that the ordinary and extraordinary waves are independent solutions of eqn. 10.20.
10.4.2
The index ellipsoid
It is customary when visualizing the electric ﬁeld and displacement vectors for the extraordinary waves to make use of the expression given for the energy contained in the electric ﬁeld in eqn. 9.44 Ue = E · D/2. We choose a value of the magnitude of E such that Ue = 1/2, which makes the notation simpler and crucially does not aﬀect the relative orientations of the ﬁeld vectors. With the coordinate axes along the Optic axis k Ee θ De
Eo
Do
Fig. 10.5 Index ellipsoid for a negative uniaxial crystal. The section perpendicular to the wave vector is shaded, and the circular section perpendicular to the crystal’s optic axis is also outlined. The electromagnetic ﬁeld vectors for a wave with extraordinary (ordinary) polarization bear the subscript e (o). The angle between Ee and De is the angle labelled α in ﬁgure 10.6.
principal axes, the energy equation becomes Dx2 /(ε0 ε1 ) + Dy2 /(ε0 ε1 ) + Dz2 /(ε0 εz ) = 1.
(10.26)
The surface described by this equation is the surface traced out by the endpoint of D and is drawn in ﬁgure 10.5. This surface is called the
Birefringence 273
274 Polarization
index ellipsoid and is drawn here for a negative uniaxial crystal. That for a positive uniaxial crystal would be an ellipsoid with the semiaxis along the optic (z)axis longer than those in the orthogonal directions. A theorem from solid geometry states that at any point on the ellipsoid surface described by eqn. 10.26 the normal to the surface is the vector with components (Dx /(ε0 ε1 ), Dy /(ε0 ε1 ), Dz /(ε0 εz )). Evidently this is the vector E, so that E points normal to the surface of the index ellipsoid. In ﬁgure 10.5 a wave vector, k, is drawn in an arbitrary direction. According to eqn. 10.16 D is perpendicular to k. Let us consider the case of ordinary polarization ﬁrst: the electric displacement, Do is perpendicular to the optic axis as shown in ﬁgure 10.5. It was proved in the last section that the displacement vector of the extraordinary waves is perpendicular to that of the ordinary waves, and is shown labelled De in the ﬁgure. The corresponding electric ﬁelds Ee and Eo are the normals to the index ellipsoid at De and Do respectively. As expected, Eo is collinear with Do , but Ee and De are not collinear.
D
E
α α
k N
H
Fig. 10.6 The ﬁeld vectors of an electromagnetic wave with extraordinary polarization travelling in a uniaxial crystal. The electric displacement, the electric ﬁeld, the wave vector, and the Poynting vector are coplanar. The shaded surface is a wavefront.
For certain directions of the wave vector the orientations of the ﬁeld vectors of the extraordinary wave are simple to describe. When k points anywhere in the plane perpendicular to the crystal’s optic axis De lies along the crystal’s optic axis. In this case the constitutive relation for the extraordinary wave simpliﬁes to De = ε0 ε3 Ee so that its electric and √ displacement vectors are parallel and the refractive index is ne = ε3 . At the same time Do and Eo lie perpendicular both to the crystal’s optic axis and to k. The second simple case is when k points exactly along the crystal’s optic axis. In this limit there is no distinction between the behaviour of the extraordinary and ordinary waves. Both have their electric ﬁelds in the plane perpendicular to the optic axis; for both polarizations D is parallel to E, and both refractive indices are n0 . When k points in a direction intermediate between the optic axis and the xOy plane then eqn. 10.23 shows that the refractive index for the extraordinary wave takes a value intermediate between n o and ne .
10.4.3
Energy ﬂow and rays
In general the electric displacement D of the extraordinary wave is not parallel to the electric ﬁeld E, and therefore its Poynting vector (N = E ∧ H) is generally not parallel to the wave vector. In other words the energy ﬂow is not perpendicular to the wavefronts! The ﬁeld vectors and a wavefront for an extraordinary wave are drawn in ﬁgure 10.6 exhibiting the general requirement that D, E and k are coplanar. It is seen that the velocity of the wavefront measured along the wave vector, vp , is less than its velocity measured along the Poynting vector, vr : vr = vp / cos α,
(10.27)
10.4
where cos α = D · E/(D E). vp is the usual wave/phase velocity and vr is called the ray velocity. A new surface, called the ray surface, is shown in ﬁgure 10.7 for the extraordinary and ordinary rays in the case of a negative uniaxial crystal. This surface is deﬁned as the surface reached after unit time by light from a point source in the uniaxial material and is obtained by rewriting and solving eqn. 10.19 in terms of the vectors N and D. The Poynting vector is radial and the wave vector is normal to the surface shown. Thus the ray surface for the ordinary wave is simply a sphere of radius c/no . In the case of extraordinary polarization the ray surface is the locus of the velocity vector vr satisfying the relation 1/vr2 = sin2 θp /ve2 + cos2 θp /vo2 ,
(10.28)
where θp is the angle between the Poynting vector and the optic axis, ve = c/ne and vo = c/no . The relation connecting the various angles is θp = θ ± α,
Birefringence 275
Optic axis N
θp
vg(o)
vg(e)
Fig. 10.7 Ray surfaces for extraordinary and ordinary waves travelling in a negative uniaxial crystal. The ray velocities are drawn for a particular choice of the direction of the Poynting vector, N.
(10.29)
where the positive/negative sign applies for negative/positive uniaxial crystals. The angles are always measured from the nearer direction of the optic axis.
Huygens’ construction Wavelet E
N
D k Wavefront
Huygens’ construction for wavefronts can now be applied to the propagation of light in uniaxial birefringent materials, with the ray surface providing the shape of the secondary wavelets. Figure 10.8 shows the propagation of a plane wave of inﬁnite extent, having extraordinary polarization, over a time interval t. Each wavelet is a surface that reaches a distance vr t from its origin in the direction θp with respect to the optic axis. A new wavefront at time t is drawn in ﬁgure 10.8 tangential to these wavelets. Although the wave vector, k, is perpendicular to the wavefronts, the Poynting vector N is not: it points from the wavelet origin to where the wavefront touches the wavelet. We note again that while E is normal to N, k is normal to D and α is the angle between them. It will generally be the case that a wave falling on a birefringent crystal has its electric ﬁeld pointing in a direction other than one which would make it precisely an ordinary or an extraordinary wave when it enters the crystal. Such a wave is then a linear superposition of components with ordinary and extraordinary polarization, and as it progresses through the crystal the phase relationship between these components changes. In most cases the paths of the component extraordinary and ordinary waves also diverge, and unless the light is travelling along the optic axis the state of polarization changes. Consequently the pure extraordinary and the ordinary waves provide the appropriate basis states of polarization
Wavefront
10.4.4
Optic axis
Fig. 10.8 Huygens’ construction for a plane wave travelling in a negative uniaxial crystal.
276 Polarization
in birefringent materials. They alone retain the same polarization as they travel through a birefringent material.
10.5
Slow axis Fast axis 45o Plane polarized Left circ. polarized QWP
Fig. 10.9 A quarterwave plate with an incident wave plane polarized at an angle of 45◦ measured in an anticlockwise sense from the fast axis. After passing the QWP the light is left circularly polarized. 2
A piece of clingﬁlm makes quite a good quarterwave plate, and a piece of clear Sellotape is a good approximation to a halfwave plate.
Wave plates
Birefringent materials are used to modify the polarization of beams, turning plane polarized light into circularly polarized light or rotating the plane of polarization. They are also used to physically separate a beam into two orthogonally polarized components. Examples of these techniques and their applications are the topics of this and the following sections. The diﬀerence in the refractive indices of light with ordinary and extraordinary polarization within a birefringent material means that a phase lag develops between these waves as they travel. A quarterwave plate or QWP, shown in ﬁgure 10.9, is a slice of a uniaxial crystal cut with the optic axis lying parallel to its faces, and of such a thickness that there is a phase diﬀerence of π/2 between the ordinary and extraordinary waves of a particular wavelength after they have travelled through the plate. A plate of double the thickness is called a halfwave plate or HWP.2 In quartz the extraordinary waves with electric ﬁeld along the optic axis will travel slower than waves with the orthogonal ordinary polarization. These directions of polarization are therefore called the slow and fast axes, respectively, of the plate. The plate thickness, d, required to give a quarterwave delay is such that λ/4 = (ne − no )d. Using the values given in Table 10.1 shows that a quartz QWP has a thickness of 13.9 µm for light of wavelength 633 nm in air. Consider next that a plane polarized beam is incident on the QWP. If its plane of polarization makes an angle of 45◦ (anticlockwise as seen looking toward the oncoming light) with the fast axis it can be resolved into equal components with polarization along the fast and slow axes, √ Ef (in) = (E0 / 2) cos (ωt − kz), √ Es (in) = (E0 / 2) cos (ωt − kz). When these emerge from the QWP their phase diﬀerence is π/2, √ Ef (out) = (E0 / 2) cos (ωt − kz), √ √ Es (out) = (E0 / 2) cos (ωt − kz − π/2) = (E0 / 2) sin (ωt − kz). Evidently the emerging light is left circularly polarized. If the incident plane of polarization is instead at −45◦ to the fast axis the emerging light will be right circularly polarized. With other orientations of the incident plane of polarization the emerging waves are elliptically polarized. This conversion process can evidently be put in reverse: when a circularly polarized beam is incident the emergent beam is plane polarized. Figure
10.5
10.10 shows a polarizing ﬁlter that is widely used to suppress reﬂections of ambient light. Only the vertically polarized component of the incident ambient light emerges through the ﬁrst polarizer. This is converted to left circularly polarized light by the QWP. On reﬂection from the screen the direction of rotation of the polarization vector is unchanged but the direction of travel is reversed, so this reﬂected beam is right circularly polarized. Passage for a second time through the QWP yields horizontally plane polarized light – which is blocked by the polarizer. A plane polarized wave passing through a HWP emerges with its polarization reﬂected in the fast axis of the HWP, as shown in ﬁgure 10.11. Suppose the incident electric ﬁeld has components along the fast and slow axes Ef (in) = E0 cos θ cos φ; Es (in) = E0 sin θ cos φ, where φ = (ωt − kz). Then the electric ﬁelds at exit are Ef (out) = E0 cos θ cos φ; Es (out) = E0 sin θ cos(φ − π) = −E0 sin θ cos φ.
Wave plates 277
Slow Fast
Linear polarizer
QWP
Mirror
Fig. 10.10 Linear polarizer and quarterwave plate used to suppress the reﬂections of ambient light from a screen. The screen is treated as a mirror. The polarization state of the light at each step is shown.
Slow axis E(in)
Thus the component of electric ﬁeld along the slow axis is reversed in travelling through the HWP.
10.5.1
Jones vectors and matrices
Jones vectors describe the polarization content of a coherent beam and the related Jones matrices describe the action of optical components on the polarization of light passing through them. The restriction to coherent states still permits a useful range of practical applications, particularly with laser beams. Plane polarized beams travelling along the zdirection and with their electric ﬁeld pointing along the horizontal xaxis and the vertical yaxis are represented respectively by twodimensional vectors 0 1 exp iφ, (10.30) exp iφ; v = h= 1 0 where φ = ωt − kz. Plane polarizers which only transmit light polarized along these axes are represented respectively by the matrices 0 0 1 0 . (10.31) ; V = H= 0 1 0 0 The eﬀects of plane polarizers on polarized beams can be summarized as follows: Hh = h; V v = v; Hv = V h = 0. Left and rightcircularly polarized beams are represented by the vectors √ √ 1 1 exp (iφ)/ 2; r = = exp (iφ)/ 2. (10.32) −i i
θ θ
Fast axis E(out)
Fig. 10.11 A halfwave plate looking toward the oncoming light. The polarization state of a plane wave is shown before and after passing through the HWP.
278 Polarization
Correspondingly the action of quarterwave plates whose fast axes are respectively horizontal and vertical can be expressed by matrices 1 0 1 0 ; Qr = . (10.33) Q = 0 −i 0 i The eﬀect of these QWPs on waves plane polarized at 45◦ to the fast axis is √ √ 1 1/√2 exp (iφ)/ 2 = , exp (iφ) = Q (10.34) −i 1/ 2 √ √ 1 1/√2 exp (iφ)/ 2 = r, exp (iφ) = Qr (10.35) i 1/ 2 The values of some Jones vectors and matrices depend on whether the complex wave form associated with a real electric ﬁeld cos (ωt − kz) is chosen to be, as here, exp [i(ωt − kz)], or whether exp [i(kz − ωt)] is the choice. There is no consensus over this choice. If the second choice were made then all the imaginary terms in r, , Q and Qr etc. would reverse their signs.
Crystal optic axis e
e
agreeing with the earlier analysis. The eﬀect of a sequence of dichroic and birefringent elements with Jones matrices J1 , J2 ...Jn would be represented by the product matrix Jn ....J2 J1 . In particular the eﬀect of a halfwave plate with its fast axis horizontal is 1 0 . (10.36) HW = Q Q = 0 −1 Overall phase factors which occur in the matrices and vectors do not aﬀect the state of polarization, so they can be dropped in any calculation used to predict the polarization state alone. If a plane polarized state described by a Jones vector S(0) is rotated by an angle θ anticlockwise the new state has a Jones vector S(θ) = R(θ)S(0),
(10.37)
o o θ
Crystal optic axis
Fig. 10.12 Wollaston prism with unpolarized beam incident. The ray polarizations are indicated by the labels o and e.
where the rotation matrix R(θ) is given by cos θ − sin θ . R(θ) = sin θ cos θ
(10.38)
Also if a polarizer or a waveplate is rotated anticlockwise in its own plane through an angle θ about the beam axis the Jones matrix is modiﬁed as follows P (θ) = R(θ)P (0)R(−θ). (10.39)
10.5.2
Prism separators
These devices take as input an unpolarized beam and ideally output two orthogonally plane polarized beams travelling in well separated directions. With some devices only one beam is in a pure state of polarization, the other beam less so. Figure 10.12 shows a Wollaston prism, in which the component prisms are cut from a uniaxial crystal, in this case calcite, so that their optic axes lie at right angles to one another, as indicated. On entry to the ﬁrst prism the Huygens’ wavelets for the extraordinary component are half ellipsoids whose axes of symmetry point upward in the diagram. Thus the ordinary and extraordinary wavefronts
10.5
remain parallel to the entry surface and travel undeviated in the ﬁrst prism. Thereafter their paths are shown in the ﬁgure labelled o and e. At the interface between the prisms the polarization components of the beam exchange their identities: the ordinary ray in the ﬁrst region becomes the extraordinary ray in the second region and viceversa. Consequently these rays are refracted at the interface away from the incident ray in opposite senses. When the angle labelled θ is set to 45◦ an angular separation of 20◦ is obtained at wavelength 589 nm. Calcite is widely used because it has high birefringence, transparency over a wide range of wavelengths, stability and is available cheaply. The contamination of the alternative polarization in either beam can be as low as 10−5 in standard calcite Wollaston prisms. Separation of one pure plane polarized component is accomplished in Glan prisms by taking advantage of the diﬀerence in the critical angles for TIR of ordinary and extraordinary rays in uniaxial crystals. Two examples using calcite are drawn here: a Glan–air prism is shown in ﬁgure 10.13, and a Glan–Thompson prism in ﬁgure 10.14. In the Glan– air prism the optic axes of the component prisms lie perpendicular to the plane of the diagram and they are separated by an air gap. The prism is cut with one hypotenuse angle around 38.5◦ so that the angle of incidence of light arriving at the calcite/air gap interface is midway between the critical angle for the ordinary ray (37.1◦ ) and that for the extraordinary ray (39.7◦ ). The ordinary ray undergoes TIR while the extraordinary ray is partially transmitted and forms the required pure plane polarized, and in addition undeviated, beam.3 The range in angle of incidence over which the extraordinary ray is transmitted and the ordinary ray undergoes TIR, is called the acceptance angle and amounts to 5◦ for a calcite Glan–air prism. The upper surface of the prism through which the ordinary ray would exit can be blackened to absorb it, or polished to give good transmission, depending on the application. A much larger acceptance cone, of up to 30◦ full angle, is obtained with the Glan–Thompson prism. The gap between the component prisms is now ﬁlled with a transparent material like Canada balsam whose refractive index (1.55) lies midway between that for an ordinary ray and an extraordinary ray in calcite. With this modiﬁcation the ordinary ray undergoes TIR at the calcite/balsam interface at angles of incidence greater than 69.2◦, while the extraordinary ray is partially transmitted at all angles of incidence. Therefore the aspect ratio of the Glan–Thompson prisms is made large, 2.5–3.0, so as to obtain the necessary large angles of incidence at the interface. The sandwich material in a Glan– Thompson prism can only withstand beam ﬂuxes of up to 1 W cm−2 without being damaged. On the other hand a Glan–air prism can cope with 100 W cm−2 . Both types of Glan prism provide only a single pure plane polarized beam travelling forward.
Wave plates 279
o Crystal optic axis 38.5o o e
Air gap
e
Crystal optic axis
Fig. 10.13 Calcite Glan–air prism with unpolarized light incident. The ppolarized ray incident at the interface is the extraordinary ray; the spolarized ray is the ordinary ray. The optic axes of the prism segments are also indicated.
3
The gap in any Glan prism is many times the wavelength of light so that there is negligible frustrated TIR. Crystal optic axis
o 72o e
Canada balsam
Crystal optic axis
Fig. 10.14 Calcite Glan–Thompson prism with unpolarized light incident.
280 Polarization
10.5.3
M
ul
til
ay
er
di e
le
ct
ric
45o
Fig. 10.15 Polarizing beam splitter made from glass prisms with a multilayer dielectric coating on the hypotenuse face.
Polarizing beam splitters and DVD readers
A beamsplitting prism designed to separate an unpolarized incident beam into pure orthogonally polarized beams emerging at 90◦ to one another is drawn in ﬁgure 10.15. The prisms are made of identical glass, either crown or ﬂint. Recall that when a beam is incident at Brewster’s angle on a stack of glass plates the ppolarized component is fully transmitted, and the spolarized component partially reﬂected at each surface. With enough such surfaces the spolarized component is almost completely reﬂected. In a polarizing beam splitter the function of the multiple layers of glass is performed by multiple dielectric layers on the hypotenuse face of one prism. Over a wavelength range of order 200 nm the transmittance of ppolarized light remains above 95% while the reﬂectance of the spolarized light is close to 100%; cross contamination of the other polarization component is typically 0.1% and 2% respectively. All other surfaces in the beam paths are given antireﬂection coatings. Figure 10.16 shows the optical layout for reading a DVD disk using a diode laser, a polarizing beam splitter (PBS), a quarterwave plate (QWP) and a photodiode detector. The laser wavelength lies in the range 635 to 650 nm. The information on the disk is carried in the form of pits etched one quarter wavelength deep in the smooth reﬂective surface. The laser beam is focused to an image spot on the reﬂective surface, and the information is retrieved by the detector in the light reﬂected from the DVD surface. When the spot falls on the clear surface the detector receives full intensity, but when the spot covers a pit there is destructive interference, in the direction of the detector optics, between the reﬂection from the pit and the lands surrounding it. The pits follow a spiral track on the DVD with a pitch of 0.74 µm and lengths between 0.4 µm and 2.0 µm. The reader shown is mounted on an arm which maintains the image spot on the spiral track as the DVD rotates. The intervals of strong and weak reﬂections and the transitions are electronically converted to strings of binary zeroes and ones. The laser light is polarized and its state of polarization at each stage along its path is indicated in the ﬁgure. Its initial polarization state is such that it is transmitted rather than reﬂected by the PBS. During the trip to and from the DVD the sequence already described for ﬁgure 10.10 is repeated with the result that the beam is orthogonally polarized on its return to the PBS and is reﬂected entirely into the detector. This arrangement prevents light reﬂected by the DVD from returning to the laser. If this were permitted then one of the laser mirrors could form a Fabry–Perot cavity with the DVD surface and hence aﬀect the laser’s operation. This deliberate decoupling makes it possible to have a compact unit in which the incident and reﬂected light share some common components. In a read/write DVD the reﬂective layer is an alloy such as GeSb
10.6
DVD
Label Reflective layer Polycarbonate
PBS
Laser
QWP
Mirror
Detector
Fig. 10.16 DVD optical readout with the polarization of the laser beam indicated. For light travelling to the DVD the symbols are drawn with full lines, and for the light returning from the DVD with broken lines. PBS signiﬁes the polarizing beam splitter.
Te whose reﬂection coeﬃcient depends on its phase state. When this material is heated to 200◦ C it cools slowly to a crystalline state which has high reﬂectance; but when it is heated to 500–700◦C it cools rapidly to an amorphous state with low reﬂectance. During the writing phase the power of the laser is raised so that it can melt the alloy: then bursts of high power and lower power produce nonreﬂective and reﬂective track segments respectively.
10.6
Optical activity
Some materials contain molecules or have a crystal structure with deﬁnite handedness, chirality, so that their refractive indices for right and left circularly polarized light diﬀer. This property is called circular birefringence. Now a plane polarized beam can be resolved into right and leftcircularly polarized components as in eqn. 10.12. Consequently when a plane polarized beam travels through a circularly birefringent material a phase diﬀerence develops between its circularly polarized components. This causes the plane of polarization of the incident light beam to rotate as it passes through such materials, a property known as optical activity. Substances which cause the plane of polarization to rotate to the right include the natural sugars and most amino acids, and are called d(extro)rotatory. Antibiotics are among the complementary class of l(aevo)rotatory molecules. Quartz can function either way because it can crystallize in mirror image forms: in one form the silicon–oxygen chains follow a righthanded helix, in the other form a lefthanded helix. Molecules whose structure has such a handedness are called chiral. They
Optical activity
281
282 Polarization
come in two forms, which are the mirror images of one another.
4 The phase is (ωt − kz) = ω(t − nd/c) for either component. Thus the phase lead of the right circularly polarized light is ω(−nr +n )d/c or (2πd/λ)(n − nr ).
The handedness of a molecule is the same viewed from either end. Thus a set of randomly oriented molecules of one handedness produce a net eﬀect on light. Optical activity can therefore be observed even in amorphous materials and liquids, provide their molecules are chiral. If the refractive indices for circularly polarized light travelling in an optically active material are n and nr , then the phase lead accumulated by the rightcircularly polarized component in travelling a distance z is 4 φ = (2πz/λ)(n − nr ).
(10.40)
Using Jones vectors the incident beam is 1 1 1 1 1 + , = 0 2 −i 2 i and the beam emerging has the Jones vector 1 1 1 1 + exp (iφ) i 2 −i 2 exp (iφ/2) + exp (−iφ/2) = [exp (iφ/2)/2] i[exp (iφ/2) − exp (−iφ/2)] cos (φ/2) . = exp (iφ/2) − sin (φ/2)
(10.41)
Thus we see that the outgoing beam has its plane of polarization rotated clockwise through an angle equal to half the phase lead of the right over the leftcircularly polarized component β = (πd/λ)(n − nr ).
(10.42)
In quartz the speciﬁc rotatory power, β/d is 21.7◦ /mm. Photoelasticity When an object made of Perspex is placed between crossed polarizers and viewed in white light the ﬁeld is dark. However, if the object is stressed, coloured bands appear across the area it covers, indicating that the stress has induced some birefringence. Regions that form a continuous band of one colour mark regions of equal stress. Models of large scale structures construced from clear plastic can be studied in this way in order to assess the stresses that are likely to be present in the full scale object.
10.7
Eﬀects of applied electromagnetic ﬁelds
Externally applied electric and magnetic ﬁelds exert forces on the electron clouds in atoms and molecules, which in turn cause changes to their
10.7
Eﬀects of applied electromagnetic ﬁelds 283
conﬁguration. This leads to a change in the relative permittivities appearing in the equation for the index ellipsoid, eqn. 10.26. The refractive index which at low values of the applied ﬁeld is n0 , becomes n = n0 + ∆n, where ∆n depends on the orientation of the applied ﬁelds E or B, and of the polarization of the light with respect to the crystal’s optic axis. Only the simpler cases will be of interest here, and to cover these cases the change is written conventionally ∆n = n3 (rE + sE 2 + rB B)/2,
(10.43)
−∆(1/n2 ) = rE + sE 2 + rB B.
(10.44)
so that Here r, s and rB are coeﬃcients which depend on the material, and on the orientations mentioned above. The corresponding optical eﬀects are called respectively the Pockels eﬀect, the Kerr eﬀect and the Faraday eﬀect and will be described in the following sections. Evidently r, s and rB are tensors whose components depend on the orientation of the applied ﬁeld and the polarization direction of the light. However the crystals are cut, and the ﬁelds applied, in such a way that usually only one component of the relevant tensor is of importance in each device. The changes in optical behaviour of crystals produced by applied electric ﬁelds have provided a useful interface between electronics and optics. Electronic signals can be used to switch light on and oﬀ, or to modulate light with an analogue signal at rates up to tens of gigahertz. Optoelectronic devices exploiting this capability are frequently used in research and telecoms. One application has already been met in Section 8.12.1: an electrooptic modulator was used to lock together the frequencies of a cavity and a laser.
10.7.1
Pockels eﬀect and modulators
The Pockels eﬀect is the birefringence induced by an applied external electric ﬁeld, and is linearly proportional to the ﬁeld strength. When the applied electric ﬁeld is reversed the eﬀect is reversed, and we will show that for this reason no Pockels eﬀect is observed in certain symmetric crystals or in liquids. Suppose that the basic cell of a crystal contains atoms at coordinates ri and that it is centrosymmetric crystal: meaning that its basic cell is symmetric under the transformation: ri → −ri . In such a material equal and opposite applied electric ﬁelds E and −E would produce the same change in the electron structure of the cell. This means that the change in refractive index will be the same for these two ﬁelds. Hence the Pockels eﬀect must vanish in centrosymmetric crystalline materials and equally, using the same argument, in any isotropic liquid. In other less
284 Polarization
symmetric crystals it is the dominant electrooptic eﬀect. The alteration in the refractive index produced by an applied electric ﬁeld E is LNO Optical axis
∆n = n3 r · E/2,
(10.45)
where the Pockels coeﬃcients depend on the relative orientation of the crystal axes, the plane of polarization and the direction of the applied electric ﬁeld. L
E(beam)
d
+

+
E(applied) Fig. 10.17 Mach–Zehnder interferometer on a lithium niobate crystal. The lower diagram is a crosssection taken at the broken line. The indiﬀused titanium waveguides are shaded, and the gold electrodes are crosshatched.
Crystals of the positive uniaxial material lithium niobate (LiNbO3 , or LNO) show a strong Pockels eﬀect and are transparent for wavelengths from 400 to 5000 nm. Lithium niobate is ferroelectric: a crystal can have a very large permanent electric dipole moment in the absence of any applied electric ﬁeld and this moment can be reversed by applying a suﬃciently strong electric ﬁeld. The mechanism involved is the formation of domains within which the dipole moments of the individual crystal cells line up parallel. In LNO alternate domains are parallel and antiparallel. An applied ﬁeld causes the polarization of all the domains to align with itself so that a very large electric susceptibility and relative permittivity result. LNO possesses not only a strong Pockels eﬀect, but also strong piezoelectric and acoustooptical responses. Potassium dihydrogen phosphate, KDP, is a similarly versatile crystalline material.5 With no voltage applied the cell does not aﬀect the polarization. Rapid switching in 1 ns or less in standard. Figure 10.17 shows a device of a type used in telecoms to modulate the intensity of electromagnetic radiation in the visible or near infrared at frequencies of order 10 GHz. It consists of a LNO crystal a few centimetres long in which waveguides are formed to make the arms of a Mach–Zehnder interferometer. These waveguides are produced by diffusing vaporized titanium into the crystal. Their crosssection is a few microns across, similar to that of the core of a single mode optical ﬁbre, and their refractive index is a little larger than that of the undoped LNO crystal. Thus light injected into one end of the waveguide is conﬁned by TIR to travel along the waveguide just as in an optical ﬁbre. Polarized light from a laser source enters via an optical ﬁbre at one end of the waveguide, and exits into another optical ﬁbre at the the other end of the waveguide. Gold electrodes are deposited alongside the arms of the Mach–Zehnder and an electric potential is applied between them with the polarity shown in the ﬁgure. This voltage produces equal and opposite electric ﬁelds in the two arms and hence opposite Pockels eﬀects. A simple alignment is shown here: the crystal optic axis, the direction of polarization and the applied electric ﬁeld are all parallel. The refractive indices in the arms of the Mach–Zehnder are modiﬁed from their value 5 The Pockels coeﬃcients for these important complex crystals, LNO and KDP, are discussed and tabulated in Polarization of Light by S. Huard, published by John Wiley and Sons, New York (1990).
10.7
Eﬀects of applied electromagnetic ﬁelds 285 1
ne in the absence of an applied electric ﬁeld to become n = ne ± r33 n3e E/2,
0.8 −1
mV is the appropriwhere E is the applied ﬁeld and r33 is 30.8 10 ate Pockels coeﬃcient for LNO at 633 nm wavelength. Thus the phase diﬀerence between light in the two arms at the point they reunite is
λ being the vacuum wavelength of the light and L being the length of the electrodes. This phase diﬀerence can be reexpressed in terms of the electrode separation d and the applied voltage V as (10.46)
If we deﬁne the electric ﬁeld of the light entering each arm to be Ein = E0 exp (iωt), then that emerging from the device is Eout = E0 [ exp (iωt) + exp [i(ωt + φ)] ], whose time averaged intensity is ∗ = 2E02 [ 1 + cos φ ]. I = Eout Eout
(10.47)
This behaviour is shown in ﬁgure 10.18 as a function of the applied voltage. The modulator is biased with a ﬁxed voltage between the electrodes which is large enough to bring the operating point on the intensity versus voltage curve into the region where the slope is linear. Any small amplitude signal voltage applied on top of this ﬁxed bias will be replicated in the variation of the intensity of the light transmitted by the modulator into the output ﬁbre. The voltage that cuts the light oﬀ completely, Vπ , is called the halfwave voltage. Evidently Vπ = λd/(2n3e r33 L).
0.6 0.4 0.2
φ = (2π/λ)r33 n3e EL,
φ = (2π/λ)r33 n3e (LV /d).
Intensity
−12
(10.48)
Taking the wavelength to be 633 nm, the length of the electrodes to be 1 cm and their separation to be 10 µm, then the halfwave voltage for the LNO modulator being considered here would be 1 V. This value lies neatly within the range of voltages employed in modern electronics. The time taken for light to pass between the two ends of the electrodes imposes an upper limit on the frequency of signals that can be used to modulate the light. This transit time, τ , equals ne L/c and only signals of frequencies well below 1/τ will be reproduced without distortion in the light intensity output from the modulator. This limit can be raised appreciably by designing the electrodes so that they carry a travelling wave which propogates along the length of the electrodes parallel to the light in the waveguide.
0
Bias point Vπ Voltage
Fig. 10.18 The intensity transmitted through the Mach–Zehnder modulator versus the applied voltage.
286 Polarization
10.7.2
Kerr eﬀect
In liquids and centrosymmetric crystals the residual electrooptic eﬀect is the Kerr eﬀect, which is proportional to the square of the electric ﬁeld. The applied electric ﬁeld induces birefringence with the electrical ﬁeld direction becoming the optic axis, and the material therefore behaves as a uniaxial crystal. If the applied ﬁeld is E then the diﬀerence between the refractive index for light polarized parallel (extraordinary polarization) and perpendicular (ordinary polarization) to the applied electric ﬁeld is expressed as ne − no = κλE 2 , (10.49) where κ is called the Kerr coeﬃcient. A material with a positive Kerr coeﬃcient therefore has polarization properties similar to those of a positive uniaxial crystal. One liquid that has often been used, nitrobenzene,6 has a very large Kerr coeﬃcient, 2.4 10−12 V−2 m, while that for water is only 4.4 10−14 V−2 m. It follows also that
Applied electric field Polarizer
∆n = ne − n = n2 I,
Analyzer
(10.50)
where n is the refractive index in the absence of an applied ﬁeld and I is the light intensity. Kerr cell
Polarization along cell with Vπ applied
Fig. 10.19 Kerr cell sandwiched between crossed linear polarizer and analyser. The polarization state of the beam travelling through the Kerr cell is indicated for the case that the applied voltage makes the cell equivalent to a halfwave plate.
Figure 10.19 shows a fast optical shutter consisting of a Kerr cell located between a crossed polarizer and analyser. The electric ﬁeld applied across the cell is oriented at 45◦ to the transmission axes of both polarizer and analyser. In the absence of any external electric ﬁeld there is no transmission. However when an electric ﬁeld is applied a phase delay is produced between light polarized along and perpendicular to the applied ﬁeld direction. This delay can be inferred using eqn. 10.49 to be ∆φ = 2πκL(V /d)2 , (10.51) where L is the length and d the width of the cell containing the active material and V is the applied electric potential. When the phase diﬀerence is exactly π the Kerr cell is equivalent to a halfwave plate with, in the case of a material with a positive Kerr coeﬃecient, the slow axis pointing along the direction of the applied electric ﬁeld. The plane of polarization of the light therefore rotates by exactly 90◦ in passing through the Kerr cell and so the beam is transmitted by the analyser. For example, a cell of width 1 cm and length 4 cm ﬁlled with nitrobenzene requires a voltage of 2260 V to produce a phase diﬀerence of π. Switching is extremely fast, capable of following signals of frequencies up to 10 GHz. This high switching speed has been exploited in the determination of the speed of light made by measuring the travel time of short pulses of light over a known path distance to and from a distant mirror. Kerr cell shutters can readily be constructed with apertures adequate to cover standard camera lenses and have been extensively used in high speed photography. 6 Unfortunately
this is a corrosive and unstable material.
10.7
10.7.3
Eﬀects of applied electromagnetic ﬁelds 287
Faraday eﬀect
Faraday, in 1845, made the ﬁrst observation of an eﬀect of an applied electromagnetic ﬁeld on light: he observed that the plane of polarization of light travelling in glass was rotated when a magnetic ﬁeld was applied along the path of the light. In the Faraday eﬀect a diﬀerence is produced between the refractive index of right and left circularly polarized light travelling along the direction of the magnetic ﬁeld ∆n = n3 rB B,
(10.52)
where B is the applied ﬁeld and rB an optical constant that depends on the material. Over a path length L a phase diﬀerence (2π/λ)∆nL develops between the two states of circular polarization. If instead the incident light is plane polarized we can use a result proved in Section 10.6: that the plane of polarization will rotate through an angle equal to half this phase diﬀerence,
Magnetic field direction inside rotator Polarizer
Polarizer
θ = (π/λ)∆nL = (π/λ)n3 rB LB, where eqn. 10.52 was used to replace ∆n. This result is generally contracted to read θ = VLB, (10.53) where V is called the Verdet constant. By convention this constant is taken to be positive if the rotation is left handed when the light travels in the direction of the applied magnetic ﬁeld. Materials containing paramagnetic ions have relatively large Verdet constants. The Verdet constant, expressing the rotation angle in radians, is typically 90 T−1 m−1 for glass doped with terbium ions, but only 3.8 T−1 m−1 for water, all measured at a wavelength in air of 633 nm. Many older texts quote the Verdet constant in the unit arcmin/amp: this value is obtained by dividing the constant in rad T−1 m−1 by 235.1. Unlike the Kerr eﬀect, the rotation of the polarization induced by an external magnetic ﬁeld is in one sense (say left handed as seen looking toward the oncoming light) when the light travels in the direction of the ﬁeld and in the reverse sense when it travels in the direction opposite to the ﬁeld (right handed when viewed looking toward the oncoming light). Thus the Faraday rotation is in the same sense when it is observed in a ﬁxed direction, for example looking along the magnetic ﬁeld direction. This feature makes it possible to construct optical isolators which permit light to travel through them in only one direction. Figure 10.20 shows such an optical isolator consisting of a Faraday rotator sandwiched between a polarizer and analyser whose transmission axes are inclined at 45◦ to each other. A typical rotator is a terbium doped glass rod located inside a powerful permanent magnet. The paths and the corresponding polarizations of light entering from both the left and the right are also indicated in ﬁgure 10.20. In each case the plane
Rotator within permanent magnet
blocked
Fig. 10.20 Optical isolator consisting of a Faraday rotator sandwiched between polarizers with their transmission direction inclined at 45◦ . The polarization states for light travelling in either direction are shown below the device as seen from the left. The Faraday rotator is taken to have a positive Verdet constant.
288 Polarization
of polarization is rotated clockwise through 45◦ by the Faraday rotator, as seen from the left hand end. Right moving light is transmitted but left travelling light is blocked. Isolators are widely used when it is necessary to prevent light from a laser being reﬂected back into the laser. Such a reﬂected beam would be eﬀectively conﬁned in a cavity formed by laser mirror and the external reﬂecting surface, and this can easily disturb the laser’s operation. When very high powers are involved the polarizers used would be Glan–air prisms rather than a sheet of Polaroid. The Verdet constant contains the factor λ−1 so that it changes rapidly with wavelength. Consequently a Faraday rotator needs to be tuned in order to function at diﬀerent wavelengths by moving the glass rod further into or out from the magnet air core.
10.8
Liquid crystals
The liquid crystal display, LCD, is now ubiquitous, providing the displays in hand calculators, laptops and mobile phones. It has also almost entirely displaced the electron beam TV and PC monitors due to a combination of advantages that are explained below. Liquid crystals are liquids which contain anisotropic molecules and which possess some degree of internal order. The molecules diﬀuse just like those in any liquid but they retain either some degree of alignment among the molecules (nematic ordering) or some degree of alignment and positional ordering of the molecules (smectic ordering). Thanks to this ordering liquid crystals are highly birefringent and optically active, properties which make them ideal elements for electronically driven displays.
N
SmA
SmC
Fig. 10.21 Distribution and alignment of molecules in nematic, smectic A and smectic C liquid crystals.
When a crystalline substance melts to become a liquid crystal the latent heat is almost as large as that evolved in a transition from a crystal to a pure liquid, and when the liquid crystal is further heated so that it changes to a pure liquid the latent heat from this phase transition is much less. This shows that the ordering in a liquid crystal is much weaker than in a crystalline solid. The liquid crystals discussed here are called thermotropic, that is to say there is a limited temperature range over which they are stable as liquid crystals. Molecules in liquid crystals of interest here have a stiﬀ rodlike centre section with more ﬂexible extensions and there are strong attractive forces between neighbours. 7 Figure 10.21 contrasts the purely orientational ordering in a nematic, N, liquid crystal and the orientational ordering and layering in smectic liquid crystals. In smectic SmA liquid crystals the molecules align normal to the layers, while in smectic SmC liquid crystals the molecules are tilted. The average direction of alignment over a local region contain7 There are other liquid crystals that are formed when a material is dissolved in a solvent, such as soap in water. These lyotropic liquid crystals are not of importance in the applications discussed here.
10.8
ing many molecules is called the director. The nematic liquid crystals are uniaxial and the director direction deﬁnes the local optic axis of the liquid crystal. Although the discovery of liquid crystals goes back to the 19th century the technological exploitation of their optical properties came only later, after George Gray had successfully synthesized liquid crystals which were stable over a useful temperature range around room temperature. The structure of liquid crystals can be chemically engineered so that the diﬀerence between the extraordinary and ordinary refractive indices, ∆n, has a value appropriate to the application, usually around 0.1 for LCD usage.
10.8.1
The twisted nematic LCD
Figure 10.22 shows the basic components of a simple twisted nematic (TN) liquid crystal display invented in 1977 by Schadt and Helfrich. A layer of uniaxial nematic liquid crystal, about 5 µm thick, is sandwiched between two glass sheets. On the inner glass surfaces a layer of polyimide is deposited, typically 100 nm thick, whose surface has been textured by gently rubbing in a ﬁxed direction. This causes the liquid crystal molecules close to the surface to align nearly parallel to the rubbing direction. By arranging the rubbing directions on the two facing polyimide surfaces to be orthogonal, the director is forced to twist through 90◦ between these two surfaces. Polarizing sheets are glued to
Field off Field on Fig. 10.22 A twisted nematic liquid crystal. The arrows indicate the transmission directions of the polarizers, and the light lines the rubbing direction on the polyimide layers. On the left transmitting light with no ﬁeld applied; on the right blocking light when an electric ﬁeld is applied.
the outer surfaces of the glass sheets with their transmission axes aligned parallel to the rubbing direction of the polyimide surface, which means
Liquid crystals 289
290 Polarization
the polarizers are crossed. When the LCD is used in transmission the panel is illuminated from below by a broad white light source.
8
The ratio of the phase lag to twice the twist is called the Mauguin parameter.
The left hand panel in ﬁgure 10.22 shows the ‘oﬀ’ state with no voltage applied to the cell. Light entering from below is polarized parallel to the rubbing direction on entering the liquid crystal. In the liquid crystal the optic axis, coincident with the local director direction, rotates through 90◦ across the cell. Evidently left and right circularly polarized waves will have slightly diﬀerent refractive indices because of the rotation of the optic axis across the liquid crystal. From the analysis made in Section 10.6 it follows that the plane of polarization of the light incident will rotate as it progresses. With the correct choice of the liquid thickness and the liquid crystal’s birefringence the plane of polarization of the incident light can be made to follow the rotation of the optic axis through 90◦ and it will then be transmitted by the upper polarizer. The requirement for this adiabatic following is that the phase diﬀerence between the extraordinary and ordinary rays is much greater than the angular twist across the liquid crystal thickness8 (2π/λ)∆n z π/2, where ∆n is this diﬀerence in refractive indices, and z the cell thickness. That is (4∆n z/λ) 1. Thus if ∆n is around 0.1 and the wavelength is 0.5 µm, then a thickness of 5–10 µm is adequate to produce a rotation through 90◦ which can be adiabatically followed by the electric ﬁeld. As a result the light is transmitted by the upper polarizer.
Polarizer Glass Polyimide Liquid crystal Polyimide Electrode
Colour filter Electrode TFT Glass
Polarizer Backlight
Fig. 10.23 Component layers of a pixel of a twisted nematic LCD.
In the ‘on’ state shown in the righthand panel an electric ﬁeld is applied perpendicular to the cell. The molecules become polarized and, while those very close to the surfaces remain pointing along the rubbing axis, those throughout the body of the liquid crystal tip so that the director points along the ﬁeld direction. With this new alignment the incident light travels along the optic axis of the liquid crystal, the plane of polarization of the incident light is no longer guided, and the light is blocked by the upper polarizer. What is shown in ﬁgure 10.22 is known as a normally white (NW) display. A normally black (NB) display is obtained if the transmission axis of the upper polarizer is set parallel to that of the lower polarizer. Further details of the structure of a twisted nematic LCD are shown in ﬁgure 10.23. A white backlight is provided by either a ﬂuorescent lamp plus diﬀuser, or by an electroluminescent sheet. The electric ﬁeld across the liquid crystal is applied via transparent conductive 100 nm thick indium tin oxide electrodes deposited on the glass ahead of the polyimide layer. It requires only a few volts across the thin layer of liquid crystal to attain the high electric ﬁeld needed to switch the molecular alignment.
10.8
However the switching is slow because the molecules have to be turned against the viscous drag: a full cycle oﬀ/on/oﬀ takes at least 15 ms to complete. LCDs used as PC monitors are divided into pixels so as to match standard formats, for example the SXGA format with 1280×1024 pixels. Each pixel contains three electrically independent subpixels, covered by respectively red, green and blue colour ﬁlters. These ﬁlters are in the form of continuous vertical bands so that the subpixels on an 18inch screen are then 0.28 mm tall and 0.093 mm wide. Individual subpixel voltages are applied using an active matrix (AM) addressing as shown in ﬁgure 10.24. Thin ﬁlm transistors (TFT) located at one corner of each pixel are used to turn the pixel voltage on and oﬀ. Control voltages are applied via electrical buses: one set of which run between the rows and the other between the columns of pixels. A data voltage is applied in sequence to each column in turn and at each of these steps a gating voltage is applied to just those rows that contain pixels that are to be turned on. If a DC voltage were applied continuously to a liquid crystal, then impurity ions would be driven onto the polyimide surfaces where they would adhere and aﬀect its ability to align the liquid crystal molecules. In order to avoid this eﬀect the voltage on any pixel is alternated between frames. LCD panels have an overall thickness of a few millimetres. The transmission coeﬃcient is restricted to 5% at best, because of the absorption in the many layers. This is still suﬃcient to outclass the electron beam tubes in brightness when using only modest backlight intensities. The displays in watches and calculators are illuminated by ambient light only, so that the power consumption is reduced to a tolerable level. In these devices a reﬂector forms the back layer. The immediate advantages of the LCD over the electron tube for use in comparable displays are the ﬂat format, light weight, avoidance of high voltage and roughly halved power consumption. A signiﬁcant problem with TN LCDs was the rapid fall oﬀ in the contrast and the ﬁdelity of the colour as the viewing direction is moved away from the perpendicular to the display surface. The origin of the problem is that when the molecules align perpendicular to the display surface the birefringence varies strongly with the viewing angle, and this is made worse by the molecules not being fully aligned. To alleviate this problem thin surface layers of diﬀusing and birefringent material are added to extend the angular range of satisfactory contrast and colour ﬁdelity.
10.8.2
Inplane switching
A more eﬀective solution is illustrated in ﬁgure 10.25 with the electrodes now located on one glass surface only. The nontransmitting ‘oﬀ’ state is shown in the lefthand panel in which light emerging from the lower polarizer has its electric ﬁeld along the director direction. Therefore the polarization is unchanged as the light crosses the liquid crystal and is
Liquid crystals 291
Data line
Common electrode Pixel electrode
LC
Drain
Scan line
Fig. 10.24 Active matrix switching for LCD. The CMOS transistor obscures only a small fraction of the pixel area.
292 Polarization
Field off
Field on
Fig. 10.25 Inplane switching LCD. The dark bands on either side of the lower surfaces are electrodes. The arrows mark the directions of the polarizers’ transmission axes. In the right hand diagram with electric ﬁeld applied the liquid crystal molecules set at 45◦ to the directions of the two transmision axes in the body of the liquid.
blocked by the upper polarizer. In the ‘on’ state, shown in the right hand panel, the applied electric ﬁeld is parallel to the glass surface rather than perpendicular to it as was the case for TN LCDs. Suppose the director of the liquid crystal, which deﬁnes its optic axis, sets at an angle θ to the direction of transmission of the lower polarizer as shown in ﬁgure 10.26. Then the electric ﬁeld of the light entering the liquid has components parallel and perpendicular to the optic axis, Ee = cos θ; Eo = sin θ, where all unnecessary factors have been removed. When these components emerge they have become
D
Ee
ire
ct or
Upper polarizer axis
Ee = cos θ; Eo = sin θ exp (−iφ), θ
Lower polarizer axis 1
Eo
Fig. 10.26 Director and polarizer axes in twisted nematic cell with ﬁeld oﬀ.
where the phase diﬀerence φ = (2π/λ)∆nd, for a cell of thickness d and a diﬀerence in the refractive indices ∆n; λ is the wavelength in free space. The component of the electric ﬁeld along the direction of transmission of the analyser is therefore EA = cos θ sin θ − cos θ sin θ exp (−iφ). Thus the time averaged intensity of the light exiting the LCD is ∗ EA = sin2 (2θ) sin2 (φ/2). IA = EA
(10.54)
10.8
The intensity is maximized by having the director at 45◦ to the polarizers, and making the liquid thickness such that the phase lag φ is 2π. In order to keep the cell thickness large enough for cheap manufacture (several microns) ∆n must be relatively small. Note that in switching between the two states shown in ﬁgure 10.25 the liquid crystal molecules remain parallel to the glass surfaces; this process is therefore called inplane switching (IPS). Because the molecules’ axes stay parallel to the glass the birefringence changes very little with the viewing angle. The contrast and colour ﬁdelity remain excellent even when the LCD is viewed up to 80◦ oﬀ axis. Nowadays LCD pictures are generally bright (∼500 cd m−2 ) and of high static contrast (∼2000:1) and can be used out of doors. The dynamic contrast attained by darkening the backlight in dark scenes can be several thousand. The newer plasma sceens have a better inherent contrast, the pixels being black when oﬀ. LCD as well as plasma response is rapid enough to display fast sports without introducing any smearing.
Multilayer dielectric mirrors
White source
R
G
B
LCDs
Mirrors
Lens to screen
Fig. 10.27 LCD projector. The LCDs can be either simple twisted nematic or polymer dispersed LCDs because only a narrow beam is required.
Compact projectors are manufactured using LCDs.9 The notional layout of such a projector is shown in ﬁgure 10.27 using three LCDs of around 20 mm lateral dimension and with a pixel pitch of about 15 µm. Mirrors with multiple layer dielectric coatings are used to selectively reﬂect the red, green or blue components of the incident white light onto the individual LCDs. The LCDs are of simple construction because no colour masks are needed and the beam is concentrated over a narrow angular range.
10.8.3
Liquid crystals 293

Polymer dispersed liquid crystals (PDLC)
Figure 10.28 shows the structure of a PDLC, in which bubbles of a nematic liquid crystal are dispersed randomly and densely within an isotropic polymer matrix. The bubbles are a few microns in diameter. Sheets of such structured material are now used to make windows which can be switched from clear to opaque in some tens of milliseconds. For this application the materials are selected to make the ordinary refractive index, no , of the uniaxial nematic liquid identical to the refractive index of the polymer. Both faces of the PDLC sheet are coated with transparent electrodes. As shown in the right hand panel, when a voltage is applied across the sheet it aligns the liquid crystal molecules perpendicular to the surfaces. Then light at normal incidence, irrespective of its polarization, enters a material with a single refractive index, no . The window is transparent. When the applied ﬁeld is removed the molecules in each bubble take up orientations that are inﬂuenced by their interac9 LCDs now share projection applications with microelectromechanical devices. These consist of an array of mirrors about 15 µm square each hinged on a separate pillar. An applied voltage tilts any mirror which is addressed through 10◦ by electrostatic attraction and this mirror then directs light into the projector lens. Intensity is controlled by dithering the mirror at a frequency of around 1 kHz so that it reﬂects light into the lens for a controlled fraction of the frame duration.
Field off
+
Field on
Fig. 10.28 Polymer dispersed liquid crystal. In the left hand panel the bubbles’ refractive index is diﬀerent from that of the polymer. In the right hand panel the refractive indices are the same for light incident normally.
294 Polarization
tion with molecules of the polymer surface and their optic axes become isotropically distributed. Thus the refractive index of the bubbles becomes n2 = (n2e + 2n2o )/3, which diﬀers markedly from the polymer refractive index. In this state the densely packed bubbles in the PDLC window scatter the incident light so eﬀectively that the surface looks uniformly dull.
10.8.4
Nematic liquid crystals have a switching rate which, though more than adequate for visual displays, is not fast enough to be of much use in the spatial light modulators mentioned in Section 7.8.1. However another class of liquid crystals, SmC∗ , can be switched about one thousand times faster. Here the star indicates that the molecules have a deﬁnite handedness: they are chiral. The molecules possess an intrinsic electric dipole moment perpendicular to both the layer normal and the director direction. As shown in ﬁgure 10.29 the electric dipoles in each layer lie parallel and a layer forms a ferroelectric domain. In one layer the director makes a ﬁxed angle with the layer normal, θ. However the chirality of the molecules causes the director direction to twist between layers so that the director direction for successive layers rotates slowly around the cone of semiangle θ shown in ﬁgure 10.29. The director direction completes a full circuit around the cone after a few thousand layers.
Director
Layer
Ferroelectric liquid crystals (FELC)
θ
normal
Dipole
Fig. 10.29 Orientation of the director and electric dipole moment in a layered SmC* liquid crystal.
A ferroelectric LCD cell is shown face on in ﬁgure 10.30. The thickness of the layer of liquid crystal is reduced to about 2 µm so that the directors are forced into one of the two orientations, OA and OB, 2θ apart. In this surface stabilized state all the molecules can be simultaneously aligned in one of these two alternative orientations by applying the electric ﬁeld in the directions indicated. Switching is rapid because the applied ﬁeld acts on the electric dipole moment of each domain, P, giving a very high torque, P ∧ E. Another useful property of the surface stabilized regime is that once switched the alignment is stable and consequently power is only required to switch, and not to maintain the state of each pixel. Crossed polarizers sandwich the FELC. In the upper panel of ﬁgure 10.30 the directors are aligned along OA so this is also the optic axis. The incident light has extraordinary polarization in the liquid crystal and is transmitted without change of polarization. It is therefore extinguished by the analyser. In the lower panel, with the electric feld reversed, the directors now align parallel to OB and this is now the optic axis. It follows that the incident light can be resolved into a component polarized along OB, and a component polarized at right angles along OC. The former has extraordinary polarization and the latter has ordinary polarization. Their electric ﬁelds on entering the liquid crystal
10.9
Further reading 295
are Ee = cos 2θ; Eo = sin 2θ. After transmission, ignoring common factors, their values become
Analyzer
Ee = cos 2θ; Eo = sin 2θ exp (iφ), Eapplied
where φ is the phase diﬀerence arising from the diﬀerence between the ordinary and extraordinary refractive indices, ∆n. We can now reuse the analysis that led to eqn. 10.54. This shows that the time averaged intensity passing through the analyser is
O
Director
A
Polarizer
Layer normal
I A = sin2 4θ sin2 φ/2 = sin2 4θ sin2 (π∆nd/λ), where λ is the wavelength in air, ∆n is the diﬀerence in the refractive indices and d is the thickness of the liquid crystal. The ﬁrst factor in this equation is largest when θ is 22.5◦ , which is fortunately within the range of angles available with SmC* liquid crystals. The second factor is maximal when d = λ/2∆n. Although ideal in their role in spatial light modulators, the FELC but are not as yet competitors for the display market.
10.9
Further reading
Polarization of Light by S. Huard, published by John Wiley and Sons (1997). A comprehensive modern text on polarization including the Mueller/Stokes formalism and a detailed analysis of devices such as modulators. Modern optics by R. D. Guenther published by John Wiley and Sons (1990) contains a comprehensive and comprehensible description of the numerous ellipsoids used to describe the properties of birefringent materials.
Analyzer
Eapplied
Layer normal
B Director 2θ
O
Polarizer
C
Fig. 10.30 Surface stabilized ferroelectric liquid crystal cell, seen face on. The alignments of the directors are shown for alternative signs of the applied electric ﬁeld across the cell.
Exercises (10.1) (a) What is the Jones matrix for a linear polarizer with its transmission axis at 45◦ to the xaxis? Can you interpret the result? (b) What is the Jones matrix for a layer of isotropic absorber which reduces the incident amplitude by a factor t?
gle, 2.5◦ , instead of 45◦ . These two prism are also glued together. What is the phase delay between the two polarizations when the thicknesses of the two prisms where the beams cross them are d1 and d2 ? What useful property does this Babinet compensator possess?
(10.2) What is the instantaneous Poynting vector for the rightcircularly polarized wave whose electric ﬁeld (10.4) Show that the Wollaston prism in ﬁgure 10.12 with is given by eqn. 10.4? θ = 45◦ separates the two beams by 20◦ .
(10.3) Consider a version of the Wollaston prism with the (10.5) An unpolarized light beam of intensity I0 is incicomponent prisms being cut with a very acute andent perpendicularly on two Polaroid sheets in se
296 Polarization ries. These are rotated in their own planes about wavelengths in calcite of the ordinary and extraorthe beam as axis. One rotates anticlockwise, the dinary waves? What fraction of each polarization other clockwise, both at angular frquency ω. What component enters the crystal? is the intensity variation with time? At what fre (10.9) Resolve the elliptically polarized wave in eqn. 10.7, quency does the polarization vector of the light when φ is π/2, into a circularly and a plane polartransmitted rotate? ized component. Write the result in terms of Jones vectors of this complex electric ﬁeld. (10.6) Determine over what range of input frequency the modulator described in Section 10.7.1 would be (10.10) Show that the angle, α, between the electric ﬁeld able to convert an electrical to an optical signal and electric displacement of the extraordinary wave without distortion. What are the corresponding is given by limits on the swing of the signal voltage around Vπ such that the response remains linear to better cos α = [n2e cos2 θ+n2o sin2 θ]/ n4e cos2 θ + n4o sin2 θ. than 0.3%? This is also the angle between the Poynting vec(10.7) Calculate the magnetic ﬁeld required to produce a ◦ tor and the wave vector, and is called the Poynting 45 rotation of the plane of polarization of light of vector walkoﬀ. Use eqns. 10.24 and 10.25; then 633 nm wavelength in a Faraday rotator made from cos α = Ee · De /(De Ee ). a 2 cm long terbium doped glass rod. If the rotator is then used with light of wavelength 533 nm in (10.11) Prove eqn. 10.23 by writing the x and zthe isolator shown in ﬁgure 10.20 what fraction of components of eqn. 10.20 in terms of the refractive a reﬂected beam would penetrate the isolator? indices. (10.8) A laser beam of wavelength 589 nm in free space (10.12) Sketch the index ellipsoid for a positive uniaxial crystal. Hence check the sign change in eqn. 10.29 is at normal incidence on a calcite plate whose opbetween negative and positive uniaxial crystals. tic axis is parallel to the surface. What are the
Scattering, absorption and dispersion 11.1
Introduction
Light travelling through matter interacts with the atoms and molecules, which results in three familiar eﬀects: absorption, scattering and dispersion. These eﬀects and the connections between them are described in this chapter and interpreted using the classical theory of how electromagnetic waves interact with matter. This approach provides many insights that, after reinterpretation, retain their value when quantum theory is developed in later chapters. In the sections immediately below this one scattering will be discussed. Scattering from electrically polarizable particles with dimensions very much smaller than the wavelength of the radiation is called Rayleigh scattering. The blue colour of a clear sky is one consequence of Rayleigh scattering, in this case of sunlight scattered from molecules in the atmosphere. It will emerge that coherent Rayleigh scattering is the underlying process in slowing light down within transparent materials . Mie scattering is the name applied to scattering of light from larger particles whose size may be anywhere from about one tenth to a hundred times the wavelength. Interference eﬀects between light scattered from diﬀerent parts of a scatterer now come into play. Short sections are used to introduce absorption and to point out correlations between absorption and dispersion that are due to atomic or molecular processes. Then the classical theory of dispersion and absorption in dielectrics based on atomic oscillators is outlined. Following this the optical properties of metals are interpreted using a classical model in which the conduction electrons are regarded as free within a metal. This model explains the shallow penetration of electromagnetic waves into metals at optical frequencies, and the accompanying high reﬂectance of metals. Free electrons undergo plasma oscillations at a frequency somewhere in the ultraviolet, which depends on the electron density in the metal. The eﬀects of such oscillations on em wave propagation are also interpreted with the model. Diﬃculties over the deﬁnition of the velocity of electromagnetic waves, and the way these diﬃculties are resolved are the topics considered in
11
298 Scattering, absorption and dispersion
the next section of the chapter. For example, at wavelengths close to the centres of sharp spectral lines the velocity of an inﬁnite sinusoidal plane wave can exceed c. However when the velocity of light is deﬁned as the velocity of energy and information transfer by wavepackets rather than that of idealized inﬁnite waves there is no violation of the postulates of the special theory of relativity. A ﬁnal section describes the excitation of surface plasma waves at the boundary between a metal and a dielectric, and their recent use in biological sensors.
11.2
Rayleigh scattering
The scattering of sunlight from gas molecules in the upper atmosphere is responsible for the blue of the sky. In addition if the blue sky is viewed through a Polaroid, looking at right angles to the direction linking the observer to the Sun, the light is found to be strongly polarized. Scattering, as in this example, in which the particles doing the scattering are much smaller than the wavelength of the radiation is called Rayleigh scattering. Whenever this is the case the electric ﬁeld is uniform across each individual scatterer, taken here to be a gas molecule. An instantaneous dipole moment is induced in the molecule of magnitude p(t) = αε0 E0 exp (iωt),
Sun Maximal scattering
(11.1)
where E = E0 exp (iωt) is the electric ﬁeld of the wave and α is the polarizability of the molecule. Each dipole, thus excited, radiates at the frequency of the incident light and in this way scatters light out of the incident beam. Using eqn. 9.58 the time averaged power scattered by one molecule is W = ω 4 α2 ε0 E02 /(12πc3 ). (11.2) If there are N molecules per unit volume the power scattered out from unit area of the beam in a small distance dz along the beam is N W dz. The incident power over unit area of the beam is ε0 cE02 /2 so that the fractional power loss is
Viewer Sun Minimum scattering
Viewer
Fig. 11.1 Rayleigh scattering at 90 ◦ . In the upper panel light is polarized perpendicular to the scattering plane, and in the lower panel it is polarized in the scattering plane.
dW/W = −[N ω 4 α2 /(6πc4 )] dz = −(8/3)π 3 α2 N dz/λ4 .
(11.3)
The strong dependence on wavelength causes blue light to be scattered about ten times more eﬀectively than red light, which accounts for the blue of a clear sky. It also explains why light from the setting Sun, which has a long path through the atmosphere, should look red. Figure 11.1 has a viewer looking at the blue sky in a direction at right angles to the line passing through the Sun. Any molecule scattering light ﬁrst absorbs the light, becoming polarized in essentially the same direction as the light absorbed, and then reemits light. The angular distribution of the scattered light can then be inferred from ﬁgure 9.7. When the molecule is excited by light polarized perpendicular to the plane of scattering, as seen in the upper panel of ﬁgure 11.1, the observer is viewing
11.2
in a direction for which the intensity is at a maximum. On the other hand, if the light is polarized in the plane of scattering, as illustrated in the lower panel of ﬁgure 11.1, then the molecular dipole points towards the observer and no light would be seen. However there can be some misalignment of the molecular dipole axis with the electric ﬁeld inducing it due to asymmetry of the structure of the molecule, and then a little light would be seen. In any case the light received is strongly polarized perpendicular to the plane of scattering. Away from this viewing direction the polarization falls oﬀ rapidly. The expression for the fractional power loss in eqn. 11.3 can be brought into a form that contains the refractive index instead of the molecular polarizability. Firstly using eqns. 9.4 and 9.5 εr = 1 + N p/E = 1 + N α.
(11.4)
Specializing to gases at low pressures, the relative permittivities are close to unity, so that to a good approximation εr − 1 = 2(n − 1). Replacing εr in the previous equation gives α = 2(n − 1)/N. Finally replacing α in eqn. 11.3 yields dW/W = −(32/3)π 3 (n − 1)2 dz/(N λ4 ),
(11.5)
and after integrating over the path length z, W (z) = W (0) exp (−βz),
(11.6)
where β = (32/3)π 3 (n−1)2 /N λ4 . The distance, 1/β, in which the intensity falls by a factor e is called the attenuation length. Light of 500 nm wavelength travelling in unpolluted air at sea level has an attenuation length of 65 km: that is a power loss in air of 12 Mm−1 (parts per million in one metre). The scattering from an individual dielectric sphere of radius a and refractive index n in air is calculated in exercise 11.10 below. A quantity called the crosssection is now deﬁned as the total scattered ﬂux divided by the incident ﬂux per unit area of the incoming beam. It is thus the equivalent area from which light is removed by the scattering sphere: σ = (8π/3)(2π/λ)4 a6 G2 ,
(11.7)
where G = (n2 − 1)/(n2 + 2). The intensity of Rayleigh scattering therefore increases with the cube of the geometric area of the scatterer, as well as falling oﬀ with the fourth power of the wavelength.
Rayleigh scattering 299
300 Scattering, absorption and dispersion
11.2.1
Coherent scattering
The scattering from the atoms or molecules in condensed matter is now examined and this will bring out the close connection between coherent scattering and refraction. Figure 11.2 shows a large area, very thin, ﬂat sheet of some transparent dielectric of thickness s on which plane sinusoidal waves are arriving at normal incidence. The electric ﬁelds felt by all the molecules such as that labelled M are identical. Consequently their induced dipole moments are all in phase and can be written p(t) = αε0 E0 exp (iωt). These dipoles radiate at the same frequency as the incident radiation. In order to calculate the electric ﬁeld at P due to scattered light the dielectric layer is divided into Fresnel zones centred on O. Then applying a result proved in Section 6.12, the total amplitude at P due to scattered radiation from the whole sheet is equal to half that produced by the ﬁrst Fresnel zone. Now the ﬁeld at P due to a dipole located on axis at O is given by eqn. 9.55 e(ωt − kz) = αω 2 E0 exp [i(ωt − kz)]/(4πc2 z) M P
O z s
Plane wavefront
Fig. 11.2 Coherent scattering from a thin plane layer of dielectric.
pointing parallel to electric vector of the incident light. From eqn. 6.55 the surface area of the ﬁrst zone is πλz, and hence the number of dipoles within this zone is Nd = πzλ(N s) = 2π 2 zN s/k, where N is the number density of the molecules. In calculating the total electric ﬁeld at P it is important to recall that the phasor formed by the electric ﬁelds in the ﬁrst Fresnel zone turns through a half circle. The resultant electric ﬁeld is therefore 2/π times the direct sum of that due to these dipoles, and is π/2 out of phase with the ﬁeld produced by a dipole located on axis. After taking these factors into account the electric ﬁeld at P due to scattered radiation is Es = (1/2)(2/π)Nd e(ωt − kz − π/2) = (2πzN s/k)αω 2 E0 exp [i(ωt − kz − π/2)]/(4πc2 z) = (N ksα/2)E0 exp [i(ωt − kz − π/2)]. Thus the total electric ﬁeld at P made up of the electric ﬁelds of the scattered and unscattered light is Ep = E0 exp [i(ωt − kz)] + Es = E0 exp [i(ωt − kz)] [ 1 − iN ksα/2 ] ≈ E0 exp [i(ωt − kz)] exp [−iN ksα/2 ],
(11.8)
where the approximation is valid when s is suﬃciently small. The eﬀect of scattering is therefore equivalent to increasing the path length in the
11.3
dielectric from s to s + N sα/2, that is by a factor which we can identify as the refractive index of the dielectric n = 1 + N α/2.
(11.9)
In turn the relative permittivity is εr = (1 + N α/2)2 ≈ 1 + N α, which agrees with eqn. 11.4. What emerges from this analysis is that the refractive index of dielectrics has its origin in the interference of coherently scattered with unscattered light. The following picture is helpful. Light travels at its free space velocity c in the open spaces between atoms/molecules; scattering produces a wave of small amplitude which is delayed in phase by π/2 with respect to the direct wave; the resultant wave is little changed in magnitude compared to the incident wave, but suﬀers a phase shift proportional to the scattering amplitude. Referring back to the previous section it can be seen that the scattering from a layer of gas was handled as if each molecule scattered incoherently from every other molecule. In direct contrast, in this section scattering from condensed matter has been treated as coherent. The reason for the diﬀerence in approach is that the scattering centres must be densely and uniformly distributed if the scattering is to be fully coherent. This is not the case in a gas, but it is nearly correct in a transparent solid like glass. If there were such a thing as a perfectly homogeneous material then the only consequence of Rayleigh scattering would be to give a refractive index diﬀerent from unity. A simple argument is used here to make it plausible that the Rayleigh scattering from a gas can also be interpreted as scattering from the density ﬂuctuations in the gas. Suppose that a unit volume of gas is divided into many equal volumes v which are small enough so that within each volume the molecules feel nearly the same electric ﬁeld. Now the number of molecules per unit volume in a gas follows a Poissonian distribution, so that each √ cell would contain on average N v molecules with an rms deviation N v. A simple √ view to take is that there are 1/v inhomogeneities, each containing N v molecules, and that each inhomogeneity scatters incoherently of all the others. This √ yields a total intensity proportional to ( N v)2 /v = N , which is exactly the intensity expected from the scattering oﬀ N incoherent scatterers!
11.3
Mie scattering
When scattering takes place from larger particles whose size approaches the wavelength of light the analysis of scattering from a single particle becomes complicated. There is interference between light scattered from diﬀerent parts of the same scatterer, and the phase delays between light
Mie scattering 301
302 Scattering, absorption and dispersion
travelling through diﬀerent thicknesses of the scatterer must be taken into account. The ﬁrst detailed study for spherical scatterers was made by Mie in 1908 and his analysis covered scattering from spheres of all diameters. The Rayleigh treatment is an adequate approximation for diameters up to roughly one tenth of the wavelength of the radiation, while at diameters greater than one hundred times the wavelength of the radiation the ray theory is an adequate approximation. Scattering from spheres in the intermediate range where neither approximation works is thus generally called Mie scattering. Only the salient features of Mie scattering from dielectric spheres will now be discussed. The interference between diﬀerent regions causes the angular distribution to acquire lobes that are in essence diﬀraction patterns. As the diameter of the sphere increases the overall angular distribution gradually becomes more forward peaked with a lesser backward peak. When the sphere diameter is λ/4 (5λ) the ratio of the scattering intensity in the forward direction with respect to that in the backward direction is around 2.5:1 (2000:1). The total amount of scattering is expressed as an equivalent area from which the scattering sphere removes light from the beam: the crosssection per sphere. The rapid rise in crosssection with radius seen in Rayleigh scattering tails oﬀ and settles down to being proportional to the radius squared, and hence to the area of the scatterer, at radii a few times λ/n, where n is the refractive index of the scatterer. The strong dependence of the scattering intensity on the wavelength characteristic of Rayleigh scattering changes to a ﬂatter dependence so that the light scattered from clouds, fog and aerosols, in which the droplet diameters are typically 10 µm, is white. White paints contain a clear polymer matrix loaded with transparent particles of a very high refractive index material, titanium dioxide; these particles having dimensions of around half the wavelength of light. The Mie scattering is intense because of the big diﬀerence between the refractive indices of the polymer (∼ 1.5) and titanium dioxide (∼ 2.76 in one crystalline form), and is nearly constant across the visible spectrum. Thus the paint, or rather the backscattered ambient light, looks white. The scattering is suﬃciently strong that a thin layer of paint can easily mask the colour of the underlying surface. A related eﬀect was described in Section 10.8.3. Micronsized liquid crystal droplets dispersed in a clear polymer matrix scatter light so eﬀectively that light passing through is totally diﬀused. However when an electric ﬁeld is applied the refractive index of the droplets becomes equal to that of the polymer and the panel is rendered transparent.
1
K.R. Weninger, B.P. Barber and S.J. Putterman, Physical Review Letters volume 78, page 1799 (1997).
In a recent application the diameter of a sonoluminescent bubble has been tracked by Mie scattering during the rapid oscillations in which the diameter changes from ten microns to a fraction of a micron in one microsecond.1 What was done was to measure the light scattered into a large solid angle centred on a scattering angle of 60◦ when the bubble was
11.4
Absorption 303
illuminated by a series of short pulses of light. The pulse duration was 0.2 ps and the repetition rate 76 MHz. In the angular range selected the integrated scattered intensity is proportional to the bubble area which provided a simple way to estimate the bubble diameter at each exposure.
11.4
Absorption
Light can be lost from a beam by absorption as well as scattering. Absorption of light takes place on electrons bound in atoms or molecules, and on free electrons in metals. When white light passes through a gas containing atoms of a single element, and the spectrum of the emerging light is viewed with a grating spectrometer, the spectrum is seen to be marked by multiple thin dark lines. These lines mark the individual narrow wavelength ranges at which absorption is strong and are characteristic of the atomic structure of the element.2 The atoms may promptly reemit the radiation or promptly emit radiation at a longer wavelength. This radiation is known as ﬂuorescence, of which radiation at the same wavelength as the incident radiation is called resonance ﬂuorescence. Phosphorescence is the term applied if there is a delay of greater than a microsecond before the secondary radiation emerges. In denser materials the absorption occurs over broad bands of wavelength rather than narrow lines, and of course some materials absorb all the light. The energy absorbed is generally converted through atomic processses and collisions to heat. Scattering of light can therefore be interpreted as the absorption of light in which the absorption is followed by prompt reradiation at the same wavelength.3 A beam of light traversing matter loses intensity through scattering out of the beam and through absorption. Coeﬃcients of absorption, βa , and scattering, βs , can be deﬁned such that in a thin layer of thickness dz the beam loses fractions of its intensity βa dz and βs dz through these two processes. Then the attenuation of a narrow beam over a distance z is given by I(z) = I(0) exp [−(βs + βa )z]. (11.10) With a broad beam the scattering from one part of the beam to another must be allowed for. At one extreme, in metals, the intensity of light falls by a factor e in a few nanometres while in normal window glass this distance is 0.3 m. Materials with an open structure such as plant leaves acquire their colour through selective absorption of the light which enters and is multiply scattered within the structure. Photosynthesis is the absorption process and what emerges is the unused green light. By contrast the colour of metals is due to the strong reﬂection from the surface. Gold appears reddish because it absorbs more strongly at the red end of the spectrum and reﬂects red light more eﬀectively. How it comes about that strong reﬂection is associated with strong absorption in this case poses a quandary that is resolved in Section 11.6.
2
The fact that light restricted to a very narrow range of wavelengths forms a line in the image plane of a spectrometer has led to the light itself being called a spectral line.
3
Raman and Brillouin scattering, which are processes that become signiﬁcant at high intensities of illumination, will be considered later after quantum theory and lasers have been introduced.
304 Scattering, absorption and dispersion
11.5
Dispersion and absorption
Refractive index
Dispersion and absorption are processes that share a common origin in atomic processes. Figure 11.3 shows schematically the variation of the
Visible spectrum
1.0
Absorption coefficient
Frequency
Frequency
Fig. 11.3 Schematic plot of the refractive index and absorption coeﬃcient for a material transparent across the visible spectrum.
refractive index and the absorption coeﬃcient of a transparent material across the infrared to ultraviolet part of the spectrum. There are clear pairings between the absorption peaks and characteristically shaped oscillations in the value of the refractive index. In general many such pairs may be seen for a dielectric. Such sharp absorption peaks account for the lines seen in the absorption spectrum of a gas. The equivalent distributions for materials like glass or water show similar correlations, but with broader lines and additional features due to the mutual interaction of the densely packed atoms. Each oscillation/peak pair appearing in ﬁgure 11.3 is due to a process in which energy is absorbed by atoms or molecules. Excitation of electrons in atoms happens through the absorption of ultraviolet and visible light, while molecular vibrations and rotations can be excited by the absorption of visible and infrared radiation down to microwaves. Transparent materials are transparent because for these materials the electronic absorptions occur in the ultraviolet and the molecular excitations in the infrared with none lying within the visible spectrum. Regions where the refractive index rises with increasing frequency (decreasing wavelength)
11.5
are said to have normal dispersion and regions where the refractive index falls as the frequency increases are said to have anomalous dispersion. Common glass therefore has normal dispersion, as can be seen in ﬁgure 1.16. A simple model of the atom as a classical oscillator will be used in the following sections to account for the shapes and correlations between the absorption and dispersion features illustrated in ﬁgure 11.3. The quantum interpretation which is presented in later chapters provides a deeper and quantitative explanation of atomic transitions. What the classical view provides are insights that do not lose their value when one uses quantum theory.
11.5.1
The atomic oscillator model
The process considered is the forced oscillation of an electron in an atom caused by the electric ﬁeld of an electromagnetic wave. Electron velocities are tiny compared to c; thus the magnetic force ev∧B, of order evE/c, is negligible compared to the electric force eE. There are two other forces acting on an electron besides the electric force due to the electromagnetic wave. The ﬁrst is the restoring electric force exerted by the stationary nucleus. Compared to the electric ﬁeld of a nucleus at an electron, of order 1011 V m−1 , the electric ﬁeld in any beam other than a high energy pulsed laser is very small. Thus the electron displacement is small compared to the atomic size and the restoring force will be linear in the displacement. The second force is an equivalent damping force used to represent the eﬀect of all the processes which dissipate the energy absorbed by the electron. We shall see later that these include the reradiation of the energy and processes such as atom–atom collisions in a gas. The equation of motion of the electron in this model is m(d2 x/dt2 ) = −eE0 exp (iωt) − ξ(dx/dt) − κx.
(11.11)
Here −e and m are the electron charge and mass respectively; the incident electromagnetic wave’s electric ﬁeld, E = E0 exp (iωt), points in the xdirection; ξ is the damping constant and −κx is the restoring force. In the absence of any electromagnetic wave the electron would undergo damped harmonic motion with natural angular frequency, ω0 , given by ω02 = κ/m. Then, putting γ = ξ/m, the equation of motion can be rewritten −eE0 exp (iωt) = m[(d2 x/dt2 ) + γ(dx/dt) + ω02 x].
(11.12)
The forced motion of the electron will have the same frequency as the driving electric ﬁeld x = x0 exp (iωt), (11.13) which when substituted in eqn. 11.12 gives −eE0 = m(−ω 2 + iγω + ω02 )x0 ,
Dispersion and absorption 305
306 Scattering, absorption and dispersion
after the common factor exp (iωt) has been cancelled. Whence x0 = (−eE0 /m)/(ω02 − ω 2 + iγω).
(11.14)
The displacement of the electron from its equilibrium position within the atom means that the atom acquires an electric dipole moment p = −ex0 exp (iωt), and the atomic polarizability is then α = p/(ε0 E) = (e2 /ε0 m)/(ω02 − ω 2 + iγω). The previous analysis of Rayleigh scattering only applies well away from resonances and it also ignores damping. Then α is eﬀectively e2 /ε0 κ.
(11.15)
The polarization of the material, that is the electric dipole moment per unit volume, is thus P = N αε0 E, (11.16) where N is the number of atoms per unit volume. Using eqn. 11.4 εr = 1 + P/(ε0 E) = 1 + N α.
(11.17)
Substituting for α from eqn. 11.15 in the last line gives εr = 1 + (N e2 /mε0 )/(ω02 − ω 2 + iγω) = 1 + ωp2 /(ω02 − ω 2 + iγω),
(11.18)
where ωp is called, in anticipation, the plasma angular frequency, and is given by ωp2 = N e2 /(mε0 ). (11.19) This is a resonant response which is strongest when the incident radiation has a frequency close to the natural oscillation frequency of the electron in the atom. Close to an individual resonance we can take ω ≈ ω0 . Then eqn. 11.18 simpliﬁes to εr = 1 + (ωp2 /ω0 )/ [ 2(ω0 − ω) + iγ].
(11.20)
The analysis is now continued for a gas, so that the relative permittivity is close to unity. Then the expression for a refractive index n is to a good approximation given by n=
√ εr = 1 + (ωp2 /4ω0 )/[ ω0 − ω + iγ/2 ].
(11.21)
The refractive index therefore has a real and an imaginary part n = nr − ini ,
(11.22)
where ni is both real and positive. Referring back to the analysis of TIR given in Chapter 9 it can be seen that if the refractive index has
11.5
an imaginary part this means that the em wave is attenuated.4 After travelling a distance z the electric ﬁeld becomes E(z) = E(0) exp [iω(t − nz/c)] = E(0) exp [iω(t − nr z/c)] exp (−ni zω/c). Therefore the intensity falls oﬀ exponentially, I(z) = I(0) exp (−2ni ωz/c),
(11.23)
giving an absorption coeﬃcient βa = 2ni ω/c proportional to ni . The real part nr must be identiﬁed as the standard refractive index which could be determined by measuring the phase delay of light passing through the dielectric. Separating the real and imaginary parts of eqn. 11.21 gives nr = 1 + ωp2 (ω0 − ω)/(4ω0 R), ni = γ
ωp2 /(8ω0 R),
(11.24) (11.25)
where R = (ω0 − ω)2 + γ 2 /4. This result conﬁrms that the Lorentzian line shape introduced in Chapter 7 is the expected atomic line shape. Figure 11.4 shows this predicted variation of the real and imaginary parts of n as a function of frequency around the natural frequency of oscillation of the electron. The vertical scale shown is arbitrary and in the case of a gas the numbers shown for nr − 1 and ni would be scaled down by a factor of order 1000. This response reproduces both the shapes and correlations between the features present in ﬁgure 11.3: each absorption peak is accompanied by a characteristically shaped rapid variation of the refractive index involving a region of anomalous dispersion. The responses of all the electrons in an atom have to be taken into account. Then the expression for the real and imaginary parts of the refractive index become nr = 1 + ωp2 [fi (ω0i − ω)/(4ω0i Ri )], ni = ωp2
i
[fi γi /(8ω0i Ri )].
i
The coeﬃcient fi is called the oscillator strength of the ith electron. Such quantities can only be calculated using the quantum theory of atomic structure and its interaction with electromagnetic radiation. 4 If the choice of complex waves were exp [i(kz − ωt)] rather than exp [i(ωt − kz)] then the sign in front of ini in eqn. 11.22 would need to be positive. In addition +iγ would be replaced by −iγ in the preceding equations.
Dispersion and absorption 307
308 Scattering, absorption and dispersion
2.5 (ω0  γ /2)
2
(ω0 + γ /2)
nr 1 and n i
1.5 1
0.5 0
0.5 1 0.6
0.8 1 1.2 Angular frequency / ω0
1.4
Fig. 11.4 Predictions of the real and the imaginary parts of the refractive index around resonance. The real part is indicated with a solid line and the imaginary part with a broken line. Note that (nr − 1) is plotted rather than nr . The vertical scale is arbitrary.
Consider the case when the damping is weak enough that γ ω0 . Then the peak absorption is at the angular frequency ω0 : the peak value of ni is ωp2 /(2γω0 ), and the width between the angular frequencies at which ni falls to half its peak value is γ. Thus if the damping at the atomic level is weak the peak resonant absorption is large, the absorption line is narrow and the anomalous dispersion is more pronounced. The assumption has been made implicitly in the preceding analysis that the electric ﬁeld felt by the atoms is spatially uniform across the dielectric. In fact the local ﬁeld felt by the atom will be the vector sum of this ﬁeld and that due to the surrounding atoms. It can be shown that, when the dielectric is a nonpolar liquid or a cubic crystalline material, the local electric ﬁeld has a simple relationship to the applied ﬁeld5 Elocal = E + P/(3ε0 ).
(11.26)
Then eqn. 11.16 becomes P = N α ε0 Elocal = N α ε0 [E + P/(3ε0 )], 5 See for example Chapter 2 of the 7th edition of Optics by M. Born and E Wolf, published by Cambridge University Press (1999). For other materials the numerical factor is diﬀerent from 1/3.
11.6
Absorption by, and reﬂection oﬀ metals
309
and substituting χε0 E for P gives the Clausius–Mossotti relation N α = χ/(1 + χ/3) = 3(n2 − 1)/(n2 + 2).
(11.27)
With this replacement the expression for the refractive index becomes (n2 − 1)/(n2 + 2) = (ωp2 /3) [(fi /ω0i )/(2(ω0i − ω) + iγi )]. (11.28) i
In the limit that the refractive index is close to unity the left hand side reduces to (n − 1)/3, so that the equation collapses to that applicable to a gas. The mean separation of atoms is smaller by a factor ten in solids, liquids and high pressure gases than it is in gases at normal temperature and pressure (NTP, 20◦ C and 105 Pa). Thanks to this denser packing the interactions between atoms in liquids and in solids is stronger and changes the absorption spectra a great deal. The narrow spectral lines observed with gases at NTP are replaced by broad bands. Excitation of molecules can take the form of vibrations in which the distance between the component nuclei oscillates, or of rotations of the molecule. The masses vibrating are now the nuclei so that the natural frequencies are correspondingly smaller and lie predominantly in the infrared/microwave region of the spectrum. A further feature of many materials is the presence of molecular structures which possess permanent electric dipole moments. An applied electric ﬁeld will cause these dipoles to align in the ﬁeld direction, which makes another important contribution to the polarization. This motion is heavily damped in liquids and solids so that at optical frequencies this contribution to the relative permittivity is smoothly and slowly falling with increasing frequency. The glasses from which so many optical components are made are transparent to visible radiation because the atomic resonances lie in the ultraviolet and the molecular resonances in the infrared. Across the visible spectrum lying between these resonances the refractive indices of glasses shown in ﬁgure 1.16 therefore have normal dispersion and are well ﬁtted by the various Sellmeier formulae, such as that given in eqn. 1.27. The poles of the terms, where their value diverges, give a good indication of the location in wavelength of the ultraviolet resonances. A clear connection between absorption and dispersion has emerged from the simple classical model used here to explain the interaction of electromagnetic radiation with matter. The connection is not restricted to this particular model: there is a deeper linkage between them which originates in the causal nature of physical processes.6
11.6
Absorption by, and reﬂection oﬀ metals
When an electromagnetic wave is incident on a metal it induces a current of free electrons. Energy transfered to the electrons is dissipated
6
A good account is given in Chapter 2 of the seventh edition of Classical Electrodynamics by J.D. Jackson, published by John Wiley and Sons, New York (1998).
310 Scattering, absorption and dispersion
in collisions with the lattice of positive ions, and appears as heat. In this way the part of the wave entering the metal is attenuated within a short distance. It was pointed out in Section 9.3 that the ratio of the conduction current to the current arising from polarization in a good conductor like copper is large up to optical frequencies. When modelling the propagation of electromagnetic waves at these frequencies in metals it is enough to take account of the response of the free electrons only. A classical model of this type developed by Drude will be used here to analyse the propagation of electromagnetic waves in a metal. In metals the least well bound electrons in each atom become detached from the parent atom and form a sea of free electrons. Travelling throughout the metal, their paths are punctuated by frequent, mostly inelastic, collisions with the lattice of positive ions, and these collisions serve to bring the ions and free electrons into thermal equilibrium. The eﬀect of a collision is also to randomize the electron’s direction after the collision. An applied constant electric ﬁeld E causes the electrons to acquire a drift velocity, v, in addition to their random motion; the total velocity is the vector sum of these two components. Newton’s second law applied to the electron drift motion gives mdv/dt = −eE, −e and m being the electron charge and mass respectively. However the electrons suﬀer collisions and after each collision the electron direction is random, which has the eﬀect of reducing the mean drift velocity after collision to zero. A measure of the mean drift velocity can be obtained by integrating the last equation over the mean time, τ , between collisions mv/τ = −eE.
(11.29)
This result shows that the collisions with the lattice have the same effect as a damping force mv/τ opposing the electric force. At room temperatures in metals such as aluminium, copper and silver τ is around 10−14 s giving a drift velocity of order 0.02 E m s−1 where the ﬁeld, E, is expressed in V m−1 . Therefore drift velocities are many orders of magnitude less than the velocities of the free electrons. Equation 11.18 for the relative permittivity of a dielectric can be easily adapted to become applicable to a metal by making two simple changes. Firstly the term arising from the restoring force exerted by the parent nucleus is deleted, and secondly the damping force per unit momentum, γ is 1/τ . The equation obtained in this way for the relative permittivity of a metal is εr = 1 + ωp2 /(−ω 2 + iω/τ ).
(11.30)
Any contribution from the electrons that remain bound in the atoms has been ignored here. As before ωp2 = N e2 /mε0 , with N now being the number density of free electrons in the metal. When the frequency of
11.6
Absorption by, and reﬂection oﬀ metals
the electromagnetic waves is well above 1/τ there are many cycles of the waves between one electron–lattice collision and the next, which means that damping becomes less and less important as the wave frequency rises. In the case of copper the number density of free electrons is around 8.5 1028 m−3 and the mean interval between collisions is approximately 2.4 10−14 s. An estimate of the plasma frequency in copper, based on these values, is 2.6 1015 Hz; the corresponding wavelength lies well into the ultraviolet. Using these values in eqn. 11.30 gives a value for the imaginary part of the refractive index in rough agreement with the observed value, but the real part is grossly underestimated. This indicates the limitations of the classical Drude model as far as quantitative predictions are concerned in the visible part of the spectrum. However the Drude model provides a consistent parametrization of the data at longer wavelengths. The intensity of light falls with distance in a metal like I(z) = I(0) exp (−βa z),
(11.31)
where the absorption coeﬃcient is given by βa = 2ωni /c. Thus the intensity falls oﬀ by a factor e in a distance called the skin depth s = c/(2ωni ) = 0.08λ/ni, where λ is the wavelength in air. The skin depth is approximately 18 nm for light of wavelength 589 nm falling on copper. Of course metals reﬂect well – which at ﬁrst sight seems hard to reconcile with the idea that they also absorb strongly. This point is now discussed. The reﬂection coeﬃcient for light falling at normal incidence on a dielectric is given in eqn. 9.79. In the case of reﬂection from metals where the refractive index can be complex this becomes R0 = (n1 − n2 )/(n1 + n2 ) 2 ,
(11.32)
so that at an air/metal interface R0 = [(nr − 1)2 + n2i ]/[(nr + 1)2 + n2i ].
(11.33)
It is clear that if the imaginary part of the refractive index is much larger than the real part, as is the case with metals, then the reﬂectance must be close to its upper limit of unity. The quandary mentioned above is resolved by noting that most of the incident light incident on a metal is reﬂected, while that small part that is transmitted is absorbed in a short distance within a metal. Figure 11.5 shows how the reﬂectances of copper, gold, silver and aluminium vary across the visible spectrum. Table 11.1 gives the values of the real and imaginary parts of the refractive index for several metals at a wavelength of 589 nm. Thin layers of
311
312 Scattering, absorption and dispersion
100 90 80
Al
Reflectance %
70 60
Cu
50 40
Au
30 20
Ag
10 0 200
300
400
500 600 700 Wavelength in nm.
800
900
Fig. 11.5 Reﬂectance oﬀ fresh metal surfaces at normal incidence. From Handbook of Optical Materials, edited by M.J. Weber and published by the CRC Press, Boca Raton, 2003. Courtesy Taylor and Francis Group, and Professor Weber.
Table 11.1 Table of refractive indices of metals at 589 nm wavelength. Metal
Real part
Imaginary part
Silver Copper Gold
0.18 0.62 0.47
3.64 2.63 2.83
silver and aluminium are deposited on glass to provide mirrors with high reﬂectance over a broad spectral range. Although multilayer dielectric coatings, which were described in Chapter 9, can give higher reﬂectance over comparable wavelength intervals these can only be deposited on items of small surface area. The primary mirrors of astronomical telescopes usually have an aluminium coating. Silver is not used despite its higher reﬂectance because it tarnishes quickly in air. Figure 11.6 shows the typical variation of the amplitude reﬂection coeﬃcients as a function of the angle of incidence for a metal. In the upper panel the curves bear a family resemblance to the corresponding ones for dielectrics shown in ﬁgure 9.12, but with larger values. Instead of the sharp change seen at Brewster’s angle for purely dielectric interfaces, the phase diﬀerence between rp and rs falls steadily from π at normal incidence to zero at grazing incidence and passes through π/2 at what is called the principal angle of incidence, θ i . The minimum of
11.6
Absorption by, and reﬂection oﬀ metals
the amplitude of the ppolarized reﬂected wave typically occurs a couple of degrees below the principal angle of incidence for metals, while the angle at which the ratio of p to spolarized amplitudes is least is only a fraction of a degree below the principal angle of incidence.
ρ = tan ψ exp(i∆),
(11.34)
where tan ψ = rp /rs  and ∆ is their phase diﬀerence. The quantities tan ψ and ∆ are measured using an ellipsometer of the sort shown in ﬁgure 11.7. A simple type of measurement involving nulling will be described. Monochromatic light is ﬁrst passed through a polarizer set at 45◦ to the plane of incidence to give s and ppolarization amplitudes which are equal and in phase. The compensator applies a phase diﬀerence between the s and ppolarized waves which will compensate (null) the phase diﬀerence occuring at reﬂection. After reﬂection these waves are therefore in phase but their amplitudes are reduced by the respective reﬂection coeﬃcients. Thus the light reaching the analyser is plane polarized but at an angle determined by the relative magnitude of the reﬂection coeﬃcients for s and ppolarization. This angle of polarization is determined by rotating the analyser so that the detector signal falls to zero. ∆ and ψ can be calculated from the the compensator and analyser settings, and from these the reﬂection coeﬃcients can be obtained by applying Fresnel’s laws. The relationships are straightforward but lead to complicated expressions. Results are at their simplest when the angle of incidence is made equal to the principal angle of incidence, θi . In this case the phase diﬀerence between the s and ppolarized components after reﬂection is π/2, so that the compensator can be a QWP. If the plane of polarization after reﬂection is at the angle ψ, then a good approximation for the refractive index of a metal surface is nr = tan θi sin θi cos 2ψ, ni = nr tan 2ψ. Measurements are made of tan ψ and ∆ at a range of angles of incidence to extract and check the complex refractive index. Variations on this technique are widely used in manufacturing and research to measure the thicknesses of single thin ﬁlms and stacks of thin ﬁlms of dielectrics where the refractive indices are already known and used as inputs. Reﬂections
s/TE
Reflectance
0.8 p/TM
0.6 0.4 0.2 0 0
Phase difference in degrees
The rapid extinction of light in a metal makes it impossible to measure the refractive index by using Snell’s law, because only the reﬂected light is available for measurement. It is not adequate to simply measure the reﬂectances of s and ppolarized light: their relative phase is needed as well in order to determine the complex refractive index. Suppose the incident light passes through a polarizer set at 45◦ to the plane of incidence. The light incident on the surface then has equal s and ppolarized components which are in phase. In general, after reﬂection the components diﬀer both in magnitude and in phase so that the reﬂected light is elliptically polarized. The ratio of the reﬂected amplitudes is
1
313
180 160 140 120 100 80 60 40 20 0 0
20 40 60 80 Angle of incidence in degrees
20 40 60 80 Angle of incidence in degrees
Fig. 11.6 The upper panel shows the reﬂectances from an air/gold interface for s and ppolarized light of wavelength 589 nm. The lower panel shows the phase diﬀerence between the reﬂected p and spolarized light. The dotted lines in both panels mark the principal angle of incidence, θi .
Detector Polarizer Analyzer QWP
θi
Metal surface
ψ
Fig. 11.7 Ellipsometer for determining the optical constants of metals. The successive polarization states are shown for a nulling measurement with light incident at the principal angle of incidence.
314 Scattering, absorption and dispersion
7
For more details see Spectroscopic Ellipsometry by H. Fujiwara, published by John Wiley and Sons, New Jersey (2007).
from successive interfaces must be included in the analysis, following the lines described in Chapter 9. A modern development is spectroscopic ellipsometry in which the parameters ψ and ∆ are measured for a range of wavelengths. Despite the complicated analysis required, ellipsometry gives results with very small errors because only angles are being measured, accurate to about 0.01◦, rather than relative intensities.7
11.6.1
Plasmas in metals
Plasmas are gases in which a proportion of the atoms or molecules are ionized into positive ions and electrons, and which are overall neutral. In its equilibrium state the spatial distributions of positive ions and electrons in a plasma are everywhere uniform and equal. One important example of a plasma is met in the upper atmosphere. The Sun’s ionizing ultraviolet radiation maintains layers of plasma, known collectively as the ionosphere, at heights between 30 and 200 km. The free electrons in metals also qualify as a plasma with a number density of order 10 28 m−3 compared to the much lower values, around 1011 m−3 , observed in the ionosphere. dz
Equilibrium z
Electrons displaced z+ξ
dz + dξ
Fig. 11.8 Equilibrium and displaced positions of the free electrons in a plasma.
Any disturbance of the electric ﬁeld causes the electrons to move at a much higher speed than the ions. This is because the electron mass is so much smaller, while the electrostatic force has the same magnitude for both an ion and an electron. Such a disturbance moves electrons out of a local region, while the positive ions are almost stationary. In this way the region acquires a net positive charge which attracts the displaced electrons. The displaced electrons then oscillate about their equilibrium position under this restoring force at the plasma frequency, ωp , previously deﬁned in eqn. 11.19. A simple analysis of plasma oscillations is made next before looking at their eﬀects on em wave propagation. In ﬁgure 11.8 the upper panel shows a plasma in equilibrium and the lower panel the eﬀect of displacing the electrons a distance ξ in the zdirection, where ξ varies in some as yet unspeciﬁed manner with z. The electron number density in the shaded region falls as a result of this movement by a factor dz/(dz + dξ). Thus the electron charge density within that region changes from −N e to −N e(1 − dξ/dz), N being the equilibrium electron number density. The net charge density in that region, including both the stationary positive ions as well as the electrons, changes from zero in equilibrium to ρ = N e(dξ/dz).
(11.35)
This induces an electric ﬁeld which is given by Gauss’s law ∇ · E = ρ/ε0 . In the case considered the ﬁeld points in the zdirection so this last equation becomes dE/dz = (N e/ε0 ) dξ/dz,
11.6
Absorption by, and reﬂection oﬀ metals
315
which when integrated gives E = N eξ/ε0 .
(11.36)
In turn this ﬁeld acts on each electron, and the equation of motion is md2 ξ/dt2 = −N e2 ξ/ε0 .
3
Referring back to eqn. 11.30, it follows that at frequencies high enough that the electron/lattice collisions are unimportant
2.5 nr and ni
This is the diﬀerential equation for simple harmonic motion, and the long term solution is ξ = ξ0 exp [i (N e2 /mε0 )t] = ξ0 exp [iωp t], (11.37) which justiﬁes the name plasma frequency for (N e2 /mε0 ).
2
nr
1.5 1
ni
0.5 0
0.2
0.4 0.6 Wavelength in µm
Fig. 11.9 Real and imaginary parts of the refractive index of sodium near the plasma frequency.
εr = 1 − ωp2 /ω 2 , which can be rewritten as n2 = 1 − ωp2 /ω 2 .
(11.38)
We therefore infer that at high enough frequencies the dominant eﬀect of em waves incident on a metal is to drive the electron plasma into forced oscillation. Figure 11.9 shows the behaviour of the real and imaginary parts of the refractive index of the alkali metal sodium around the plasma frequency predicted by eqn. 11.38. In the case of sodium the measured wavelength for the plasma oscillation, 210 nm, and that predicted using the free electron density, 209 nm, are in particularly good agreement and lie well in the ultraviolet. The onset of plasma oscillations has a profound eﬀect on the transmission of electromagnetic waves. Below the plasma frequency n2 is negative and hence the refractive index is purely imaginary. Then the reﬂection coeﬃcient given in eqn. 11.33 simpliﬁes: R0 ≈ (n2i + 1)/(n2i + 1) = 1.
(11.39)
Reﬂection is therefore very strong. Above the plasma frequency n2 is now positive and the refractive index is purely real and small, so that the reﬂection coeﬃcient is also small. In the case of the plasmas in the upper atmosphere the plasma frequency is around 3 MHz. Below this frequency short range radio transmissions can travel large distances round the Earth thanks to reﬂections from the plasma. Above the plasma frequency the plasma becomes transparent and for this reason communication with satellites is at frequencies well above 3 MHz. When a spacecraft reenters the Earth’s atmosphere the ionization produced around it is so intense that communication at all frequencies is lost. Attention must be drawn to two apparent violations of the theory of special relativity in the preceding analysis. First notice that above the
In ﬁgure 11.5 only silver’s reﬂectance shows any sign of a dip to mark the plasma frequency. Electrons are not free classical particles but inhabit quantized energy bands so that the predictions of the simple model need some modiﬁcation. See for example Chapters 4 onward in Optical Properties of Solids by F. Wooten, published by Academic Press (1972).
316 Scattering, absorption and dispersion
plasma frequency the refractive index is less than unity. Secondly if we look back at ﬁgure 11.4 it is seen that in the regions of anomalous dispersion the refractive index can also become less than unity. It seems that in these cases electromagnetic waves would travel faster than c, their velocity in free space! The apparent violation of the special theory of relativity in regions of anomalous dispersion was already noted by contemporaries of Einstein and raised as a fundamental objection to the theory. In order to resolve these inconsistencies more careful consideration must be given to what is meant by the velocity of light.
vg
11.6.2
Group and signal velocity
Up to this point the deﬁnition used for the velocity of light has been vp
vp = ω/k,
Coordinate
Fig. 11.10 Group and phase velocities of a wavepacket. There would be many thousands of individual wave oscillations inside a wavepacket from an actual light source.
6 5
ω/ωp
4 3 2 1 0 0
ω = kc 2
kc/ωp
4
6
Fig. 11.11 Angular frequency variation with the wave number above the plasma frequency ωp . The dispersion relation for free space is also shown.
where ω and k are the angular frequency and the wave number of an inﬁnite plane wave such as exp [i(ωt − kz)]. We shall call this velocity the phase velocity of light, as was done previously in Chapter 10. Inﬁnite waves which span all space and time and have a unique frequency are however never met in nature. Rather electromagnetic radiation consists of wavepackets which are of ﬁnite duration and extent and which can be resolved into a superposition of inﬁnite plane waves. There is even a problem in principle with measuring the velocity of a perfect sinusoidal wave. In order to recognize a place on the wavetrain for the purposes of timing its departure and arrival it must be marked in some way – which would change the wave from being a pure sinusoid. It therefore makes sense to consider the velocity of wavepackets as the true measure of the speed of light. The wavepacket peak in ﬁgure 11.10 indeed deﬁnes where the ﬁelds are large and where the Poynting vector is large. The form of the wavepacket is taken to have an electric ﬁeld E(z, t) = E(ω) exp [i(ωt − kz)]dω, centred around wave number k0 and angular frequency ω0 , where E(ω) varies slowly with ω. The phase velocity ω/k varies with frequency or equivalently with the wave number. Putting ∆ω = ω − ω0 and ∆k = k − k0 = ∆ω(dk/dω), this can be rewritten E(z, t) = exp [i(ω0 t − k0 z)] E(ω) exp [i∆ω(t − z dk/dω)] dω, (11.40) where the ﬁrst term describes the rapidly oscillating waves in ﬁgure 11.10 and the second term the envelope. Each wave in the envelope is a function of t − (dk/dω)z, so the velocity of the envelope is vg = dω/dk,
(11.41)
which is called the group velocity. Surprisingly a wavepacket does not necessarily travel at a velocity close to the mean velocity of its inﬁnite
11.6
Absorption by, and reﬂection oﬀ metals
317
plane wave components. In ﬁgure 11.10 the waves within the wavepacket travel at the phase velocity while the wavepacket itself, that is the envelope of the waves, travels at the group velocity. If one could travel parallel to and alongside the wavepacket at its velocity and somehow detect the internal waves, then they would appear to travel through the wavepacket. If the wave velocity exceeds the group velocity the waves would appear to ﬂow forward under the envelope. Energy and information travel at the group velocity so, provided the group velocity remains less than c, it is of little consequence if the phase velocity exceeds c. 5
Other useful expressions for the group velocity can be derived from eqn. 11.41 (11.42)
vg = (c/n)[1 + (λ/n)(dn/dλ)]; 1/vg = 1/vp + (ω/c)(dn/dω);
(11.43) (11.44)
1/vg = 1/vp − λ d(1/vp )/dλ,
(11.45)
where λ is, as usual, the free space wavelength. Evidently in free space where there is no dispersion vg = vp = c. An example that brings out the importance of the group velocity is that of electromagnetic waves travelling in the ionosphere. Rewriting eqn. 11.38 gives ω 2 = n2 ω 2 + ωp2 = k 2 c2 + ωp2 ,
(11.46)
where k is the wave number in the plasma. This behaviour is drawn as a solid line in ﬁgure 11.11 for waves with frequencies above the plasma frequency, and asymptotically approaches the form for electromagnetic waves in free space ω = kc. The corresponding phase and group velocities are displayed versus the angular frequency in ﬁgure 11.12. Close to the plasma frequency the phase velocity is very large, while the group velocity is close to zero. At high angular frequencies both velocities converge on c. Figure 11.13 shows an ionogram recording of the apparent heights at which short pulses of radio waves are reﬂected from the ionosphere. The ordinate is directly proportional to the delay between sending and receiving the signal back after reﬂection oﬀ the ionosphere. The abscissa is the frequency of the electromagnetic waves used. The peaks in the ionogram indicate reﬂection from individual ionized layers at increasing heights with correspondingly increasing electron densities. As the frequency of the transmitter is increased toward the plasma frequency of a particular layer the delay climbs steeply. Evidently the velocity of the signal penetrating the plasma layer slows down, reaching a minimum at the plasma frequency for that layer. This matches the behaviour expected if the velocity of the signal is indeed the group velocity. Some further care is required in deﬁning the velocity at which information can travel. Consider the electromagnetic radiation with wavelength in the region of anomalous dispersion shown in ﬁgure 11.4. One version
Velocity/c
vg = vp − λ(dvp /dλ);
4 3 2
vp
1 0
vg 0
0.5
1 1.5 (ω ωp)/ωp
2
Fig. 11.12 Angular frequency variation with the group and phase velocities above the plasma frequency ωp .
318 Scattering, absorption and dispersion
700 Ordinary 600
Extraordinary
Range in km
500 400 300 200 100 0 0
1
2
3 4 Frequency in MHz
5
6
7
Fig. 11.13 Ionogram showing the apparent height at which pulses of radiation are reﬂected from the ionosphere versus the frequency. Courtesy Dr.M. Rietveld of the EISCAT Scientiﬁc Association, N9027 Ramfjordmoen, Norway.
of the expression for the group velocity is vg = (c/n)[1 + (λ/n)(dn/dλ)].
K.E Oughstum and N.A. Cartwright in the Journal of Optics, A4 (2002) S125 conclude that ‘Superluminal energy and information transfer is not physically possible within the framework of the Maxwell–Lorentz theory in linear, causally dispersive systems.’
Therefore, when dn/dλ is large and positive, which it certainly will be across a narrow spectral line, vg can also become larger than c. A narrow line has a correspondingly sharp absorption peak so that radiation entering the medium is strongly absorbed. Consequently the frequency components in the wavepacket closest to the centre of the spectral line are preferentially absorbed, while the components of the wavepacket with frequencies above and below the spectral line are less attenuated. In such cases the energy and information all travel at a velocity less than c. This velocity at which information travels is termed the signal velocity . A diﬀerent aspect of the distinction between group and phase velocity is met with when electromagnetic waves enter birefringent materials. Figure 10.8 shows how the Poynting vector of an electromagnetic wave with extraordinary polarization generally points in a diﬀerent direction from the wave vector normal to the wavefronts. Energy travels at the ray velocity in the direction of the Poynting vector, while the wavefronts travel at the wave velocity. Now a wavepacket in a beam consists of inﬁnite plane waves whose directions are distributed over a range of angles as well as wavelengths. These intefere constructively and travel with the group velocity in a direction determined by their Poynting vectors.
11.6
Absorption by, and reﬂection oﬀ metals
319
A surprising conclusion can be made from the analysis in this and the preceding chapter. Not only will the group velocity of light diﬀer in magnitude from the phase velocities of the inﬁnite sinusoidal plane waves making up the wavepacket in a dispersive medium, but also in a birefringent material a wavepacket of extraordinary waves will generally not even travel in the direction normal to the wavefronts of the constituent inﬁnite plane waves.
11.6.3
Surface plasma waves
In addition to plasma waves that ﬁll the volume of a metal, and which are for this reason called bulk plasma waves, surface plasma waves (SPW) also occur. These waves travel along the interface between a metal and a dielectric. One device ﬁrst used by Kretschmann and Raether in 19688 for exciting SPWs is shown in the upper panel of ﬁgure 11.14. The prism is made from a dielectric with relatively large refractive index, √ np = εp , for example quartz. A thin layer of metal is deposited on the prism base and the lower surface of this metal ﬁlm is put in contact with a dielectric whose refractive index is much less than that of the prism. Because the metal layer is so thin, there can be total internal reﬂection at what is eﬀectively a prism/dielectric interface. The evanescent wave can excite a surface plasma wave travelling along the metal/dielectric interface provided the phase velocity of the incident light matches that of a surface plasma wave. If this match is achieved the incident light feeds the surface plasma wave, with the result that the reﬂectance drops well below the 100% which normally signals TIR. The intensity distribution of the surface plasma wave as a function of distance from the metal/dielectric interface is illustrated in the lower panel in ﬁgure 11.14. It typically extends of order one wavelength into the dielectric, but less far into the metal. The upper panel of the ﬁgure also illustrates how to detect the presence of a surface plasma wave when using a white light source. After reﬂection the light emerges from the prism and falls on a diﬀraction grating. The resulting spectrum will show a gap at any wavelength for which TIR is suppressed when coupling to a surface plasma wave has occured. Because the surface plasma wave is sensitive to the physical and chemical content of a layer around one wavelength thick at the dielectric surface the device described is currently used as a biosensor. This application will be described after analysing the production of surface plasma waves. If the material on which the metal is placed is nonmagnetic then the surface plasma waves are ppolarized, and this case will be analysed now. The coordinate system is indicated in the upper panel of ﬁgure 11.14 with the plane x = 0 located on the metal/dielectric boundary. Then the magnetic ﬁeld of the surface plasma wave points in the ydirection and has values BD = BD0 exp (−κD x) exp [i(ωt − kz z))],
(11.47)
8
See Surface Plasmons by H. Raether, published by SpringerVerlag, Berlin (1988).
Grating θp z x
Metal Dielectric
Metal SPW Intensity Dielectric
Fig. 11.14 The Kretschmann–Raether scheme for exciting surface plasma waves is shown in the upper panel. The lower panel shows the exponential decline in the intensity of the surface plasma wave with distance from the contact surface. The vertical scale is magniﬁed in the lower panel.
320 Scattering, absorption and dispersion
BM = BM0 exp (κM x) exp [i(ωt − kz z))],
(11.48)
where the subscripts D and M refer to the dielectric and metal respectively. The wave vector kD is given by 2 = kz2 − κ2D . kD
Also 2 2 kD = ω 2 /vD = εD ω 2 /c2
where vD is the wave velocity in the dielectric and εD its relative permittivity. Eliminating kD from the last two equations gives κ2D = −εD ω 2 /c2 + kz2 ,
(11.49)
with a similar relation for the wave vector in the metal κ2M = −εM ω 2 /c2 + kz2 .
(11.50)
Next we apply Maxwell’s equation, eqn. 9.16, to the surface plasma wave in the dielectric and keep only the zcomponent −κD BD = iεD ωEDz /c2 , where EDz is the component of electric ﬁeld in the zdirection. Rearranging this last equation gives EDz = +i(κD c2 /ωεD )BD .
(11.51)
Similary for the wave in the metal we get EMz = −i(κM c2 /ωεM )BM .
(11.52)
Finally we can impose the twin requirements that the zcomponent of the electric ﬁeld is continuous at the surface, and that the magnetic ﬁeld is continuous at the surface: BD = BM and EDz = EMz . Then dividing eqn. 11.51 by eqn. 11.52 gives κD /εD = −κM /εM .
(11.53)
This condition can only be satisﬁed if εM is negative; which is the situation in the case of a metal when the frequency of the incident electromagnetic radiation is less than the plasma frequency. When this is the case there can be a ppolarized surface plasma wave. Note also that its electric ﬁeld is not transverse to the direction of propagation but lies in the zOx plane. Proceeding further by squaring the last equation gives κ2D /ε2D = κ2M /ε2M . Using eqns. 11.49 and 11.50 to substitute for κD and κM yields ε2M /ε2D = [kz2 − εM (ω/c)2 ]/[kz2 − εD (ω/c)2 ],
11.6
Absorption by, and reﬂection oﬀ metals
321
from which it follows that kz2 = (ω/c)2 εM εD /(εM + εD ),
(11.54)
κ2D/M = −(ω/c)2 ε2D/M /(εM + εD ).
(11.55)
while
At high frequencies, but still below the plasma frequency, εM = 1 − ωp2 /ω 2 , and substituting this value for εM in eqn. 11.54 gives (11.56)
Equation 11.56 is the surface plasma wave’s dispersion relation and is drawn as a curved tends to line in ﬁgure 11.15. As kz tends to inﬁnity ω √ the limit ωp / (1 + εD ), while near the origin it tends to ckz / εD . In the same ﬁgure the steeper, straight broken line follows the dispersion relation for light in free space incident at an angle θp , ω = kz c/ sin θp . Evidently light in free space cannot excite a surface plasma wave because its phase velocity always exceeds that of the surface plasma wave; put another way their dispersion curves never cross away from the origin. However the phase velocity along the zdirection of light incident √ in the prism is c/( εp sin θp ), and this ﬂatter dispersion relation can intersect the dispersion curve for the surface plasma wave. Intersection is guaranteed if the slope is less than that of the dispersion curve for √ √ the SPW at the origin: εp sin θp > εD . The conditions for exciting a surface plasma wave can be produced either by varying the angle of incidence with monochromatic light, or by scanning in wavelength at a ﬁxed angle of incidence, as is illustrated in ﬁgure 11.14. A sharp drop in the reﬂectance is the required signal. Using eqn. 11.55, and taking ε D to be 2.0 and εM to be −10, the intensity of the surface plasma wave is predicted to fall oﬀ by a factor e in a distance of 0.11 (free space) wavelengths in the dielectric and 0.024 wavelengths in the metal. Gold is the metal prefered for the prism coating because it does not tarnish and also because the dips in reﬂectance are very narrow and deep. A gold layer which is around 50 nm thick gives an optimally deep and narrow dip. When used in a biosensor the metal surface is coated with an appropriate agent to which the target (for example a bacterium such as E. coli) will bind. The ﬂuid under test is then made to ﬂow past the prepared surface. If this ﬂuid contains any targets these are trapped by the surface agent and the accumulation of such material changes the relative permittivity εD of the surface layer, with a corresponding change in the frequency of the surface plasma wave. The resulting change in the wavelength at which the dip in reﬂectance occurs is large enough that a few picograms of target biomaterial per square millimetre is detectable. This approach has the ﬂexibility of separating functions: the detector is sensitive yet biologically nonspeciﬁc, while the biological agent is designed to trap a speciﬁc biotarget.
Angular frequency
kz2 = (ω/c)2 εD (1 − ωp2 /ω 2 )/(εD + 1 − ωp2 /ω 2 ).
Free space
Prism
SPW
zcomponent of k
Fig. 11.15 Dispersion relations for a surface plasma wave, for an electromagnetic wave in free space and for an electromagnetic wave in the prism. The condition for which the wave in the prism and surface plasma wave match in frequency and wave number is marked by a spot.
322 Scattering, absorption and dispersion
From eqn. 11.54 it is apparent that when the metal and dielectric electric permittivities are close to cancelling the wavelength of the plasma wave along the surface direction becomes extremely short. This provides a practical means of coupling optical waves into nanostructures, thus linking optoelectronics and nanotechnology. At present this development is in its infancy.
11.7
Further reading
Optical Physics, third edition, by S. G. Lipson, H. S. Lipson and D. S. Tannhauser, published by Cambridge University Press (1995). A very imaginative book which concentrates ﬁrmly on understanding. It should prove useful in several areas besides the present chapter. ‘How light interacts with matter’ by V. F. Weisskopf in Scientiﬁc American, August 1968, pages 60–71 is very also informative.
Exercises (11.1) Show that for a plasma vg = c2 /vp . (11.2) Calculate the plasma frequency in a layer of the ionosphere where the electron number density is 1012 m−3 . What will the wave and group velocities be at a frequency 10 MHz? (11.3) The power in a beam of light falls by a factor e in passing through 3 m of glass. What is this loss in dB km−1 ? (11.4) Red ink is allowed to dry to form a thin solid layer on a glass sheet. It now appears in reﬂected light to be greenish. Why is this? (11.5) Prove eqns. 11.42.
dN = dΩ(ω 4 p20 sin2 θy )/(32π 2 c3 ε0 ).
(11.6) Calculate the reduction of intensity of light of wavelength 589 nm after travelling through 200 km of air at NTP. The refractive index of air is 1.000292 and the number density of molecules is 3 1025 m−3 . (11.7) One empirical formula introduced to ﬁt narrow resonances seen in the refractive index is ε=1+
(11.9) Monochromatic unpolarized light travelling in free space in the xdirection is scattered from a sphere of refractive index n, and radius a, this radius being much less than the wavelength of the light λ in free space. (a) Deduce using eqn. 11.27 that the electric dipole moment of the sphere is p0 = 4πa3 Gε0 E, where E = E0 cos (ωt) is the electric ﬁeld of the light and G = (n2 − 1)/(n2 + 2). (b) Hence show that the time averaged ﬂux scattered in solid angle dΩ in a direction making an angle θy with the dipole axis when dipole points along the yaxis is
2
2
[Ai λ /(λ −
λ2i )].
Express the two constants for a given spectral line (Ai and λi ) in terms of the electron natural frequency and the plasma frequency. (11.8) Calculate the reﬂectance on copper of light of wavelength 589 nm using the data given in Table 11.1.
(c) Write the corresponding expression if the dipole points along the zaxis. (d) Show that sin2 (θy ) + sin2 (θz ) = 1 + cos2 θx , where (θx , φx ) are the polar angles with respect to the xaxis etc. The solid angle between the cones of semiangle θx and θx + dθx around the xaxis is 2π sin (θx )dθx . (e) Hence show that the ﬂux of unpolarized radiation scattered into an angular interval between θx and θx + dθx is dN = sin θx dθx [ω 4 p20 (1 + cos2 θx )]/[32πc3 εo ].
11.7
Further reading 323
(f) If the diﬀerential crosssection for scattering is deﬁned as (dropping the subscript x)
(g) Hence show that the total Rayleigh crosssection from a single sphere is
dσ/dθ = (dN/dθ)/F,
σ = (8π/3)k 4 a6 G2 ,
where F is the incident ﬂux per unit area, then where k = 2π/λ. show that the diﬀerential crosssection for Rayleigh (11.10) Why is it that no plasma surface wave is excited at scattering of a sphere is the metal/prism interface in ﬁgure 11.14? dσ/dθ = πk 4 a6 G2 (1 + cos2 θ) sin θ.
(11.11) Prove eqn. 11.44.
This page intentionally left blank
The quantum nature of light and matter 12.1
Introduction
Just a century ago experiments with electromagnetic radiation led to a totally new insight, that not only matter comes in discrete packets but so too does electromagnetic radiation. These packets of radiation, ﬁrst proposed by Planck and Einstein, are called photons, and electromagnetic radiation of frequency f Hz consists of photons each carrying a quantum of energy hf , where h is Planck’s constant with a value 6.626 10 −34 J s. Just how small this is can be appreciated by noting that a one watt torch emits ∼1018 photons per second. Three pieces of experimental evidence were crucial to the acceptance of the concept of the quantization of radiation. The ﬁrst came from the measurement of the spectrum of the radiation from a perfect absorber and emitter of radiation – a socalled black body. The second piece of evidence was provided by Lenard’s and Millikan’s studies of the photoelectric eﬀect in which em radiation liberates electrons from a metal surface. A third piece of evidence was obtained when Compton measured the change in wavelength of Xrays scattered by electrons. Somewhat later in 1924 de Broglie proposed that if electromagnetic waves had a particle nature, then for consistency matter should possess wave properties. Shortly thereafter interference eﬀects were observed with electrons, and more recently they have been observed with other particle types including atoms and molecules. The experiments which demonstrated the particle nature of electromagnetic radiation and the wave nature of electrons are described in the ﬁrst section of the chapter. In 1911 Rutherford discovered the basic features of atomic structure. It became clear that an atom contains a nucleus that carries most of the mass and all the positive charge, but whose diameter is only one hundredthousandth of the atom’s diameter. Around this nucleus the much lighter, pointlike, negatively charged electrons circulate at distances of order 10−10 m under the electrical attraction of the nucleus. Their total charge is equal and opposite to that of the nucleus. It was originally suggested that the electrons travelled in orbits like planets around the Sun, but this picture of the atom had a serious ﬂaw. The electrons in orbits would be accelerating radially and hence they would radiate continuously. As a result they would lose energy and spiral
12
326 The quantum nature of light and matter
rapidly into the nucleus. Bohr got around this diﬃculty by arbitrarilyquantizing the orbits. He postulated that electrons only occupy certain orbits for which the angular momentum is an integral multiple of h/2π and that while in these orbits the electrons do not radiate. In addition he postulated that an electron can move from one allowed orbit to another in a transition which is instantaneous and which is accompanied by the emission or absorption of a photon carrying the energy diﬀerence between the two orbits. Bohr’s model was able to account for the existence of lines in atomic spectra and to explain quantitatively the striking regularities in the frequencies of the spectral lines of hydrogen and hydrogenlike atoms. These features had previously been incomprehensible, so that this new understanding showed that here were the beginnings of a theory based on quantization. The successes, and crucial weaknesses, of the Bohr model will be described in the second part of the chapter. Today the Bohr model still provides a useful conceptual step towards the comprehensive theory of quantum phenomena, known as quantum mechanics. If electromagnetic radiation, and matter, show both wave and particle properties then the obvious question to ask is how these properties can be integrated into a coherent conceptual and mathematical framework. In the third part of this chapter the interpretation of the wave–particle duality is discussed. Put very brieﬂy, the wave intensity over a region of space takes on a new meaning as the probability of ﬁnding the associated particle in that region. However with probability comes uncertainty, and this is illustrated with examples. Each degree of freedom has an associated pair of conjugate variables, as for example in the case of a single particle the position and momentum in each dimension. These pairs can only be measured with limited precision, even with ideal measuring instruments. This limitation is quantiﬁed by Heisenberg’s uncertainty principle for pairs of conjugate variables. Complemenarity can also apply to pieces of information other than kinematic variables. In all cases, kinematic or not, exact knowledge of one piece of information implies total uncertainty about the other. As an example, in Young’s double slit experiment one may either know which slit the photon passes through or one may observe the two slit interference pattern, but not both. If the paths are indistinguishable then interference is seen, but if the path is known (welcher Weg information) interference no longer occurs. The ﬁnal section of the chapter is used to bring out the connections between photons, wavepackets, coherence volumes and modes of the electromagnetic ﬁeld.
12.2
The black body spectrum
Black body radiation is electromagnetic radiation contained within an enclosure whose walls are maintained at a constant uniform temperature
12.2
and, as discussed in Section 1.8, its spectrum is independent of the material of the enclosure. The practical form of a black body was described in Section 1.8. Figure 12.1 shows the spectra of black body radiation at three temperatures, like those that had been measured before 1900 by Lummer and Pringsheim, and by others. Their method was to disperse the black body spectrum with a diﬀraction grating and then to measure the heating eﬀect of each wavelength segment of the spectrum. 0.6
5000K 0.5 classical prediction for 3000K
3
Energy density in J m µm
1
0.4
0.3
0.2
4000K
0.1 3000K
0 0
0.2
0.4
0.6
0.8 1 1.2 1.4 Wavelength in µm
1.6
1.8
2
Fig. 12.1 Black body radiation spectra at 3000 K, 4000 K and 5000 K. Planck’s quantum predictions which are indicated with full lines ﬁt the data. The classical prediction for 3000 K is shown with a broken line.
It is straightfoward to make a prediction of the spectrum using classical theory, but the result is spectacularly wrong! The density of modes of electromagnetic radiation has been calculated in Section 9.8.1 to be ρ(f )df = 8πf 2 df /c3 ,
(12.1)
where f is the frequency. In arriving at this value electromagnetic radiation is regarded as having two degrees of freedom, corresponding to
The black body spectrum 327
328 The quantum nature of light and matter
the two polarizations. Classical thermodynamics then predicts that the energy should be partitioned equally between the modes with each mode having a mean energy in thermal equilibrium of kB T , where T K is the temperature and kB is Boltzmann’s constant, 1.381 10−28 J K−1 . Thus the energy spectrum predicted is W (f )df = 8πkB T f 2 df /c3 ,
(12.2)
with W (f ) measured in J m−3 Hz−1 . The equivalent distribution in wavelength is given by Wλ (λ)dλ = W (f )df, hence Wλ (λ) = W (f )df /dλ = 8πkB T /λ4 ,
(12.3)
and is indicated for a temperature of 3000 K by the broken line in ﬁgure 12.1. At high frequencies (short wavelengths) this prediction is wildly in error, a feature known as the ultraviolet catastrophe. Each of the rapidly increasing number of modes at high frequency has the same energy, which implies an inﬁnite energy in the electromagnetic ﬁeld! In 1901 Planck discovered what he took to be a temporary mathematical ﬁx which brought the prediction into agreement with the data, and he imagined that his procedure would somehow be incorporated into classical theory. In fact this was the ﬁrst step in revealing the quantum basis of nature. What Planck proposed was that radiation is absorbed and emitted by the walls of the container in packets or quanta of energy, E = hf = h ¯ ω,
(12.4)
where ω is the angular frequency of the radiation and h is known as Planck’s constant; h ¯ is simply h/2π. If there are n quanta in the mode its energy is nhf . Then the probability of there being n quanta in any mode within the enclosure is given by the Boltzmann distribution for systems in thermal equilibrium P (n) ∝ exp (−Energy/kB T ) = exp (−nhf /kB T ). Putting x = exp (−hf /kB T ) and normalizing so that the probabilities add up to unity gives P (n) = xn / xn = xn (1 − x). n
Hence the mean number of photons in a mode is n=
nP (n) = (1 − x)
n
n
nxn = (x − x2 )
d n x dx n
d = (x − x ) [1/(1 − x)] = x/(1 − x) = 1/(x−1 − 1). dx 2
12.2
The black body spectrum 329
Therefore n = 1/ [ exp (hf /kB T ) − 1 ].
(12.5)
Using the expression for the density of modes given above in eqn. 12.1, the energy spectrum of black body radiation was predicted by Planck to be W (f )df = nhf ρ(f )df = (8πhf 3 /c3 )df /(exp (hf /kB T ) − 1),
(12.6)
again in J m−3 Hz−1 . In a material of refractive index µ there would be an additional factor µ3 in the numerator W (f )df = (8πhf 3 µ3 /c3 )df /(exp (hf /kB T ) − 1).
(12.7)
Planck found that this expression would ﬁt the the observed black body spectra at all temperatures for which it was measured with the same value of h in each case. Quantization had solved the inherent weakness of classical wave theory and provided an excellent ﬁt to the data. Thus quantization had to be taken seriously! The current measured value of Planck’s constant is 6.626 10−34 J s making h ¯ equal to 1.0546 10−34 J s. Note that in the limit of low frequencies and high temperatures, that is when hf /kB T is very small, Planck’s formula reduces to the classical expression. Several simple properties of the black body spectrum follow on from this analysis. The wavelength at which the spectrum peaks, λpeak , is obtained by diﬀerentiating eqn. 12.6 and setting the result to zero. This gives λpeak = 2.898 10−3/T, (12.8) which was discovered experimentally in 1893 by Wien, and is known as Wien’s law. The energy density in electromagnetic radiation in equilibrium at a temperature T can be obtained by integrating eqn. 12.6 over frequency. This gives ∞ 4 4 W= W (f )df = 8π 5 kB T /(15c3 h3 ), (12.9) 0
where we use the result that the deﬁnite integral equals π 4 /15. The units of W are J m−3
∞ 0
x3 dx/(exp x − 1)
The ﬂux of energy per unit time across unit area in the enclosure in one sense is of interest. This quantity is the irradiance or intensity of black body radiation measured in W m−2 . Figure 12.2 shows a hemisphere drawn on a selected surface. Those modes whose normals lie between the two cones of semiangles θ and θ + dθ make up a fraction sin θdθ/2 of all the modes. Thus the total energy in these modes crossing unit area per unit time is F = (sin θdθ/2)(Wc cos θ).
θ dθ
Fig. 12.2 Notional surface drawn within an enclosure containing black body radiation.
330 The quantum nature of light and matter
Integrating this result over all angles gives the irradiance or intensity of black body radiation π/2 F = (Wc/2) sin θ cos θ dθ = Wc/4. (12.10) 0
Thus the intensity is 4 4 T /(15c2 h3 ) = σT 4 . F = 2π 5 kB
(12.11)
This variation as the fourth power of the absolute temperature is known as Stefan’s law, and the Stefan–Boltzmann constant, σ, is 5.76 10 −8 W m−2 K−4 . The spectrum of the cosmic microwave background radiation which ﬁlls the universe has been measured with microwave dishes on board the COBE and WMAP satellites. The measurements reveal that this cosmic microwave background has a spectrum that deviates by only parts in one hundred thousand from the black body radiation spectrum at a temperature of 2.75 K. In fact it is the most perfect black body radiation spectrum ever observed.
12.3
The photoelectric eﬀect
The photoelectric eﬀect occurs when visible light or ultraviolet light falls on a metal or an alkali metal causing electrons to be emitted from the surface irradiated. By 1902 Lenard had shown that for each such metal there exists a threshold frequency below which no photoelectrons are produced, and this threshold frequency does not change however high the intensity of the incoming radiation. This behaviour is impossible to explain with em wave theory alone. According to wave theory an electron in a metal would be forced to oscillate at the incoming wave frequency, this oscillation would build up in amplitude over time and eventually the electron would break free from the surface. Crucially, this sequence predicted by classical wave theory should proceed whatever the frequency of the incident electromagnetic radiation. Lenard also observed that the maximum energy of the emerging electrons depends solely on the frequency of the radiation and not on its intensity: facts which are again inconsistent with classical wave theory. Above threshold the number of electrons emitted was, as expected, proportional to the intensity of the radiation. In 1905 Einstein proposed that in the photoelectric eﬀect a single electron absorbs one of Planck’s energy quanta from the incident radiation and escapes from the surface. Supposing it requires an energy φ to release an electron lying exactly at the surface of the metal, then all electrons will emerge with kinetic energy less than or equal to KEmax = hf − φ.
(12.12)
12.3
The photoelectric eﬀect 331
Table 12.1 Cutoﬀ wavelengths, frequencies and work functions for several metals from CRC Handbook of Chemistry and Physics, 77th edition, published by the Chemical Rubber Company Press, Boca Raton, Fl., (1997). Courtesy Taylor and Francis Group. Metal
Wavelength in nm
Frequency in THz
Work function in eV
Sodium Lithium Cesium Copper Nickel
451 428 580 267 241
665 701 517 1120 1240
2.75 2.90 2.14 4.65 5.15
φ depends on the metal and is called the work function of the metal. Electrons originating deeper in the metal will lose energy through collisions on their way out, so that KEmax is the maximum electron energy. Einstein’s proposal introduces the required cutoﬀ frequency fco = φ/h; radiation below this frequency cannot cause any photoemission of electrons. Millikan, who was highly sceptical, set out to test Einstein’s predictions in a series of experiments that extended Lenard’s studies, and on which he, Millikan, continued working for over a decade. An outline of his later apparatus is shown in ﬁgure 12.3. A monochromator is used to select light from a strong line in the spectrum of a mercury lamp, this light then travels through a small window to fall ﬁnally on the metal surface being studied. The interior of the outer container is evacuated so that electrons do not lose energy in collisions with gas molecules, and so that the metal surface is not contaminated either. Because surface impurities change the threshold frequency so much, each metal surface was scraped immediately before being irradiated. Electrons ejected from the metal surface travelled to a copper Faraday cup and the resulting current was measured. A net negative voltage was applied to the cup relative to the metal and this voltage was increased until the current vanished. If this cutoﬀ voltage is V and e is the electron’s charge, then the maximum electron kinetic energy would be eV . One advantage accruing from the use of a copper cup is that copper has a high threshold frequency with the result that there was negligible photoemission caused by light reﬂected onto the cup. Millikan conﬁrmed that there exists a threshold frequency for each metal he tested, and that with radiation of any lower frequency no electrons are emitted, however intense the radiation. Some modern determinations of the cutoﬀs for polycrystalline pure metal surfaces are given in Table 12.1. Secondly Millikan showed that when the kinetic energy of the highest energy photoelectrons was plotted against the frequency of the radiation the dependence ﬁtted Einstein’s predicted linear relation KEmax = hf − φ = h(f − f0 ).
(12.13)
+ e path Mch. Alkali metal
photon path from Hg source
Faraday cup
Fig. 12.3 Millikan’s apparatus to study the photoelectric eﬀect. Mch is a monochromator. The light passes through a window in the container.
332 The quantum nature of light and matter
3
Stopping voltage
2.5 2 Sodium
1.5
Lithium
1 0.5 0 0.4
0.5
0.6
0.7
0.8
0.9
1
1.1
1.2
15
Frequency / 10 Hz Fig. 12.4 Millikan’s measurements of the stopping voltage against the radiation frequency for sodium and lithium. The voltages are corrected for the contact potentials that are present at contacts between diﬀerent metals in the circuit. Adapted from R. A. Millikan, Physical Review 7, 355 (1916), by courtesy of the Amercan Physical Society.
This dependence is evident in the data collected by Millikan and shown in ﬁgure 12.4. From the slope of the lines ﬁtted to the data for each metal Millikan extracted values for h which agreed with each other and with the value found by Planck. Millikan also conﬁrmed that the number of photoelectrons is proportional to the intensity of the radiation. Later in 1927 Lawrence and Beams measured the delay between the moment that the radiation ﬁrst arrives at the surface after the source is switched on and the moment at which electron emission commences. Classically the wave energy is spread over the whole surface of the wavefront and it would take some time before any individual electron could accumulate enough energy to escape from the surface. The delay should therefore increase as the intensity diminishes. By contrast quanta are localized so that emission should commence instantaneously when the light is turned on. Lawrence and Beams found that the delay was less than the resolution of their timing methods and gave an upper limit of 3 10−9 s for the delay at low light intensities. A tighter limit for the delay was obtained by Forrester, Gudmundson and Johnson in 1955. They studied the photoelectric eﬀect using light modulated at high frequencies and observed that the detector signal followed the modulation faithfully, showing that any delay was much less than the modulation period. Their conclusion was that the delay is signiﬁcantly less than 10−10 s. The absence of any
12.4
The Compton eﬀect 333
delay makes implausible a semiclassical picture, in which electromagnetic radiation is absorbed and emitted as quanta but travels purely as waves.
12.4
Photon out (f)
The Compton eﬀect
θ
Compton, following through the consequences of the quantization of radiation, appreciated that when Xrays scatter from matter the underlying process is the scattering of an individual quantum of electromagnetic radiation from a single electron which is initially at rest. This process is pictured in ﬁgure 12.5. The particle aspect of the quantum of energy is now explicit with electromagnetic radiation being pictured as consisting of particles called photons each carrying Planck’s quantum of energy. If the incoming photons had frequency f0 Compton assigned to each an energy hf0 . He also assigned a momentum hf0 /c to the photons, which takes the relationship expressed in eqn. 9.54 between wave energies and momenta and applies it to an individual photon travelling in free space. The special theory of relativity gives the same relationship, E = pc. Suppose now that the scattered photon has frequency f , that the electron recoils with velocity βc, and that the scattering angles are as shown in ﬁgure 12.5. The electron must be treated relativistically because the kinetic energy of the recoiling electron is comparable to its rest mass energy. Initially its energy before being struck is the rest mass energy mc2 , while its energy andmomentum afterwards are mγc2 and mβγc respectively, where γ = 1/ (1 − β 2 ). Energy is conserved in the scatter so that hf0 + mc2 = hf + mc2 γ. (12.14) The components of momentum along and at right angles to the incident photon’s direction are also conserved so that hf0 /c − hf cos θ/c = mγβc cos φ, hf sin θ/c = mγβc sin φ,
(12.15) (12.16)
where θ and φ are the respective scattering angles of the photon and the electron relative to the incident photon’s direction. Rearranging and squaring the energy equation, eqn. 12.14, gives h2 (f02 + f 2 − 2f f0 ) = m2 c4 (γ − 1)2 . Squaring the momentum equations eqns. 12.15 and 12.16 and adding the results gives h2 (f02 + f 2 − 2f f0 cos θ) = m2 γ 2 β 2 c4 . Taking the diﬀerence between the last two equations yields 2h2 f f0 (1 − cos θ) = 2m2 c4 (γ − 1).
Photon in (f 0)
φ
Electron recoil
Fig. 12.5 Compton’s particle interpretation of Xray scattering from an electron.
334 The quantum nature of light and matter
The energy equation, eqn. 12.14, can be used to replace the right hand side of the last equation and this gives 2h2 f f0 (1 − cos θ) = 2h(f0 − f )mc2 , which when divided by 2hf f0 mc2 leaves (h/mc2 )(1 − cos θ) = (1/f − 1/f0 ). Reexpressing this equation in terms of the wavelengths produces a simpler form λ − λ0 = (h/mc)(1 − cos θ). (12.17)
Collimator Target
Xray tube
Crystal
Collimator
Ionization chamber
Fig. 12.6 Compton’s apparatus for studying the wavelength change in Xray scattering from various light elements.
This change in wavelength of the scattered Xrays is called the Compton eﬀect. It depends only on physical constants and the scattering angle, and not at all on the nature of the scattering material. Numerically the term that quantiﬁes the wavelength shift, h/mc, is 0.0024 nm which shows that it is necessary to use short wavelength radiation. If the experiment were attempted with visible light the change in wavelength would, even now, be hard to detect let alone measure precisely. Compton’s Xray photons had energies very much larger than the kinetic energy and binding energy of the electrons in the atoms which he used, and hence the assumptions that the target electrons are free and at rest is fully justiﬁed. Compton’s apparatus is sketched in ﬁgure 12.6. Radiation from the molybdenum Xray source is scattered oﬀ a graphite block. Collimators are used to select narrow beams of Xrays, a collimator being made of a pair of lead plates each pierced by a single small hole. The scattering angle oﬀ graphite is selected by positioning a ﬁrst collimator appropriately. After emerging from this collimator the Xrays are Bragg scattered oﬀ a calcite crystal, with a second collimator set so that the angle of reﬂection is equal to the angle of incidence. Finally the Bragg scattered Xrays enter an ionization chamber ﬁlled with a gas whose atoms are readily ionized by Xrays. Under an applied voltage a current ﬂows proportional to the Xray ﬂux. Figure 12.6 is drawn for the case that the scattering angle between the incident and scattered Xrays at the carbon target is 90◦ . At each such setting of the scattering angle from graphite Compton varied the Bragg scattering angle at the calcite crystal in steps and and at each step recorded the current. The resulting angular distribution of the scattered Xray intensity was then converted to a distribution in wavelength using Bragg’s law. Figure 12.7 shows the spectra observed at three diﬀerent scattering angles oﬀ a carbon target. The wavelength of the incident radiation is indicated by the broken line in each panel and the right hand peak contains the photons scattered from electrons. Compton found that the shift in wavelength of this right hand peak ﬁtted his prediction precisely in magnitude, in its dependence on the scattering angle, and in its independence of the target material. In 1949 Hofstadter and McIntyre detected both the photon and the recoil electron, checking that they emerge in coincidence, to within 10 −9 s.
12.5
de Broglie’s hypothesis
In 1924 de Broglie pointed out that if electromagnetic waves possess particle properties, then it might be reasonable to suppose that material particles such as electrons should possess wave properties. He proposed that the relations connecting the wave and particle properties of em radiation should also apply to material particles. Thus the frequency of the wave f associated with a particle of total energy E would be given by Planck’s relation E = hf. (12.18) The parallel relation for the momentum of a photon is
80 0
φ = 45
60 40 20 0
70
72 74 76 78 Wavelength in pm.
80
120
p = E/c = h/λ.
100
de Broglie therefore proposed a similar relationship for material particles p = h/λ.
100
(12.19)
Intensity
12.5
120
Intensity
The ﬁt achieved to the black body spectrum, the explanation of the features of the photoelectric eﬀect and the prediction and measurement of the Compton eﬀect were the key elements in establishing that electromagnetic radiation has a particle nature.
de Broglie’s hypothesis 335
This is called the de Broglie relation, and λ is known as the de Broglie wavelength of material particles.
80 φ = 90
60
0
40 20
It is important to note that the energy concerned is the total energy of the material particle given by an expression from the special theory of relativity E 2 = p 2 c2 + m 2 c4 , (12.20)
0
where m is the rest mass. This collapses to E = pc for the massless photon and is the expression applied by Compton. At the other extreme, when velocities are small compared to c, p mc and
120
E = mc2 (1 + p2 /m2 c2 )1/2 ≈ mc2 + p2 /2m,
80
where the binomial expansion is used. p /2m is the usual kinetic energy of Newtonian mechanics, while mc2 is the rest mass energy. de Broglie’s ideas were conﬁrmed when in 1926 Davisson and Germer, and simultaneously G. P. Thomson, demonstrated the wave properties of electrons. Davisson and Germer accelerated electrons through potentials of tens of volts and then scattered them from a nickel crystal. They observed diﬀraction from the regular atomic layers in the crystal. Thomson, on the other hand, passed electrons through thin ﬁlms of randomly oriented microcrystals, and observed sharp circular diﬀraction rings at angles satisfying the usual Bragg condition nλ = 2d sin θ,
(12.21)
72 74 76 78 Wavelength in pm.
80
100 Intensity
2
70
0
φ = 135
60 40 20 0
70
72 74 76 78 Wavelength in pm.
80
Fig. 12.7 Wavelength distribution of Xrays scattered from graphite. The vertical line indicates the wavelength of the direct beam (the Kα line of Molybdenum.). Taken from A. H. Compton, Physical Review 22, 411 (1923), by courtesy of the American Physical Society.
336 The quantum nature of light and matter
where n is an integer, d is the crystal plane spacing, λ the electron wavelength and θ the angle between the electron path and the crystal surface. If tiny material particles like electrons have wave properties then what implications does this have for macroscopic objects? A mass of one gram moving at 1 m s−1 has a de Broglie wavelength of 6.6 10−31 m so that we cannot expect to see diﬀraction of everyday objects. However because the de Broglie wavelength of electrons of around 2 eV is 1 nm diﬀraction eﬀects will eventually impose a lower limit on the size of the gates in ﬁeld eﬀect transistors, and hence on the ultimate density of components in electronic processors. Nowadays the wave nature of material particles is exploited when using electron microscopes and neutron diﬀraction to explore the structure of matter.
12.6
The Bohr model of the atom
In parallel to the discovery of the quantum properties of electromagnetic radiation, experiments on atomic structure led to the appreciation that quantization was crucial in understanding the atom and atomic spectra. Rutherford in 1911 used αparticles (bare 4 He nuclei, each with charge +2e) to bombard thin metal foils and had observed that substantial numbers were scattered into the backward hemisphere and some almost straight backward. His observations could only be consistently explained if the object within the atom which was scattering the αparticles carries most of the atomic mass, positive charge and is very much smaller than the atom. This scatterer is the nucleus which consists of neutral neutrons and charged protons and is typically 10−15 m across. An electron has 1/2000th the mass of the proton or neutron and carries an equal and opposite charge to the proton. The electrons in an atom circulate around the nucleus in orbits extending to 10−10 m in diameter. Classically the electrons in such an atom must radiate continuously because they are accelerating radially. As a result the electrons would be expected to spiral rapidly into the nucleus while radiating over a broad frequency range. By contrast isolated atoms radiate at discrete wavelengths, the spectral lines, and their wavelengths form patterns that invite explanation. Hydrogen has a particularly simple atomic structure, with one electron orbiting a single proton nucleus. The emission spectrum of hydrogen gas excited by an electric discharge was already known at the time to consist of several series of spectral lines that could be ﬁtted with one overall empirical formula 1/λ = RH (1/n2 − 1/p2 ),
(12.22)
where RH is a constant known as the Rydberg constant. n and p are positive integers with the restriction that p > n. The currently accepted value of the Rydberg constant is 1.09678 107 m−1 . Thus the spectrum
12.6
consists of spectral series, each with a ﬁxed value of n while p runs through the sequence n + 1, n + 2, n + 3, etc. The Lyman series is generated by the combinations n = 1, p = 2, 3, 4, .... It starts with the Lyman αline at 121.6 nm (n = 1, p = 2) and terminates at 91.3 nm (n = 1, p = ∞). Among the other series, also named for their discoverers, are: the Balmer series (n = 2) lying in the visible and ultraviolet; the Paschen series (n = 3), the Brackett series (n = 4) and the Pfund series (n = 5), all lying in the infrared part of the spectrum. An explanation of these spectral features was achieved by Bohr in 1913 using a simple quantized model of the atom. In the Bohr model the electrons in atoms are pictured as travelling in stable circular orbits which satisfy the requirement that the angular momentum is exactly an integral multiple of h ¯ . The transition from an orbit of higher energy to one of lower energy is instantaneous and is accompanied by the emission of a photon which carries oﬀ the energy diﬀerence. Similarly an electron in a lower energy orbit can absorb a photon and instantaneously jump to a higher energy orbit provided the photon energy exactly matches the diﬀerence between the electron’s energy in the two orbits. The photon frequency, f , is given by hf = ∆E,
(12.23)
where ∆E is the diﬀerence in energy between the two atomic states. It is now shown how the pattern of spectral lines for hydrogen emerges naturally from these simple postulates. Suppose that in an atom containing a single electron the radius of the electron’s orbit is r, its speed is v and m is its mass. The quantization condition is mvr = n¯ h, or v = n¯ h/mr. (12.24) A justiﬁcation for Bohr’s quantization condition is revealed when the de Broglie relation from eqn. 12.19 is used to replace the momentum in eqn. 12.24. This gives n¯ h = pr = hr/λ, thus nλ = 2πr.
(12.25)
Therefore the quantization condition requires that one complete orbit should contain an integral number of electron wavelengths. If the orbit length did not satisfy this condition then the electron wave, after travelling many times around the orbit, would interfere destructively with itself. The electron is maintained in the stable orbit by the Coulomb attraction of the nuclear charge. Thus its radial equation of motion is Ze2 /(4πε0 r2 ) = mv 2 /r,
(12.26)
The Bohr model of the atom 337
338 The quantum nature of light and matter
where −e is the electron charge and Ze is the nuclear charge, Z being unity for hydrogen. Using eqn. 12.24 to replace v in eqn. 12.26 gives Ze2 /4πε0 = n2 h ¯ 2 /mr, so that r = 4πε0 (n2 h ¯ 2 /mZe2 ).
(12.27)
The radius of the ﬁrst orbit in a hydrogen atom a∞ = 4πε0 (¯ h2 /me2 ) = 0.05292 nm
(12.28)
is called the Bohr radius and characterizes the linear size of atoms; other orbits have radii n2 larger. This prediction for the atomic size is consistent with the measured spacing of atoms in condensed matter if the outer electron orbits of one atom touch or slightly overlap those of an adjacent atom. A quantity needed later is the angular frequency of rotation of an electron in the nth Bohr orbit Ωn = v/r = n¯ h/mr2 = (Ze2 /4πε0 )2 (m/n3 h ¯ 3 ).
(12.29)
Finally the total energy of the electron in the nth orbit, made up of the kinetic and the potential energy, is En = mv 2 /2 − Ze2 /(4πε0 r). The term mv 2 can be replaced using eqn. 12.26 to give En = −Ze2 /(8πε0 r) ¯ 2 ). = −(Ze2 /4πε0 )2 (m/2n2 h
(12.30)
From this result the photon energy emitted/absorbed in a transition between the pth and nth orbit can be calculated, h2 )[1/n2 − 1/p2 ], ∆E = Ep − En = (Ze2 /4πε0 )2 (m/2¯
(12.31)
where p > n is required for this to be positive. The wavelength of the photon emitted in such a transition is given by 1/λ = ∆E/hc = R∞ Z 2 [1/n2 − 1/p2 ],
(12.32)
R∞ = (1/4πε0 )2 (me4 /4π¯ h3 c).
(12.33)
where This last expression needs some correction because the nucleus has been assumed to be immobile (of inﬁnite mass). The corrected value in the case that the nucleus has mass M is R = [mM/(M + m)](R∞ /m) = µ(R∞ /m),
(12.34)
where µ is called the reduced mass of the electron. Similarly the Bohr radius must be corrected ¯ 2 /µe2 . a0 = 4πε0 h
(12.35)
12.6
0 eV
Lyman
Balmer
1.51 eV 3.39 eV
Transitions
13.6 eV Spectral lines Energy 14 eV
Wavelength in nm
10 eV
100
6 eV
200
2 eV
400
800
Fig. 12.8 The energy levels in the hydrogen atom are displayed together with transitions producing the Lyman and Balmer series. For clarity only the ﬁrst few transitions are drawn. The spectral lines are shown below. Each series converges to a limit corresponding to a transition in which the electron just escapes from the atom with zero kinetic energy.
The prediction for the Rydberg constant from eqn. 12.33 using the accepted values for the constants agreed with the measured value. Thus the Bohr model predicts the experimentally observed spectral series and the Rydberg constant with precision.1 This is very convincing evidence for the quantization of atomic orbits. Figure 12.8 shows the energy levels of the hydrogen atom and some of the transitions. As the quantum number n in eqn. 12.30 increases the discrete energy levels pack more and more tightly together, converging on zero binding energy. Above this energy electrons are free and can have any positive energy. It follows that an electron in an atom can absorb any photon whose energy exceeds the electron’s binding energy and emerge as a free photoelectron. Only the Lyman series is seen in the absorption spectrum of hydrogen gas at low temperatures. The reason lies in the distribution of electrons between the energy levels in a collection of hydrogen atoms in thermal 1 Nowadays the success of quantum theory means that eqn. 12.33 is accepted: the measurement of the Rydberg constant is therefore one among several experimental measurements from which the constants appearing in eqn. 12.33 are determined.
The Bohr model of the atom 339
340 The quantum nature of light and matter
equilibrium. In thermal equilibrium at a temperature T K the number of hydrogen atoms with an electron in an excited state of energy ∆E above the lowest energy or ground state is given by the Boltzmann distribution N (∆E) = N (0) exp [−∆E/kB T )],
(12.36)
where N (0) is the number in the ground state. Therefore whenever the temperature is low enough that kB T is very much smaller than the excitation energy of the ﬁrst excited state the ground state is almost the only one occupied. Thus at a temperature of 20 ◦ C, for which kB T is 0.025 eV, absorptions in hydrogen gas will, almost entirely, involve Lyman line transitions from the ground state. It must be the case that at the large scale the quantum description becomes equivalent to the classical description because at the macroscopic scale classical mechanics does work well. Bohr therefore proposed a correspondence principle which states that at very high quantum numbers the quantum description merges with the classical description of phenomena. For example the angular frequency of a photon emitted in a transition between adjacent levels given by eqn. 12.31 approaches, in the limit of very large quantum numbers n, ωn,n+1 = (Ze2 /4πε0 )2 m/(n¯ h)3 .
(12.37)
This is precisely the angular frequency of rotation of the electron in the nth orbit met with in eqn. 12.29. Now according to classical wave theory the angular frequency of the radiation emitted by a rotating electron is just the electron’s rotation frequency, so we see that in a transition at high enough quantum numbers the quantum prediction has converged on the classical prediction.
12.6.1
Beyond hydrogen
The Bohr model fails to explain the spectrum of neutral helium, the next simplest element beyond hydrogen, and it was only following the development of quantum mechanics that a comprehensive explanation of all atomic spectra became possible. The Bohr model works moderately well for the alkali metals, whose atomic structure has a single electron outside a core of electrons tightly bound to the nucleus. Thus the electric ﬁeld felt by the singleton electron is that of the nucleus and the core of electrons, which is much like that of a hydrogen atom. A result useful in interpreting atomic structure within the Bohr model was obtained by Moseley in 1914. He compared the wavelengths of the sharp lines appearing in the Xray spectra produced when elements are bombarded by high energy electrons. These lines would be emitted when the projectile electron ejects an electron from one orbit and an electron in a higher energy orbit drops into the orbit vacated. For the
12.7
shortest wavelength, and therefore highest energy Xray line, the Kα line, Moseley found that the wavelengths ﬁtted a single expression 1/λ = R(Z − a)2 ,
(12.38)
where Z is the atomic number, a is around unity and R is approximately equal to the Rydberg constant. This result is consistent with the interpretation that the Kα line is emitted in a transition between the second and ﬁrst Bohr orbit and that Ze is the nuclear charge. As with the alkali metals there is essentially a single electron involved, this time the electron closest to the nucleus. The net electric ﬁeld due to the outer electrons is small at its orbit and is responsible for the small correction ae to the nuclear charge Ze. This result is highly important: it reveals that the integral charge on the nucleus, Z, and hence the number of electrons in the atom, is equal to atomic number. The atomic number determines the position of an element in the periodic table, which is now seen to reﬂect the fact that the number of electrons in an atom determines its chemical properties.
12.6.2
Weaknesses of the Bohr model
It has been already noted that the Bohr model fails to explain the main features of spectra of elements beyond hydrogen. Its hybrid nature is also very unsatisfying: ﬁrst a quantization condition is imposed for closed orbits and after that classical mechanics is used. As a result there is no explanation as to why transitions occur and hence no way of calculating rates at which transitions occur. Furthermore in many physical situations there is no obvious periodic condition that can be imposed comparable to the condition on the angular momentum of an electron orbiting a nucleus. The resolution of all these diﬃculties only came with the development of quantum mechanics.
12.7
Wave–particle duality
The acceptance of the existence of wave and particle behaviour of both electromagnetic radiation and matter led to a search for ways to reconcile these two types of behaviour. A closely connected issue was the replacement of Bohr’s interim solution of grafting quantum conditions onto classical mechanics by a new coherent mathematical structure. In this section the interpretation of the wave–particle duality is discussed. Later in the following chapter the new quantum mechanics and its application to atomic structure is described. The connection between the particle and wave properties of light is statistical, and this is equally true for electrons or other material particles. To be explicit, the probability of ﬁnding a photon in a given volume
Wave–particle duality 341
342 The quantum nature of light and matter
dV is simply determined by the instantaneous energy density, I, of the electromagnetic wave over the same volume P dV = IdV / IdV , (12.39) where P is called the probability density. The integral is taken over the whole of space to ensure that the total probability of the photon being found somewhere is unity.
Photon count
4 3 2 1 0
Photon count
40
20
0
Photon count
1000
500
0 Fig. 12.9 Distribution of photons in the detection plane in Young’s two slit experiment for 10, 1000 and 20 000 photons. The broken curves indicate the classical interference pattern.
Young’s two slit experiment, met ﬁrst in Chapter 5, provides a simple application of this statistical picture. Suppose the observation screen is a pixelated detector with granularity much ﬁner than the fringe widths, and that all the pixels are equally eﬃcient in detecting photons. Fur
12.7
ther, suppose that an extremely low intensity, monochromatic source is used, suﬃciently weak so that at any given moment there is only ever a single photon within the volume between the source slit and the detector screen in ﬁgure 5.1. Figure 12.9 shows typical histograms of the photon distribution across the detector after 10, 1000 and 20 000 photons have been detected. For comparison the intensity calculated earlier in Chapter 5 is superposed in each case. An individual photon may hit anywhere across the screen, apart from locations where the wave intensity is precisely zero. Only the probability for arriving at each pixel is known, and the probabilities of reaching a given pixel are identical for each and every photon emerging from the source slit. The distribution observed when the number of photons is small is extremely ragged and does not resemble the wave intensity very closely. As the number increases the resemblance becomes ever closer, a convergence that is purely statistical. Recall that if the number of photons expected to strike a pixel is √ n, with n being large, the statistical error is√ n. Thus the fractional error on the number arriving at a pixel is 1/ n, and this fractional error falls as n rises. In a standard laboratory demonstration of Young’s two slit experiment there are very large numbers of photons arriving at the screen per second so that any visual perception of the underlying statistical ﬂuctuations is impossible. A two slit experiment with a very low source intensity was ﬁrst carried out in 1906 by Taylor and has been repeated many times. He used a photographic plate as the detector in an exposure lasting several weeks, after which the plate was processed and the anticipated fringes revealed. In this experiment and numerous similar ones there is at most only a single photon within the interference apparatus at any given moment. This has to mean that a photon is interfering with itself. The interference pattern does not change when multiple photons are in the apparatus so that we can conclude that there too each individual photon interferes with itself only. It is when photons have high energies that their particle behaviour is most noticeable. For example a single optical photon can produce an electron via the photoelectric eﬀect and this electron can be multiplied to give a detectable current pulse in a photomultiplier or avalanche photodiode. Particle behaviour is also evident in atomic, nuclear and molecular transitions. It is harder to associate particle behaviour with low frequency radio waves because each photon makes only a miniscule contribution to the total electric ﬁeld and each such contribution cannot be individually detected. However the statistical ﬂuctuation in the number of photons will be recognizable at low signal levels.
Wave–particle duality 343
344 The quantum nature of light and matter
12.8
The uncertainty principle
Uncertainty is inherent in quantum theory but this uncertainty is quantiﬁable. We start by considering a thought (gedanken) experiment which Heisenberg used to illustrate the uncertainty that occurs when simultaneous measurements are made of position and momentum. Figure 12.10 shows a microscope used to detect ultraviolet photons scattered from an electron. The precision in the measurement of the electron’s position is ﬁxed by the resolution of the microscope. Using eqn. 6.19 the angular radius of the Airy disk is taken to be ∆φ = λ/D, where D is the objective diameter and λ the wavelength of the radiation. Then the resolution in the electron’s position is ∆y = f ∆φ = λ/(2 sin θ), where f is the distance of the electron from the lens and θ is the semiangle subtended at the electron by the objective. Now the angle at which the photon enters the objective can be anywhere within an angular range of θ from the lens axis. Thus its momentum in the ydirection is uncertain by an amount ∆py = 2p sin θ, where p is the momentum of the incoming photon. Momentum conservation then requires that this is also the uncertainty in the electron momentum. Multiplying together the uncertainties in the electron momentum and position gives
Objective Aperture D
f θ
∆py ∆y = λp = h,
Scattered photon yaxis
Incident photon
Recoil electron
Fig. 12.10 Heisenberg’s microscope being used to determine the location of an electron.
where de Broglie’s relation has been used in making the second equality. The tightest limit on the product of uncertainties is obtained when the distributions in position and momentum have Gaussian rather than ﬂat distributions, in which case ∆py ∆y = h ¯ /2. If there are any instrumental errors in the measurement this can only increase these uncertainties. Taking into account that nonGaussian distributions are possible and that there can be additional instrumental errors leaves only a limit on the product ∆py ∆y ≥ h ¯ /2,
(12.40)
which is one expression of Heisenberg’s uncertainty principle. This principle brings into sharp relief a fundamental diﬀerence between classical and quantum theory. In the classical view it is imagined that the momentum of the light beam probing the electron can be indeﬁnitely reduced to make ∆py as small as required. However the reduction in ∆py can only be achieved in practice by increasing the photon wavelength, which
12.8
increases ∆y, but leaves the product of uncertainties unchanged. The uncertainty principle applies to the measurement in each of the three dimensions, from which it follows that simultaneous measurements of the vector position r and vector momentum p have uncertainties that satisfy ∆px ∆py ∆pz ∆x∆y∆z ≥ h ¯ 3 /8. (12.41) The product on the left hand side of this equation can be pictured as a volume element in a sixdimensional phase space. Three of the coordinates are the spatial coordinates, while the other three are the coordinates in momentum space. The inequality in the equation expresses the requirement that there is a limiting precision within which the kinematics of any particle can be known. The shape of this volume is dictated by the circumstances of the measurements, and is not necessarily cubical or spherical. Spatial coordinates and time are treated in a uniﬁed way within the special theory of relativity: they merge to form a single spacetime location (x, y, z, ct). Similarly energy and momentum form an energy– momentum vector (px , py , pz , E/c). One implication of this uniﬁcation is that there must exist an energy–time uncertainty relation of the form ∆E∆t ≥ h ¯ /2.
(12.42)
The interpretation of this uncertainty relation diﬀers from that of the previous example because time is not a quantity being measured. Rather ∆t is the time taken to make the measurement of the energy, and ∆E is the resulting uncertainty in the energy measured. Thus if a measurement is made of the energy of an excited state of an atom which decays with a lifetime τ the measurement of its energy has an uncertainty of at least h ¯ /2τ . The consequent uncertainty in the angular frequency of the light emitted in transitions from this state to the ground state is ∆ω ≥ 1/(2τ ). This can be recognized as the natural line width described in Section 7.3.1. In that section the origin of the natural linewidth was ascribed to a damping process, which is the classical equivalent of the exponential decay of excited states. Two other processes were described which in a typical source broaden the spectral lines much more than their natural width: these are collisions and the Doppler eﬀect. The eﬀect of collisions is to reduce the time available for measurement, while the Doppler effect arises from the thermal motion of the atoms which emit the photons. Quantum uncertainty relations bear a resemblance to the bandwidth theorems proved for classical wavepackets in Chapter 7. Expressed here as a wavefunction rather than a distribution in electric ﬁeld the time and angular frequency distributions of a Gaussian wavepacket have the form ψ(t) ∼ exp (−iωt) exp [−(t − t0 )2 /2σt2 ] (12.43) and angular frequency distribution ψ(ω) ∼ exp [−(ω − ω0 )2 /2σω2 ],
(12.44)
The uncertainty principle 345
346 The quantum nature of light and matter
for which σt σω = 1. The corresponding distributions in time and angular frequency are  ψ(t) 2 ∼ exp [−(t − t0 )2 /σt2 ]
(12.45)
and  ψ(ω) 2 ∼ exp [−(ω − ω0 )2 /σω2 ],
(12.46) √ for √ which the root mean square deviations are ∆ω = σω / 2 and ∆t = σt / 2. Then ∆t∆ω = 1/2. (12.47) For any other distribution the product is greater, so we have a bandwidth theorem ∆t∆ω ≥ 1/2. The corresponding relation linking position and wave vector errors is ∆x∆k ≥ 1/2. If these expressions are multiplied by h ¯ the outcomes resemble the quantum uncertainty relations! There is an important distinction which this direct conversion obscures. In the classical view the shape of a wavepacket could be measured precisely to give the shape parameters σx and σk . In the quantum picture ∆x and ∆k are the uncertainties in the measurements on an individual photon. The laws of classical physics are deterministic and statistical analysis is used as a practical tool only in dealing with systems containing very large numbers of particles. A good example is to be found in the kinetic theory of gases. By contrast, statistical behaviour is fundamental to even the simplest quantum systems. It is informative to view diﬀraction as an artefact of the uncertainty principle. For example consider the diﬀraction pattern produced by a slit of width d on which light of wavelength λ is incident. The uncertainty in the lateral position of any photon going through the slit is d and using the uncertainty principle this implies an uncertainty in its transverse momentum of at least ∆py = h ¯ /2d. Consequently the angular spread of the photons is at least ¯ /(2dp) = λ/(4πd), ∆θ = ∆py /p = h which indeed reproduces the proportionality of the angular size of the ﬁrst bright fringe to the quotient (wavelength)/(slit width).
12.9
12.9
Which path information 347
Which path information λ D/d
It goes without saying that to produce a two slit interference pattern the electromagnetic waves should pass through both slits. On the other hand a photon, a particle, ought to pass through one slit or the other. It is only because there is no information about which slit the photon passed through that an interference pattern is observed. With a symmetric pair of slits the wave amplitudes at each slit are equal and the probability of the photon being at either slit is 0.5. That is all we can know when we observe the interference pattern with its visibility of 100%. If on the other hand the photon is tagged in some way to identify its path the two slit interference pattern is inevitably destroyed. In such cases the wave amplitude at the slit where the photon is not detected must be zero, and the wave at the other slit naturally produces a single slit diﬀraction pattern. Einstein suggested that it might be feasible to determine which slit the photon had passed through by measuring the recoil of the board in which the slits are cut, and still allow an interference pattern to be seen. Bohr demonstrated that this method fails because of the uncertainty principle. Figure 12.11 shows the arrangement. The momentum transfer to the board will depend on whether the photon is deﬂected leftward by the right hand slit or rightward by the left hand slit. The diﬀerence in momentum transfer between the two cases is pd/D, where p is the photon momentum, d is the slit spacing and D is the slit/screen separation. In order that the measurement made on the board’s lateral momentum can be sensitive to this diﬀerence the precision of measurement should satisfy ∆py < pd/D. Next consider the eﬀect of the board’s movement sideways on the fringe pattern. Unless the position is known to better than the fringe spacing the fringe pattern will be smeared out and lost. Thus the lateral position of the board must be known to a precision ∆y < λD/d. Multiplying these two requirements together gives ∆py ∆y < pλ. The right hand side of the equation is just h. Therefore the proposed measurement to label the slit through which the photon passes and preserve the interference fringes would violate the uncertainty principle. One can either observe which slit the photon travels through or observe the two slit interference fringes, but not both. Put another way, the interference relies on the choice of paths taken by the photon being indistinguishable. If knowledge of the path (welcher Weg information) is complete the interference fringes are no longer detectable.
Screen
D
Board
d
Light source Fig. 12.11 Young’s two slit experiment.
348 The quantum nature of light and matter
There can however be intermediate situations in which there is partial information about the path and correspondingly an interference pattern with reduced visibility. This can be demonstrated with the apparatus pictured in ﬁgure 12.12. Monochromatic light is incident through a lens onto the slits, which are each covered by a polarizer. The transmission
CCD B Extraordinary
Polarizer slit a
Axis slit b Polarizer
Wollaston prism Ordinary CCD A
Fig. 12.12 An experiment to demonstrate partial information about the photon’s path and the resulting reduced fringe visibility. The Wollaston prism and CCDs are mounted on a common frame which can be rotated about the axis indicated. The diagram is adapted from that of L.S. Bartell, Physical Review D21, 1698 (1980). Courtesy of Professor Bartell and the American Physical Society.
2
See Section 10.5.2.
axes of these polarizers are set at +45◦ and −45◦ to a line perpendicular to the paper for slits a and b respectively. The Wollaston prism2 separates the ordinary and extraordinary components of the light so that they are focused by the lens onto the CCDs A and B respectively. The Wollaston prism and the two CCDs are mounted in a frame that can be rotated about the axis indicated by the dotted line in the ﬁgure. If the amplitudes of the electric ﬁelds on slits a and b are both E, then their √ components with polarization perpendicular to the paper are both E/ 2 and are extraordinary waves on entering the Wollaston prism. Similarly the√ components linearly polarized in the plane of the paper are both E/ 2, and have ordinary polarization in the Wollaston prism. Consequently interfrence fringes are seen at both CCDs, and it is impossible to say through which slit any photon travelled. Next suppose that the Wollaston prism and the detectors are rotated together through 45◦ around the dotted line axis. Light emerging from slit a now enters the Wollaston prism with ordinary polarization, while the light from slit b has extraordinary polarization. Therefore all the light from slit a(b) is directed to CCD A(B). The path is known now and there are no longer any two slit fringes to be seen at either CCD A or CCD B. Finally the WollastonCCD array can be set at some intermediate orientation so that there is partial knowledge about the slit through which each photon passes. For example if the choice of orientation gives a 90% chance that any photon arriving at CCD A(B) comes from slit a(b), then the fringe pattern reappears but with a visibility reduced to around 60%.
12.10
12.10
349
Wavepackets and modes
In earlier chapters the emission of electromagnetic radiation from sources was pictured as taking the form of a stream of ﬁnite length wavepackets. These wavepackets can now be recognized for what they are: the wavetrains describing individual photons emitted in an atomic transition.3 The duration of wavepackets is determined by the lifetime of the initial state τ and hence the corresponding quantum uncertainty in the photon energy is ∆E = h/τ . Any instrumental errors will lead to a larger overall uncertainty. A further important connection can now be established between the uncertainty principle and the modes of an electromagnetic ﬁeld introduced in Chapter 9. The latter are solutions of Maxwell’s equations which satisfy the boundary conditions imposed by the particular optical system involved. They are the independent orthogonal states of the electromagnetic ﬁeld in this system. One such mode is the Gaussian mode which propagates freely in a Fabry–Perot cavity. In the simple case of the electromagnetic ﬁeld within a closed box discussed in Section 9.8 the modes have discrete wavelengths: the components of the wave vectors were shown to be integral multiples of π/L, where L is the box dimension considered. Thus the volume in wave vector space occupied by a single mode is ∆kx ∆ky ∆kz = π 3 /(Lx Ly Lz ). h gives Rearranging this and noting that ∆ki = ∆pi /¯ ¯ 3. Lx Ly Lz ∆px ∆py ∆pz = π 3 h This simple argument reproduces, apart from a numerical factor, eqn. 12.41 for the minimum volume in sixdimensional phase space within which a photon can be conﬁned. Thus it can be inferred that, in general, and not just in the case of a rectangular box, the modes of the electromagnetic ﬁeld for a particular set of boundary conditions represent the ﬁnest granularity into which the sixdimensional phase space can be divided consistent with the uncertainty principle.
12.10.1
Wavepackets and modes
Etendue
Etendue was deﬁned in Chapter 4 as the product of the area and the solid angle of the light beam transmitted through an optical system. Multiplying the irradiance of the incident light beam by the etendue determines the radiant ﬂux through the system. Here we examine the implications of the uncertainty principle on how ﬁnely the etendue can be resolved.
3
Laser radiation is distinctive because a wavepacket contains large numbers of photons that have been emitted in phase. A detailed description of lasers is given in Chapter 14.
350 The quantum nature of light and matter
Beams in optical systems are usually paraxial so that the component of the photon momentum in the direction of the beam axis, the zdirection, is much greater than the transverse momentum components. Thus the uncertainty in this longitudinal component of momentum is almost the same as that in the total momentum, ∆pz ≈ ∆p = h ¯ ∆k = h ¯ ∆ω/c, where ∆ω is the spread in angular frequency of the light beam. Therefore the uncertainty in the longitudinal momentum is determined by the frequency spread. Suppose the beam has a width in the lateral xcoordinate of ∆x, and an angular spread of ∆θx . The uncertainty in the position of a photon is thus ∆x, and that in the angle it makes with the xaxis is ∆θx . There are corresponding uncertainties in the ydirection. The beam etendue is then T = ∆x∆y∆θx ∆θy = (∆px /p)(∆py /p)∆x∆y.
(12.48)
Now it was shown in Section 5.5.3 that the etendue from an aperture into its coherence area is Tc = λ2 . (12.49) Thus if the beam proﬁle exactly matches the coherence area, it follows that (12.50) ∆x∆y∆px ∆py = λ2 p2 = h2 , which reproduces the limit on this product imposed by the uncertainty principle. The result demonstrates that coherence area of the beam is the smallest meaningful division of the etendue.
12.11
Afterword
The interpretation of quantum phenomena introduced in this chapter and used hereafter is the generally accepted view originating with Bohr. In essence it accepts the probabilistic connection between waves and particles and does not consider any deeper explanation. Einstein, for one, was extremely uncomfortable with the idea that nature is probabilistic. Numerous experiments designed to seek ﬂaws in the predictions of the standard interpretation have been carried out and have provided valuable insights into quantum behaviour. However, more than 80 years later, tests have yet to reveal any discrepancies between the predictions based on the standard interpretation and the data. The remaining chapters of the book will build on the understanding developed in this present chapter of the dual nature of electromagnetic radiation and matter. In the 1920s and 1930s a new consistent quantum
12.12
Further reading 351
mechanics replaced the attempt by Bohr to describe atomic structure with an ad hoc mixture of classical mechanics and quantum conditions. The wholesale success obtained with this new quantum mechanics in predicting the details of atomic structure and spectra, including features not accounted for by the Bohr model, is described in Chapter 13 onwards. Chapter 13 also describes the associated discovery of the electron’s intrinsic angular momentum, or spin, and relates how the circular polarization of electromagnetic radiation is connected to the photon spin. Photons and electrons obey contrasting quantum statistics, which also diﬀer from classical statistics. In Chapter 14 the interaction of radiation with matter, the principles of lasers and their applications are outlined. Detectors of radiation are described in Chapter 15 while an account of modern communication systems based on optical ﬁbres is presented in Chapter 16. Chapter 17 is used to give an account of the interactions of electromagnetic waves, particularly laser beams, with atoms. The standard semiclassical theory is used. A further step in developing the quantum theory of radiation is taken in Chapter 18, where creation and annihilation operators for photons are introduced, a step that is known as second quantization. Modern experimental techniques for manipulating photons are introduced in the same chapter, in particular the study of correlations between photons and the production of photons in entangled states.
12.12
Further reading
Crucial Experiments in Modern Physics by G. L. Trigg, published by Van Nostrand Reinhold company (1971). This gives accounts of selected experiments with details drawn from the original papers.
Exercises (12.1) Solar radiation falling on satellites causes the emis (12.3) Calculate the wavelengths of a photon with energy 1 eV, and of an electron with kinetic energy 1 eV. sion of electrons. In order to reduce the charging eﬀect that this produces, the surface can be coated with platinum whch has a high work func (12.4) Calculate the relative proportions of hydrogen atoms in the ground and ﬁrst excited state in gas tion, 6.33 eV. Calculate the cutoﬀ frequency and in thermal equilibrium at 100 000 K and at 1000 K. wavelength for the photoelectric eﬀect on platinum. Would the Balmer and Lyman absorption lines be Roughly how much of the solar spectrum’s energy seen in the spectrum of radiation after passing does this exclude? through hydrogen gas at the two temperatures? (12.2) What is the wavelength shift of an Xray when scattered through 17◦ by an electron initially at rest? (12.5) In a certain atomic transition the parent and In ﬁgure 12.7 what is the origin of the left hand daughter states have lifetimes 3 10−8 and 4 10−8 s. peak in each panel which appears undisplaced in What is the natural width of the transition in terms wavelength from the incident Xrays? of the photon energy?
352 The quantum nature of light and matter (12.6) The Sun’s surface temperature is around 6000 K. What is the peak wavelength of the spectrum and what is the surface irradiance (intensity)? What is the peak wavelength of radiation from a black (12.12) body at room temperature, and of the cosmic background radiation at 2.75 K?
wavelength of 500 nm and a spread of 0.1 nm in wavelength. Its angular spread is 0.01 rad. What is the coherence volume? In the process of photoluminescence a material is irradiated with light and light of longer wavelengths is reemitted by the material. Why is this light of longer rather than shorter wavelengths?
(12.7) Calculate the longest and shortest wavelengths in the Balmer, Paschen and Brackett series in the hy(12.13) In ﬁgure 12.12 the Wollaston prism and CCDs are drogen spectrum. rotated so that there is a 90% probability that the (12.8) In positronium an electron is bound to a positron, light reaching CCD A is from slit a. Calculate the the positron being the antiparticle of the electron visibility at the centre of the fringe pattern on eiwhich has the same mass as the electron, but opther CCD. The illumination of the slits can be asposite electric charge. What is the wavelength of sumed equal. the Lyman αline of positronium? (12.9) Calculate Planck’s constant from the slopes of the (12.14) The cosmic microwave background last interacted with matter when the photons were still energetic lines drawn through the data points in ﬁgure 12.4. enough to ionize hydrogen gas. After this decou(12.10) A camera is used to photograph a distant scene. pling era the wavelength of the photons stretched calculate the uncertainty in the transverse momenas the fabric of the universe expanded. If the photum of a photon arriving at the image plane in ton energy at the spectral peak was 1.5 eV at the terms of the f/#, and the wavelength. Hence caldecoupling era what was the temperature in the culate the image resolution. universe at that time? By what factor has the universe since expanded? (12.11) A quasimonochromatic light beam has a mean
Quantum mechanics and the atom 13.1
Introduction
Despite its successes the Bohr model of the atom is conceptually unsatisfactory in having quantum conditions grafted onto classical mechanics. Its successes lay in explaining the spectra of hydrogenlike atoms. Crucially it does not provide a way to calculate the relative intensities of spectral lines and fails to explain the spectra of other elements adequately. The more comprehensive and purely quantum mechanical analysis of atoms developed in the 1920s and 1930s is presented in this chapter. In the ﬁrst part of the chapter the elements of quantum mechanics are introduced leading to Schroedinger’s equation, a wave equation which describes the motion of material particles. Generally we shall only be concerned here with the motion of electrons. Schroedinger’s equation is applied ﬁrst to determine the motion of an electron in a onedimensional square potential well. This simple case of a quantum well serves to demonstrate the basic properties of quantized bound states of electrons. Practical examples of quantum wells appear in quantum well lasers and optical modulators. Then the case of motion in an harmonic oscillator potential is treated, whose energy states turn out to resemble those of the em ﬁeld. After this the motion of an electron in a hydrogenlike atom is analysed. Quantization of the energy and angular momentum of the electron emerge in a natural way from the analysis. Pauli’s exclusion principle solved the riddle of why the electrons in an atom do not all enter the lowest energy level: there can never be more than one electron in any given quantum state. The exclusion principle and the related discovery of the intrinsic angular momentum, or spin of the electron are described in the next section. This leads into a discussion of multielectron atoms, the splitting of energy levels and the selection rules governing transitions. In the succeeding section of the chapter measurements to determine the momentum and the angular momentum of photons are recounted It emerges that the photon too has an intrinsic angular momentum (spin) and that in circularly polarized beams the photons have their spins aligned with their direction of travel. A comparison of the con
13
354 Quantum mechanics and the atom
trasting statistical behaviour of electrons and photons follows. The ﬁnal section compares the classical and quantum interpretations of line widths and decay rates.
13.2
Dirac produced a relativistic theory which was developed into a comprehensive relativistic ﬁeld theory of electromagnetic radiation and its interaction with charged particles by Feynmann, Tomonaga and Schwinger, and is known as quantum electrodynamics (QED).
An outline of quantum mechanics
The new mathematical framework to deal with quantum phenomena was developed by Schroedinger, Heisenberg, Born and Jordan. This nonrelativistic quantum mechanics is adequate for our purposes but is strictly applicable only if the material particles travel at velocities very much less than c. The key achievement was to arrive at a wave equation for the electron analogous to the electromagnetic wave equation for the photon. This wave equation replaces Newton’s equation of motion for material particles. An underlying assumption is that the universal laws of conservation of energy, momentum and angular momentum should remain valid in quantum mechanics. It is postulated that the behaviour of a single electron or a set of electrons is described by a complex wavefunction which contains all possible information that exists about the system. This wavefunction, Ψ(qn , t), is a function of time and all the independent variables, written as a set {qn }. These variables could be the spatial coordinates for a single electron (plus its polarization – if it should have any). The interpretation of the wavefunction parallels the interpretation of electromagnetic waves when locating a photon: the probability for ﬁnding a system with variables in a range dV = dq1 dq2 ... around q1 , q2 ,... is deﬁned to be P (q1 , q2 , ...) dV = Ψ∗ ΨdV.
(13.1)
The wavefunction used is normalized, meaning that a numerical factor is inserted so that integrating P dV over the full range of the independent variables gives unity. In the case of a single electron P dV is simply the probability of ﬁnding the electron within the spatial volume dV . The formal development of quantum mechanics presented below has basic features that apply equally to photons and electromagnetic waves. The treatments of electrons and photons diverge because the wave equations are diﬀerent and, as we shall see later, because any number of photons can share the same quantum state while only a single electron can ever occupy any given quantum state. Quantities that are measurable for a particle or a system of particles are known as observables. Position, momentum, orbital angular momentum, polarization and energy are all observables. It goes without saying that the measurements of observables give real and not complex numbers. As a ﬁrst example these ideas are illustrated for a free electron moving in one dimension. It is described by a wavefunction which is simply a plane wave travelling in the xdirection √ Φk = (1/ L) exp [i(kx − ωt)], (13.2)
13.3
where L is a very large range in x to which the electron is restricted and which will be increased to inﬁnity as required.1 The operators for momentum and the total energy are deﬁned as follows and indicated by placing hats over the respective symbols for the observable ∂ , (13.3) ∂x ∂ ˆ = +i¯ (13.4) E h . ∂t When these operators act on a plane sinusoidal wavefunction they give pˆ = −i¯ h
∂Φk =h ¯ kΦk = pΦk , (13.5) ∂x ∂Φk ˆ k = +i¯ =h ¯ ωΦk = EΦk . h (13.6) EΦ ∂t The quantities p and E appearing on the right hand side, without hats, are the values that would be obtained in measurements of the momentum and kinetic energy respectively.2 It is argued that these operators for momentum and total energy should apply equally for wavefunctions in general because any wavefunction can be Fourier analysed into linear sums of sinusoidal waves.
Schroedinger’s equation 355
1 Plane waves provide a simple example for discussion and ﬁnite realistic wavepackets are all linear sums of plane waves. Plane sinusoidal waves extend to inﬁnity and the range L is needed to give a normalizable wavefunction. The values of measurable quantities are correctly predicted when the limit L → ∞ is taken. On occasion care is needed when taking the limit.
pˆΦk = −i¯ h
The motion of any free electron would be described by a wavepacket made up of a superposition of plane sinusoidal waves, analogous to a photon wavepacket. The Fourier transform of the wavepacket is a frequency distribution from which a momentum distribution can be calculated. Then the result of measuring the electron momentum once would be some random value within this distribution. However making measurements on a large set of electrons with identical wavepackets would reproduce the distribution determined by Fourier analysis. The mean value of such a set of measurements is called the expectation value. Because the operators are complex and the quantities measured are real it follows that the waves for electrons and other material particles must themselves be complex, unlike electromagnetic ﬁelds which are real. Here the wave equation for electrons will be introduced ﬁrst and then solutions obtained for simple potentials.
13.3
Schroedinger’s equation
Schroedinger constructed an operator equation with which to analyse the nonrelativistic motion of electrons. The starting point is to write the law of conservation of energy for an electron moving in some potential V (r) E = V (r) + p2 /2m, (13.7) where E is the total energy and p2 /2m is the kinetic energy of the electron. Generalizing eqn. 13.3, the operator equivalent of p2 = p2x + p2y + p2z
2
This prediction of an exact value for the momentum appears to violate the uncertainty principle. However the likelihood of ﬁnding the electron in any interval, Φ∗ Φdx = dx/L, is the same everywhere. Then the position, which is the conjugate variable to the momentum, is indeterminate.
356 Quantum mechanics and the atom
is simply h2 (∂ 2 /∂x2 + ∂ 2 /∂y 2 + ∂ 2 /∂z 2 ). −¯ h2 ∇2 = −¯ This result can be used to convert eqn. 13.7 into a notional equation between operators i¯ h∂/∂t = V (r) − (¯ h2 /2m)∇2 .
(13.8)
A valid wave equation results if the operators act on the electron wavefunction Ψ(r, t). The result is i¯ h∂Ψ(r, t)/∂t = V (r)Ψ(r, t) − (¯ h2 /2m)∇2 Ψ(r, t),
(13.9)
which is called Schroedinger’s time dependent equation. Its solution describes the motion of the electron in the chosen potential V (r). If, as here, the potential does not vary with time the solution factorizes to give Ψ(r, t) = ψ(r) exp (−iEt/¯ h) (13.10) which when substituted in Schroedinger’s equation gives its time independent form Eψ(r) = V (r)ψ(r) − (¯ h2 /2m)∇2 ψ(r), ψ x
x
where E is the electron kinetic plus potential energy. In the case that the electron is free this reduces to Eψ(r) = −(¯ h2 /2m)∇2 ψ(r).
∞ ∂ψ ∂x
(13.11)
(13.12)
A solution is x
x
ψk (r) = exp (ik · r), and if this is substituted in eqn. 13.12 it gives, as expected, E = h ¯ 2 k 2 /2m or E = p2 /2m.
∞ ∂2ψ ∂ x2
x
Fig. 13.1 The panels show unphysical discontinuities. In the left hand panel the wavefunction, has a discontinuity, it jumps up, which would require inﬁnite momentum. In the right hand panel the derivative has a similar discontinuity, and this would require inﬁnite energy.
Schroedinger’s equation is linear in ψ so that wavefunctions which satisfy the equation can be superposed to give another valid wavefunction just as solutions of the electromagnetic wave equation can be superposed. There are several crucial diﬀerences between, on the one hand Maxwell’s equations and electromagnetic waves, and on the other Schroedinger’s equation and electron waves. Schroedinger’s equation is complex and the electron waves are complex and not directly measurable: Maxwell’s equations are real and the electromagnetic ﬁelds are real and directly measurable. Note that it has often been useful in the preceding chapters to perform calculations using complex ﬁelds whose real parts are the actual electromagnetic ﬁelds. Another diﬀerence is that for Schroedinger’s equation to apply the motion of the electron must be nonrelativistic, whereas Maxwell’s equations are fully relativistic. All the external inﬂuence on an electron is absorbed into a static potential V in Schroedinger’s equation, which is only adequate to describe electrostatic ﬁelds. Despite
13.3
these limitations a basic understanding of atomic states and their radiation is achieved by applying Schroedinger’s equation. Any solution of Schroedinger’s equation must satisfy several simple requirements. Firstly the wavefunction must be ﬁnite everywhere in order that the probability of ﬁnding the electron is ﬁnite everywhere. The next requirement is that the wavefunction is continuous and single valued everywhere. If instead the wavefunction jumped discontinuously, as shown in the left hand panel of ﬁgure 13.1, the derivative would become inﬁnite at that point. Thus a measurement of momentum made over a region including this point would yield an inﬁnite momentum. Similarly the ﬁrst derivative must be continuous everywhere. Were this to jump discontinuously, as shown in the right hand panel of ﬁgure 13.1, then the second derivative would be inﬁnite. Referring back to eqns. 13.7 and 13.9 we see that this is impossible when both the energy and the potential are ﬁnite everywhere. These requirements on the continuity of the wavefunction and its derivative are essential tools when joining up solutions of Schroedinger’s equation at boundaries where the potential changes. We shall see that boundary conditions are at the root of the quantization of energy and other measurable quantities.
Energy xcoord. 0
V0
a
Fig. 13.2 The energy levels of eigenstates in the square potential well.
The square potential well
Before tackling the motion of an electron in the Coulomb potential of the nucleus the motion of an electron of mass m in a onedimensional square potential well will be studied. This example provides an uncluttered ﬁrst view of a quantum wavefunction for an electron in a potential well. The potential is drawn in ﬁgure 13.2, it has a value −V0 over the region −a/2 < x < a/2 and is zero elsewhere. Within the attractive well Schroedinger’s equation is (−¯ h2 /2m)d2 ψ/dx2 = (E + V0 )ψ (internal),
(13.13)
(13.14)
Bound states of the electron, for which E is negative and the kinetic energy, (E + V0 ), is positive are considered ﬁrst. A solution inside the well which is symmetric about the origin is ψi = Ai cos (ki x), where ki =
(13.15)
2m(E + V0 )/¯ h and Ai is some constant. Externally ψe = Ae exp (∓ke x)
8 6
4
while outside the potential it becomes (−¯ h2 /2m)d2 ψ/dx2 = Eψ (external).
10
ke a / 2
13.3.1
Schroedinger’s equation 357
(13.16)
√ where ke = −2mE/¯ h and Ae is another constant. The upper sign in the exponent is taken for x > a/2 and the lower sign for x < −a/2. The
2 0 0
5 ki a / 2
10
Fig. 13.3 Graphical method for obtaining solutions of Schroedinger’s equation for the square well potential of depth V0 and width a. ki is the wavenumber within the well, and ke that outside it. The curves are discussed in the text.
358 Quantum mechanics and the atom
opposite choices of sign for the exponentials would give wavefunctions growing exponentially with the distance from the well. These can be rejected because they grow inﬁnitely.
1 ψ(x)
0.8 0.6 0.4 0.2 0
n=0
Applying the requirements that the wavefunction and its ﬁrst derivative are continuous at the wall at x = a/2 gives a/2
0
+a/2
Ai ki sin (ki a/2) = ke Ae exp (−ke a/2).
1 n=1
ψ(x)
0.5 0 1
(13.17)
From the deﬁnitions of ki and ke we also have a/2
0
+a/2
1 n=2
0.5 ψ(x)
Dividing one equation by the other gives ke = ki tan (ki a/2).
0.5
0
0.5 1 a/2
0
+a/2
1 n=3
0.5 ψ(x)
Ai cos (ki a/2) = Ae exp (−ke a/2) and
0
(ki a/2)2 + (ke a/2)2 = ma2 V0 /(2¯ h2 ).
The last two equations can be solved simultaneously either by computer or graphically as exhibited in ﬁgure 13.3 where (ke a/2) is plotted as a function of (ki a/2) for a given potential V0 . The relation found in eqn. 13.17 is represented by the full lines, while the quarter circle represents eqn. 13.18 with ma2 V0 /(2¯ h2 ) taken to be 100. Simultaneous solutions to eqns. 13.17 and 13.18 lie at the points where these curves intersect. A second set of wavefunctions which are antisymmetric about the origin also satisfy Schroedinger’s equation for the square well. The waves inside the well have the form ψi = Bi sin (ki x),
0.5 1 a/2
0
+a/2
Fig. 13.4 The four wavefunctions of lowest energy satisfying the square well boundary conditions. They are labelled with the number of nodes within the well. Broken lines mark the well edges where classical motion would terminate.
(13.18)
(13.19)
where Bi is some constant. Outside the well ψe = Be exp (∓ke x)
(13.20)
where Be is another constant. For these wavefunctions the continuity conditions lead to a diﬀerent transcendental equation ke = −ki cot (ki a/2).
(13.21)
This equation is plotted with broken lines in ﬁgure 13.3. On this plot the simultaneous solutions to eqns. 13.21 and 13.18 lie at the intersections of the broken lines and the quarter circle. Then on ﬁgure 13.2 the energy levels of all seven solutions are shown using full and broken lines for the states with even and odd wavefunctions respectively. Finally the wavefunctions of the four lowest energy (most tightly bound) states are plotted in ﬁgure 13.4. The preceding analysis shows that bound states are restricted to discrete energies. Only then can the sinusoidal waves inside the well join smoothly onto a wave that decays exponentially outside the well. At
13.4
Another departure from classical behaviour illustrated by the solutions of the square well potential is that the wavefunction of a bound electron does not vanish in the region outside the well where its kinetic energy has become negative. Classically the electron would be conﬁned between the walls where its kinetic energy is positive and it would never penetrate beyond the well. The quantum wavefunction decays exponentially in this region and is analogous to the evanescent electromagnetic wave occuring in total internal reﬂection and described in Section 9.5. The parallel extends further to include frustrated total internal reﬂection. Figure 13.5 shows a potential barrier of ﬁnite width. Electrons incident from the left are reﬂected to give standing waves. In addition the electron’s wavefunction penetrates the potential barrier, and at the far side this exponentially decaying wave joins smoothly onto an oscillatory wave that travels away from the boundary. Electrons can therefore travel through a region where their kinetic energy is negative and emerge on the far side, a possibility which is absolutely forbidden to them in classical mechanics. This purely quantum process is called barrier penetration or tunnelling.
13.4
Eigenstates
The wavefunctions that are solutions of Schroedinger’s equation for simple potentials like the square well, and including the case of the free electron with zero potential energy, are known as energyeigenfunctions. The corresponding energies are called energy eigenvalues and the electron is said to be in an eigenstate of energy. An eigenstate may be an eigenstate of several observables with each taking unique values for a given eigenstate. These are then known as compatible or simultaneous observables: examples are the energy, the angular momentum and a component of the angular momentum of an electron in a hydrogen atom. In an eigenstate the measurement of these compatible observables leaves the electron in the eigenstate. It is worth repeating that the eigenvalues of energy are discrete when the potential localizes the electron in a potential well, but
Energy
V
E 0 x 1 0.5 ψ(x)
other energies the requirement of continuity at the boundary makes it necessary to have a sum of a decaying and an increasing exponential outside the well. No matter how little the electron’s energy diﬀers from the discrete value picked out by the solution of Schroedinger’s equation in ﬁgure 13.3 the exponentially increasing component of the wave outside the well will tend to inﬁnity at an inﬁnite distance and cannot describe electron states localized in the well. This restriction to states with discrete energies is a feature which distinguishes quantum mechanics from classical mechanics. Discrete energy states are met in atoms, molecules and in nuclei. Unbound electrons, that is to say electrons with positive energies, have wavefunctions that are oscillatory both inside and outside the potential well. The continuity conditions at the boundary can now be satisﬁed at any positive energy and so there is a continuum of allowed states extending from zero energy upwards.
Eigenstates 359
0
0.5 1 x
Fig. 13.5 Potential barrier and wavefunction penetration.
360 Quantum mechanics and the atom
continuous from zero up to any conceivable positive value when an electron is free. The existence and the properties of eigenstates generalize to systems of electrons and other material particles. Such a system has a set of eigenstates {φi } of observables such as A with eigenvalues {ai } respectively. With the standard notation the operator corresponding to A is ˆ and this acts on the wavefunction φi in the following way: A, ˆ i = ai φi , Aφ
(13.22)
meaning that any measurement of the observable A on the eigenstate φi always gives the eigenvalue ai .
13.4.1
Orthogonality of eigenstates
An intrinsic property of eigenstates is their orthogonality in the sense that the overlap integrals between the wavefunctions of any pair of them over all the free variables vanishes. Suppose φi and φj are two such eigenfunctions of an electron; then writing Schroedinger’s equation for φi and multiply it by φ∗j , h ¯2 2 ∗ ∇ + V φi = φ∗j Ei φi , φj − 2m then repeating the process with the wavefunctions the other way about h ¯2 2 ∇ + V φ∗j = φi Ej φ∗j . φi − 2m Subtracting one equation from the other and integrating the result over a volume much larger than the potential well gives 2 ∗ (Ei − Ej ) φj φi dV = −(¯ h /2m) (φ∗j ∇2 φi − φi ∇2 φ∗j ) dV = −(¯ h2 /2m) ∇ · (φ∗j ∇φi − φi ∇φ∗j ) dV . Using Gauss’ theorem the right hand side becomes a integral over a surface enclosing the volume considered, (Ei − Ej ) φ∗j φi dV = −(¯ h2 /2m) (φ∗j ∇φi − φi ∇φ∗j ) · dS. This new surface integral is evaluated over a surface suﬃciently remote from the potential well that the wavefunctions and their derivatives have vanished. Hence φ∗j φi dV = 0, Ej = Ei . (13.23) Eigenfunctions are usually normalized for convenience so that φ∗i φi dV = 1.
(13.24)
13.5
Thus
φ∗i φj dV = δij ,
(13.25)
where δij is the Kronecker δ deﬁned by δij = 0 for i = j, δij = 1 for i = j.
(13.26) (13.27)
With a free electron √ φk = (1/ L) exp [i(kx − ωt)] where the subscript k is a continuous variable rather than an integer label. Then using the Dirac δ function introduced in Chapter 7 and eqn. 7.17 φ∗k φk dx = δ(k − k ).
(13.28)
Similar properties hold good for the modes of electromagnetic radiation in an optical setup, because these are the eigenstates of Maxwell’s equations satisfying the boundary conditions imposed by the optical components. For example the standing electromagnetic waves possible in a Fabry–Perot cavity have discrete wavelengths and their waveforms are mutually orthogonal.
13.5
Expectation values
In the more general case that a system is not in an eigenstate of an observable, the value that is obtained by measuring the observable can only be predicted statistically. Quantum mechanics predicts the expecˆ and deﬁned by tation value of an observable A, which is written A, the equation ˆ = ψ ∗ Aψ ˆ dV , A (13.29) where ψ is normalized. The equation is to be interpreted in this way. Suppose that the same measurement of A is made on each of a large number of systems which have been prepared in exactly the same way so that they have identical wavefunctions ψ – such a hypothetical collection of systems is called an ensemble. Then the average value of the observable measured over the ensemble equals the expectation value. In the case of an eigenstate of the observable A the expectation value is simply the eigenvalue of A for that eigenstate. Any wavefunction ψ of a system which has an observable A can always be expanded as a linear superposition of the normalized eigenfunctions {φi } of A. Assuming for simplicity that the eigenvalues are discrete, ci φi . (13.30) ψ= i
Expectation values 361
362 Quantum mechanics and the atom
Then the expectation value of A in a state with wavefunction ψ is deﬁned to be ˆ ˆ dV A = ψ ∗ Aψ ˆ j dV = c∗i cj φ∗i Aφ i,j
=
c∗i cj aj
φ∗i φj dV
i,j
=
c∗j cj aj .
(13.31)
j
The interpretation of this result is that Pj = c∗j cj is just the probability that the measurement ﬁnds the value aj , or equally the probability that the system is found to be in the eigenstate φj . When the eigenvalues are continuous as for a free electron described by eqn. 13.2 ψ = c(k)φk dk. (13.32) A similar analysis to that given above then yields ˆ = c∗ (k)c(k)a(k) dk, A
(13.33)
where a(k) is the value obtained when A is measured on the eigenstate with momentum k. Now P (k)dk = c∗ (k)c(k)dk
(13.34)
is the probability that the measurement ﬁnds a momentum eigenvalue lying between k and k + dk. All actual measurements yield real values so that expectation values of observables have to be real. Thus for any observable A ∗ ∗ ˆ ∗ ˆ ψ Aψ dV = ψ Aψ dV = ψ Aˆ∗ ψ ∗ dV , and then
ˆ = ψ ∗ AψdV
ˆ ∗ ψ dV . (Aψ)
(13.35)
Operators which have this mathematical property for all wavefunctions of a system are called hermitean; thus observables are always represented by hermitean operators. The expectation values of observables are unchanged if the wavefunction is multiplied by a phase factor exp (iα) where α is real. Thus there is always a phase ambiguity in the wavefunctions.
13.5
13.5.1
Expectation values 363
Collapse of the wavefunction
If the measurement of the observable A on a system with wavefunction ψ gives the value ai , the system must immediately thereafter be in the eigenstate with wavefunction φi , and no longer in the state with wavefunction ψ. A second measurement of the observable A will again give ai , and so would further measurements. This result is profoundly diﬀerent from anything met in classical mechanics. The change is discontinuous: up to the exact moment of the measurement the system is evolving according to the wavefunction ψ and immediately afterwards its wavefunction has become φi . This step is known as the collapse of the wavefunction. Schroedinger highlighted the logical diﬃculty of an external observer causing the wavefunction to collapse by using a cat fable. The cat is locked in a box together with a mechanism which will release a lethal gas if and when a single radioactive nucleus decays. It is then argued that the wavefunction of the contents of the box should be a superposition of two wavefunctions: the ﬁrst for the undisturbed mechanism and a live cat; the second for an activated mechanism and a dead cat. Later Schroedinger opens the box and observes the contents. At this instant the wavefunction of the contents collapses to either one that contains a live cat, or to another that contains a dead cat. The generally favoured resolution of this paradox of a cat simultaneously alive and dead is through what is called decoherence. Broadly speaking any interaction of a quantum system with its surroundings is equivalent to making a measurement. For example the air molecules striking the cat are suﬃcient to collapse its wavefunction.3 Carrying the idea of having wavefunctions for macroscopic objects to its logical conclusion requires the universe to possess a wavefunction. In this extreme case all possible observers are part of the wavefunction. Whether this is a correct view; and whether, and how the wavefunction of the universe could collapse have led to considerable speculation.
13.5.2
Compatible, or simultaneous observables
Eigenstates of a system are usually eigenstates of several observables, and it requires knowledge of all of these to completely specify an eigenstate. Here we consider the case where there are just two of these compatible observables, A and B. There is a set of eigenfunctions {φ} for which ˆ j = aj φj ; Bφ ˆ j = bj φj , Aφ and some eigenfunctions will share the same eigenvalues for A but have diﬀerent eigenvalues of B. Measurement of A followed by a measurement of B on any arbitrary state will result in its wavefunction collapsing into an eigenfunction, φj . Further measurements thereafter of A and B
3
See for example ‘Decoherence and the the transition from quantum to classical’ by W. J. Zurek, Physics Today, October 1991.
364 Quantum mechanics and the atom
yield aj and bj respectively. The expectation value for the product of compatible observables ∗ ˆˆ ∗ ˆ Aψ ˆ dV , ψ ABψ dV = cj cj a j b j = ψ ∗ B j
holds true whatever arbitrary state ψ of the system is considered. Conˆ−B ˆ Aˆ always vanishes. This opsequently the expectation value of AˆB erator is called the commutator of A and B and is written with square ˆ B]. ˆ If, as in the case being considered brackets [A, ˆ B ˆ ] = 0, [A,
(13.36)
ˆ are said to commute. Aˆ and B The uncertainty principle revisited The operators of conjugate variables like the position and momentum do not commute
∂ ∂ − −ih x ψ = i¯ hψ. [x ˆ, pˆ ]ψ = x −ih ∂x ∂x For any such pair of conjugate variables, which we denote by F and G ˆ ] = i¯ [ Fˆ , G h.
(13.37)
This relationship is now shown to be directly related to the uncertainty principle. Writing δF for the operator Fˆ − Fˆ , the variance is deﬁned as (∆F )2 = Fˆ 2 − Fˆ 2 = [ Fˆ 2 − 2Fˆ Fˆ + Fˆ 2 ] = (δF )2
(13.38)
Again using eqn. 13.37 [ δF, δG ] = i¯ h, so that 2 ImδF δG = i¯ h, and hence δF δG2 > h ¯ 2 /4.
(13.39)
Using the Schwarz inequality and eqn. 13.38 (∆F )2 (∆G)2 = (δF )2 (δG)2 ≥ δF δG2 . Finally using eqn. 13.39 to replace the right hand side in this equation gives (∆F )2 (∆G)2 ≥ h ¯ 2 /4 (13.40) which reproduces the uncertainty principle. In general we can say that if the operators of two observables do not commute, then simultaneous measurements of their observables will obey an uncertainty relation.
13.6
13.6
The harmonic oscillator potential
365
The harmonic oscillator potential
A frequently met dynamical system is that of a mass undergoing simple harmonic motion in the xdimension under a restoring force linear in the displacement from the origin, kx. Its eigenstates are of particular interest because they are exact parallels to the eigenstates of black body radiation. This parallel will prove of value in Chapter 18. The potential energy is the integral of the force x V = kx dx = kx2 /2, 0
and this function is displayed in ﬁgure 13.6. Schroedinger’s time independent equation for this potential is then −(¯ h2 /2m) d2 ψ/dx2 + (kx2 /2)ψ = Eψ.
4.5 4 3.5 3 2.5 2 1.5 1 0.5 0
(13.41) d2 ψ/ds2 + (λ − s2 )ψ = 0, where λ = 2E/¯ hω0 , with ω0 = k/m being the natural frequency of oscillation of the mass in this potential. This equation has analytic solutions of the form ψ(s) = H(s) exp (−s2 /2),
Energy / h f 0
h2 and s = αx, the equation can be rewritten Setting α4 = km/¯
(13.42)
where H(s) is a polynomial. Solutions with a term exp (s2 /2) are excluded because they diverge at inﬁnity and are not conﬁned to the potential well. The derivatives are
2
0 s=αx
2
Fig. 13.6 The energy levels of eigenstates in the harmonic potential well.
dψ(s)/ds = [dH(s)/ds] exp (−s2 /2) − s H(s) exp (−s2 /2), d2 ψ(s)/ds2 = [d2 H(s)/ds2 ] exp (−s2 /2) − 2s [dH(s)/ds] exp (−s2 /2) +(s2 − 1) H(s) exp (−s2 /2), and when these are substituted into eqn. 13.41 the result is [d2 H(s)/ds2 ] − 2s [dH(s)/ds] + (λ − 1) H(s) = 0.
(13.43)
Finite polynomial solutions to this equation can be obtained provided that λ takes the discrete values4 λ = 2n + 1,
(13.44)
where n is any nonnegative integer. When λ is replaced by its deﬁned value, 2E/(¯ hω0 ), it immediately follows that the energy of the n’th eigenstate is En = h ¯ ω0 (n + 1/2). (13.45) The lowest energy solutions, labelled with the value of n as a subscript, are H0 (s) = 1; H1 (s) = 2s; H2 (s) = 4s2 − 2; H3 (s) = 8s3 − 12s. (13.46)
4
See Chapter 4 of the third edition of Quantum Mechanics by L. I. Schiﬀ, published by McGrawHill Kogakusha Ltd., Tokyo (1968).
366 Quantum mechanics and the atom
ψ(s)
These functions are called Hermite polynomials. There is a simple recurrrence relation derived from eqn. 13.43 that can be used to generate further members of the sequence Hn+1 (s) = 2sHn (s) − 2nHn−1 (s). n=0 2
0 s
2
4
ψ(s)
4
n=1 2
0 s
2
4
ψ(s)
4
n=2 2
0 s
2
4
ψ(s)
4
The resulting Gauss–Hermite solutions to the Schroedinger equation for the harmonic well, Hn (s) exp (−s2 /2), are shown in ﬁgure 13.7. Of these the lowest energy solution has the familiar Gaussian shape. Corresponding energy levels are indicated by the horizontal lines in ﬁgure 13.6. The classical state of lowest energy would be that in which the mass is at rest at x = 0 and has zero kinetic energy. By contrast the lowest energy quantum state with the Gaussian wavefunction has a positive energy h ¯ ω0 /2. This is called the zero point energy and is a characteristic of quantum systems. In the square well potential the lowest energy state is displaced upward in energy from −V0 so that there too the particle is not at rest in its lowest energy state. It is striking that the energy spectrum of the eigenstates in the harmonic well match those postulated by Planck for the modes of the em ﬁeld of angular frequency ω0 – apart from the displacement upward by the zero point energy. This feature will be considered further in Chapter 18, and it will emerge that the parallel is exact: when there are no photons in a mode of the electromagnetic ﬁeld it too has a zero point energy h ¯ ω0 /2! In the intervening chapters this uniform dispacement upward by h ¯ ω0 /2 of all the energy levels of any mode of electromagnetic radiation can be safely neglected because only the transitions between states are discussed.
n=3 4
2
0 s
2
4
Fig. 13.7 The four wavefunctions of lowest energy in a harmonic well, labelled with the quantum number n. The vertical dotted lines indicate the points at which the kinetic energy is zero, and where the classical motion terminates.
13.7
The hydrogen atom
Schroedinger’s analysis of the electron motion within the Coulomb potential due to the nuclear charge provided a detailed and precise description of atomic structure, which incorporated all the successes of the Bohr model. In a hydrogenlike atom with a nucleus carrying a charge Ze and having mass M , the Coulomb potential felt by the single electron of mass m is −Ze2 /(4πε0 r) at a distance r from the nucleus. This potential is drawn for the hydrogen atom in ﬁgure 13.8. Then the Schroedinger time independent equation is −(¯ h2 /2µ)∇2 ψ − Ze2 /(4πε0 r) = E,
(13.47)
where µ = mM/(M +m) is the reduced mass of the electron and E is the total energy of the atom. The same approach is taken as that applied in seeking solutions to the square well potential. Although the analysis is more complicated the solutions in this case are analytic. Here only the results are discussed, but full details of the solution can be found in
13.7
The hydrogen atom
367
Table 13.1 Table of radial eigenfunctions Rnl (r). n
l
Rnl (r)
1 2 2 3
0 0 1 0
2(Z/a0 )3/2 exp (−Zr/a0 ) √ (1/ √ 8)(Z/a0 )3/2 [2 − Zr/a0 ] exp (−Zr/2a0 ) (1/ 24)(Z/a0 )3/2 [Zr/a0 ] exp (−Zr/2a0 ) ( 4/3/81)(Z/a0 )3/2 [27 − 18Zr/a0 + 2(Zr/a0 )2 ] exp (−Zr/2a0 )
many standard texts on atomic physics or quantum mechanics.5 The eigenfunctions separate into radial and angular components, which are best written in spherical polar coordinates ψnlm (r, θ, φ) = Rnl (r) Ylml (θ, φ).
(13.48)
Each solution is identiﬁed by three integral quantum numbers n, l, ml . The function Rnl contains an associated Laguerre polynomial F (r), and Ylm is a spherical harmonic function. In general these have the forms Rnl (r) = exp (Cr/n) rl F (r), Ylml (θ, φ) = sin (ml θ) exp (iml φ) G(θ),
(13.49) (13.50)
where C is a constant and F (r) and G(θ) are polynomials in r and θ respectively. The exact forms of the eigenfunctions are given in Tables 13.1 and 13.2 for a few of the lowest values of n, l and ml . These wavefunctions are orthonormal in the usual sense that the volume integrals over all space are ∗ ψnlm ψn l ml dV = δn,n δl,l δml ,ml . (13.51) l Valid combinations of the quantum numbers are restricted to the following values:
0
l = 0, .... , n−2 , n−1; ml = −l , −l+1 , −l+2, ..... , l−2 , l−1, l.
10
(13.52)
The energy of an electron, its orbital angular momentum and a component of its angular momentum, which can be chosen to be the zcomponent, are the three compatible observables and their eigenvalues are speciﬁed by the quantum numbers n, l and ml respectively. The predicted energy eigenvalues duplicate those found with the Bohr model En = −µZ 2 e4 /[(4πε0 )2 2¯ h2 n2 ],
Energy in eV
5
n = 1, 2, 3, ..... ;
(13.53)
5 For example the third edition of Quantum Mechanics by L. I. Schiﬀ, published by McGrawHill Kogukusha, Tokyo (1968).
15
20 20
0 radius / a0
20
Fig. 13.8 The Coulomb potential due to the hydrogen nucleus and the electron energy levels. a0 is the Bohr radius.
368 Quantum mechanics and the atom
Density
Table 13.2 Table of angular eigenfunctions Ylm (θ, φ).
0.5 0.4 0.3 0.2 0.1 0 0
5
10 15 radius / a0
Density
0.15
n=2
0.1
l=0
20
25
5
10 15 radius / a0
20
25
Density
1
±1
2
0
2
±1
2
±2
√
1/ 4π
(3/4π) cos θ (3/8π) sin θ exp (±iφ)
(5/16π)[3 cos θ − 1] (15/8π) sin θ cos θ exp (±iφ) 2
(15/32π) sin2 θ exp (±2iφ)
0.15
n=2
0.1
l=1
Lz = xpy − ypx ,
The total orbital angular momentum operator, L, is given by ˆ2 = L ˆ 2x + L ˆ 2y + L ˆ 2z . L
[Lx , Ly ] = i¯ hLz . 5
10 15 radius / a0
(13.55)
ˆ 2 commutes with all the components Lx , Ly and Lz , but these comL ponents do not commute with each other. For example
0.05 20
25
0.1 n=3 l=0
(13.56)
This explains why the total orbital angular momentum and only one of its components can be compatible observables. For completeness the forms of these operators in spherical polar coordinates are ∂ 1 ∂2 1 ∂ ˆ 2 = −¯ L sin θ + (13.57) h2 sin θ ∂θ ∂θ sin2 θ ∂φ2 ∂ ˆ z = −i¯ and L h . (13.58) ∂φ
.
Density
0 0
from which the quantum mechanical operator can be obtained by replacing the position and momentum by their operator equivalents ∂ ∂ ˆ −y . (13.54) Lz = −i¯ h x ∂y ∂x
0.2
0.05
0 1
Angular momentum operators can be constructed from the momentum operators as follows. In classical mechanics the zcomponent of the vector angular momentum, L, is
0.05
0 0
m
so that n is called the principal or energy quantum number. n=1 l=0
0.2
0 0
Ylm (θ, φ)
l
0 0
5
10 15 20 25 radius / a0 Fig. 13.9 Radial electron density distributions in the hydrogen atom for the lowest energy eigenstates. The dotted lines indicate where the kinetic energy changes sign.
When these operate on the wavefunctions they give ˆ 2 ψnlm = l(l + 1)¯ h2 ψnlml (13.59) L l ˆ ¯ ψnlml . (13.60) and Lz ψnlml = ml h h, of the orbital anIt follows that l speciﬁes the magnitude, l(l + 1)¯ gular momentum, while ml h ¯ speciﬁes its component in the zdirection.
13.7
The historical spectral notation is used in labelling the eigenstates of the electron: an electron with orbital angular momentum quantum number l = 0, 1, 2, 3, 4, 5.... is said to be in an s, p, d, f, g, h,.... state. A pair of electrons with quantum numbers n = 2, l = 1 are described as being in the conﬁguration 2p2 . The radial distribution of the electron 2 probability density r2 Rnl is shown in ﬁgure 13.9 for a few values of n and 2 l, where the factor r is included to take account of the growth in the volume element as the radius increases, dV = r 2 sin θ dθ dφ dr. The dotted lines in this diagram indicate where the kinetic energy changes sign. Each wavefunction has (n − l) radial nodes. Figure 13.10 shows diametral sections containing the zaxis through the l = 1 electron probability distributions. The threedimensional distributions are obtained by rotating these planar distributions around the axis indicated by the broken line in ﬁgure 13.10. Note that the three 2p electronic wavefunctions give a combined electron density distribution that is spherically symmetric. Such spherically symmetric distributions always result when there is an electron in each of the 2l + 1 substates for a given l value; that is to say when the electron subshell is full. Figure 13.11 indicates the alignment of the orbital angular momentum vector about the zaxis for the eigenstates with l = 2. The orbital angular momentum vector makes a well deﬁned angle with the zaxis cos−1 {ml / [l(l + 1)]}, but its azimuthal position is indeterminate, reﬂecting the fact that only one component of orbital angular momentum can be an observable compatible with the total angular momentum. The eigenstates have a deﬁnite parity, that is to say the result of reﬂecting the coordinates in the origin causes the wavefunction to change by a factor ±1. A wavefunction with orbital angular momentum has a parity of (−1) . The wavelengths of photons emitted in transitions between the energy levels are discrete. Therefore if the light from a source is passed through a diﬀraction grating the photon will strike a point in the image plane that is determined by its wavelength. Detecting this location yields the wavelength and hence the eigenstates of the atom both before and after the transition. In general if the lifetime of the state decaying is τ the uncertainty in the energy of the photon is at least h ¯ /τ . If other transitions exist with energies lying within this range it is no longer possible to identify the states involved from the photon wavelength alone. Eigenstates sharing the same value of n but diﬀerent values of l and ml have identical energies and such eigenstates are termed degenerate. As will be seen later the degeneracy is lifted by relativistic eﬀects and by placing the atom in a constant magnetic or electric ﬁeld. The most intense lines in the hydrogen spectrum involve transitions in which l changes by unity ∆l = ±1,
(13.61)
The hydrogen atom
369
l=1; m=0
l=1; m=+1 or 1
Fig. 13.10 Diametral sections through the probability distributions in the hydrogen atom for electrons in the 2p shell. +2 h/2π +h/2π 0 h/2π 2 h/2π
m = +2 m = +1 m=0
m = 1 m = 2
Fig. 13.11 Orientations of the angular momentum vector for the eigenstates with the orbital angular momentum quantum number equal to 2.
370 Quantum mechanics and the atom
which is one example of a selection rule in optical spectra. The corresponding rule for the associated changes in the magnetic quatum number is ∆m = ±1, 0. (13.62) Transition rates for these electric dipole transitions are calculated in Chapter 17, where the selection rules are also discussed. Transitions between the states of a hydrogen atom in which the change in is diﬀerent from those speciﬁed in eqn. 13.61 occur at far lower rates and are known collectively as forbidden transitions.
13.8
The Stern–Gerlach experiment
Spatial quantization in which a component of the electron’s orbital angular momentum in some arbitrarily selected direction is quantized is another unexpected quantum feature. The details of spatial quantization are now developed and the experimental test of its consequences carried out by Stern and Gerlach will be described. We focus ﬁrst on a single electron in an atom. As it orbits nucleus it forms a current loop which has correspondingly a magnetic moment µ = Ai,
(13.63)
where A is the area of the loop and i is the current. If the electron’s orbit has radius r, the area A = πr 2 , while the current is the product of the electron charge multiplied by the number of times it passes any given point in its orbit in one second. Suppose the electron velocity is v, then this current is i = −e(v/2πr).
Magnet
Field lines
Hence the magnetic moment Magnet
Deflected atoms
Oven Collimator
Screen
Fig. 13.12 The Stern–Gerlach experiment. The upper panel shows a section through the magnet perpendicular to the beam direction. The lower panel shows a vertical section containing the incident beam.
µ = −evr/2 = −µB L/¯ h,
(13.64)
where L is the orbital angular momentum and µB is a natural unit of magnetic moment deﬁned to be e¯ h/2m, and called the Bohr magneton. Vectorially µ = −µB L/¯ h. (13.65) An external magnetic ﬁeld applied in what we take to be the zdirection will break the spherical symmetry of the atom’s environment and deﬁne a suitable quantization axis. The energy of the atom is altered by an amount Em = −µ · B = µB ml B, (13.66) where ml h ¯ is the zcomponent of the orbital angular momentum. ml is therefore called the magnetic quantum number. In a nonuniform magnetic ﬁeld varying in the zdirection the atom will experience a force F = −µB ml
∂B . ∂z
(13.67)
13.9
Stern and Gerlach carried out an experiment in 1922 which displayed this eﬀect of spatial quantization directly. Their experiment is illustrated in ﬁgure 13.12. Atoms evaporated from liquid silver in an oven are collimated into a beam by a series of slits in metal plates. The atoms then travel between the poles of a magnet whose faces are shaped to give a magnetic ﬁeld strength which varies strongly across the gap and hence has a large value of ∂B/∂z. Beyond the magnet the atoms travel some distance to a screen where they are detected. When the magnet is oﬀ the atoms travel undeviated and form a line image of the collimator slit on the screen. The force given by eqn. 13.67 lies in the zdirection, pointing from one pole piece to the other. In the classical view the magnetic moments of the atoms would be oriented in random directions and their angular deﬂections in the direction of the ﬁeld gradient would simply broaden the collimator slit image. However when the magnet was turned on the image on the screen changed to a pair of line images of the collimator slit well separated from one another in the direction of the ﬁeld gradient. This is precisely what is expected from eqn. 13.67, demonstrating the reality of spatial quantization. However a disturbing anomaly was noted. The number of components into which an atomic beam splits should be odd because 2l + 1 is an odd number; but in practice the number of components seen with some atomic elements, including silver, is even. A simple explanation of this anomaly was revealed a few years later with the discovery of the spin of the electron.
13.9
Electron spin
In 1925 Uhlenbeck and Goudsmit proposed that the electron has an intrinsic angular momentum or spin whose magnetic quantum number is 1/2. Therefore the magnitude and the quantized component of this spin have values s = (1/2)(3/2)¯ h , ms = ±¯ h/2. (13.68) The total angular momentum, j, of an electron must be the vector sum of its intrinsic and its orbital angular momenta j = s + l,
(13.69)
with magnetic quantum numbers mj in the sequence −j, −j + 1, ... ,+j. Overall an electron in an atom therefore