Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                

Advances in Optics Reviews Vol. 3

2018, Sergey Y. Yurish Editor. Advances in Optics: Reviews Book Series, Volume 3.

Preface It is my great pleasure to introduce the third volume of new Book Series ‘Advances in Optics: Reviews’ started by the IFSA Publishing in 2018. Three volumes were published in this year. The ‘Advances in Optics: Reviews’ Book Series is published as an Open Access Books in order to significantly increase the reach and impact of these volumes, which also published in two formats: electronic (pdf) with full-color illustrations and print (paperback). The third of three volumes of this Book Series has organized by topics of high interest. In order to offer a fast and easy reading of each topic, every chapter in this book is independent and self-contained. All chapters have the same structure: first an introduction to specific topic under study; second particular field description including sensing or/and measuring applications. Each of chapter is ending by complete list of carefully selected references with books, journals, conference proceedings and web sites. The Vol.3 is devoted to various topics of applied optics and contains 17 chapters written by 49 experts in the field from 14 countries: Australia, China, India, Israel, Italy, Japan, Malaysia, Mexico, The Netherlands, Poland, Taiwan, UK, USA and Vietnam. ‘Advances in Optics: Reviews’ Book Series is a comprehensive study of the field of optics, which provides readers with the most up-to-date coverage of optics, photonics and lasers with a good balance of practical and theoretical aspects. Directed towards both physicists and engineers this Book Series is also suitable for audiences focusing on applications of optics. A clear comprehensive presentation makes these books work well as both a teaching resources and a reference books. The book is intended for researchers and scientists in physics and optics, in academia and industry, as well as postgraduate students. I shall gratefully receive any advices, comments, suggestions and notes from readers to make the next volumes of ‘Advances in Optics: Reviews’ Book Series very interesting and useful. Dr. Sergey Y. Yurish Editor IFSA Publishing Barcelona, Spain.

Advances in Optics: Reviews Book Series, Volume 3 Sergey Y. Yurish Editor Advances in Optics: Reviews Book Series, Volume 3 International Frequency Sensor Association Publishing Sergey Y. Yurish Editor Advances in Optics: Reviews Book Series, Vol. 3 Published by International Frequency Sensor Association (IFSA) Publishing, S. L., 2018 E-mail (for print book orders and customer service enquires): ifsa.books@sensorsportal.com Visit our Home Page on http://www.sensorsportal.com Advances in Optics: Reviews, Vol. 1 is an open access book which means that all content is freely available without charge to the user or his/her institution. Users are allowed to read, download, copy, distribute, print, search, or link to the full texts of the articles, or use them for any other lawful purpose, without asking prior permission from the publisher or the authors. This is in accordance with the BOAI definition of open access. Neither the authors nor International Frequency Sensor Association Publishing accept any responsibility or liability for loss or damage occasioned to any person or property through using the material, instructions, methods or ideas contained herein, or acting or refraining from acting as a result of such use. ISBN: 978-84-697-9439-5, e-ISBN: 978-84-697-9440-1 BN-20180420-XX BIC: TTB Acknowledgments As Editor I would like to express my undying gratitude to all authors, editorial staff, reviewers and others who actively participated in this book. We want also to express our gratitude to all their families, friends and colleagues for their help and understanding. Contents Contents Contents............................................................................................................................ 7 Contributors................................................................................................................... 15 Preface ............................................................................................................................ 19 1. Health and Wellness Fiber Optic Sensors in IoT Application............................... 21 1.1. Introduction ...................................................................................................................... 21 1.2. Internet of Things ............................................................................................................. 22 1.2.1. Communication Models Used by IoT .................................................................................... 24 1.2.2. Main Existing Applications of IoT ........................................................................................ 26 1.2.3. Leading Companies .............................................................................................................. 27 1.3. Fiber Optic Sensors for Health and Wellness Application ............................................... 28 1.3.1. Working Principles and Applications ................................................................................... 29 1.3.1.1. Intensity-Modulated Fiber Optic Sensors ................................................................... 30 1.3.1.2. Interferometric Fiber Optic Sensors ........................................................................... 33 1.3.1.3. Wavelength-Modulated Sensors ................................................................................. 35 1.3.2. Multicore Fiber ..................................................................................................................... 37 1.4. IoT Systems for the Family Based on Fiber Optic Sensor ............................................... 38 1.5. Conclusion and Future Prospect ....................................................................................... 41 References ............................................................................................................................... 41 2. Multi-Point Temperature Sensor Consisting of AWG, SOA, and FBGs in Linear-Cavity Fiber Lasing System ................................................................... 49 2.1. Introduction ...................................................................................................................... 49 2.2. Linear-Cavity Fiber Sensor Consisting of SOA, AWG, and FBGs.................................. 50 2.3. Analysis of Multi-Channel Lasing ................................................................................... 52 2.3.1. Analysis for SOA Nonlinearity .............................................................................................. 52 2.3.2. Analysis of Multi-Wavelength Lasing ................................................................................... 54 2.4. Experimental Results ....................................................................................................... 56 2.4.1. SOA Nonlinearity .................................................................................................................. 56 2.4.2. ASE Spectrum and AWG Transmittance ............................................................................... 57 2.4.3. Multi-Wavelength Lasing ...................................................................................................... 58 2.4.4. Two-Wavelength Lasing with FBGs...................................................................................... 60 2.4.5. Simultaneous Temperature Sensing ...................................................................................... 62 2.4.6. Increase of the Temperature Sensing Range ......................................................................... 63 2.5. Conclusion ....................................................................................................................... 64 Acknowledgements ................................................................................................................. 64 References ............................................................................................................................... 65 Appendix ................................................................................................................................. 66 3. Review of Fabry-Pérot Fiber Sensors ...................................................................... 69 3.1. Introduction ...................................................................................................................... 69 3.2. Basic Theory .................................................................................................................... 70 3.3. Applications ..................................................................................................................... 71 7 Advances in Optics: Reviews. Book Series, Vol. 3 3.3.1. Strain Sensing ....................................................................................................................... 71 3.3.1.1. Prolate Spheroidal FP Cavity Strain Sensor ...............................................................71 3.3.1.2. Spherical FP Cavity Strain Sensor ..............................................................................75 3.3.1.3. Cylindrical FP Cavity Strain Sensor ...........................................................................76 3.3.1.4. Analysis of Temperature Sensitivity for the Fiber FP Strain Sensor ..........................83 3.3.1.5. Summary of this Section ............................................................................................. 84 3.3.2. Refractive Index Sensing ....................................................................................................... 84 3.3.2.1. Method of Measuring Fluid RI in FP Cavity ..............................................................84 3.3.2.2. Method of Measuring Fluid RI out of FP Cavity ........................................................88 3.3.2.3. Summary of this Section ............................................................................................. 91 3.3.3. Temperature Sensing............................................................................................................. 92 3.3.3.1. Theory ........................................................................................................................92 3.3.3.2. Applications of Temperature Sensor Based on FPI ....................................................93 3.3.3.3. Summary of this Section ............................................................................................. 98 3.4. Concluding Remarks and Perspectives ............................................................................. 98 References ............................................................................................................................... 99 4. Multi-Parameter Integrated Optical Sensor Based on Multimode Interference and Microring Resonator Structures ............................................. 103 4.1. Introduction..................................................................................................................... 103 4.2. Multimode Interference Structures ................................................................................. 104 4.3. Microring Resonator ....................................................................................................... 105 4.4. Two-Parameter Sensor Based on 4×4 MMI and Resonator Structure ............................ 106 4.5. Three-Parameter Sensor Based on 6×6 MMI and Resonator Structure .......................... 114 4.6. Four-Parameter Sensor Based on 8×8 MMI and Resonator Structure ............................ 118 4.7. Conclusions..................................................................................................................... 125 References ............................................................................................................................. 125 5. Coherent Gradient Sensor for Curvature Measurement in Extreme Environments .......................................................................................................... 129 5.1. Introduction..................................................................................................................... 129 5.2. The CGS System............................................................................................................. 130 5.3. Curvature Measurements in Cryogenic Medium ............................................................ 134 5.3.1. Governing Equations .......................................................................................................... 134 5.3.2. Error Analysis ..................................................................................................................... 136 5.3.3. Curvature Measurement in Cryogenic Vacuum Chamber .................................................. 138 5.4. Curvature Measurements in Multiple Media .................................................................. 139 5.4.1. Refraction Analysis ............................................................................................................. 139 5.4.2. Experiment Verification ...................................................................................................... 140 5.5. The Multiplication Method for Sparse Interferometric Fringes ...................................... 144 Acknowledgements................................................................................................................ 151 References ............................................................................................................................. 151 6. Photo-Emf Sensors and Talbot Effect: Measurement of Displacement and Vibration .......................................................................................................... 153 6.1. Introduction..................................................................................................................... 153 6.2. Non-Steady-State Photo-Electromotive Force Effect ..................................................... 155 6.3. Applications .................................................................................................................... 159 6.3.1. Measurements of Displacements ......................................................................................... 159 8 Contents 6.3.2. Determination of Low Frequency, Out-of-Plane Vibrations ............................................... 162 6.4. Conclusions .................................................................................................................... 165 References ............................................................................................................................. 166 7. Advances in Label-Free Sensing of Bacteria by Light Diffraction Phenomenon............................................................................................................ 169 7.1. Introduction .................................................................................................................... 169 7.2. The Biophysical Model of the Bacterial Colony ............................................................ 170 7.3. The Optical System for Analysis of Light Diffraction on Bacterial Colonies ............... 178 7.3.1. The Optical Wave Field Transformation in Proposed Optical System ............................... 179 7.3.2. The Configuration of the Experimental Optical System for Bacteria Identification ........... 183 7.4. Bacteria Identification Based on Fresnel Diffraction Patterns of Bacterial Colonies.... 184 7.4.1. The Bacteria Sample Preparation ....................................................................................... 184 7.4.2. The Experimental Fresnel Diffraction Patterns of Bacterial Colonies ............................... 185 7.4.3. The Analysis of the Diffraction Patterns ............................................................................. 188 7.5. The Use of the Fresnel Diffraction Patterns of Bacterial Colonies for Evaluation of the Efficiency of Antibacterial Factors ..................................................................... 190 7.6. The Perspectives for Exploiting of Light Diffraction on Bacterial Colonies Using Digital Holography ............................................................................................. 193 7.7. Conclusions .................................................................................................................... 194 Acknowledgements ............................................................................................................... 194 References ............................................................................................................................. 194 8. Integrated Terahertz Planar Waveguides for Molecular Sensing ..................... 197 8.1. Introduction .................................................................................................................... 197 8.2. THz Frequency Sensitive Detection ............................................................................... 198 8.2.1. Waveguide Configuration and Terahertz Spectral Characterization .................................. 198 8.2.2. Integration of a Superstrate and the Sensing Method ......................................................... 201 8.3. Phase Sensitive Detection .............................................................................................. 204 8.3.1. Waveguide Configuration and Terahertz Spectral Characterization .................................. 204 8.3.2. Integration of a Superstrate and the Sensing Method ......................................................... 206 8.4. Conclusions .................................................................................................................... 210 Acknowledgements ............................................................................................................... 210 References ............................................................................................................................. 210 9. Integrated-Optics Solutions for Biomedical Optical Imaging ............................. 213 9.1. Introduction .................................................................................................................... 213 9.2. Designs at 1300 nm ........................................................................................................ 214 9.2.1. Material System .................................................................................................................. 214 9.2.2. Working Principle of the Electro-Optic Switch ................................................................... 214 9.2.3. Akinetic Beam Scanner Layout and Its Working Principle ................................................. 216 9.2.4. Multiple-Reference TD-OCT Layout and Its Working Principle......................................... 217 9.2.5. Design Parameters of the Electro-Optic Switch ................................................................. 217 9.2.6. Two-Mode Interference Beam Splitter/Combiner Design ................................................... 219 9.3. High-Speed Spectrometer Designs ................................................................................. 220 9.3.1. Material System at 800 nm.................................................................................................. 221 9.3.2. Electro-Optic Switch Design at 800 nm .............................................................................. 222 9 Advances in Optics: Reviews. Book Series, Vol. 3 9.3.3. Ultrahigh-Resolution Spectrometer Layout and Its Working Principle .............................. 223 9.3.4. Broadband Spectrometer Layout and Its Working Principle .............................................. 225 9.4. Conclusions..................................................................................................................... 226 Acknowledgements................................................................................................................ 226 References ............................................................................................................................. 227 10. Video Based Heart Rate Estimation Using Facial Images from Video Sequences ............................................................................................ 229 10.1. Introduction................................................................................................................... 229 10.2. Introduction to Component Analysis ............................................................................ 232 10.2.1. Independent Component Analysis ..................................................................................... 232 10.2.2. Principal Component Analysis .......................................................................................... 232 10.3. Dynamic Heart Rate Estimation Using Component Analysis....................................... 233 10.3.1. Experimental Setup ........................................................................................................... 233 10.3.2. Experimental Results Using ICA Method .......................................................................... 234 10.3.3. Experimental Results Using PCA ...................................................................................... 236 10.4. Distance between the Subject and Video Camera......................................................... 238 10.4.1. Varying Distance with Fixed Video Duration ................................................................... 239 10.4.2. Fixed Distance with Varying Video Duration ................................................................... 241 10.5. Conclusion .................................................................................................................... 242 References ............................................................................................................................. 244 11. Implementing Differential Signalling in Free Space Optical Communication Link ............................................................................................. 247 11.1. Introduction to Free Space Optics................................................................................. 247 11.2. FSO Communications ................................................................................................... 249 11.2.1. Background ....................................................................................................................... 249 11.2.2. FSO Structure ................................................................................................................... 251 11.3. Turbulence .................................................................................................................... 256 11.4. Channel Model.............................................................................................................. 259 11.5. Differential Signalling .................................................................................................. 261 11.6. Differential Signalling Configuration ........................................................................... 263 11.7. Differential Signalling and Turbulence......................................................................... 264 11.7.1. Optimal Detection Threshold Level .................................................................................. 264 11.7.2. Correlation between Channels .......................................................................................... 265 11.7.3. Channel Modelling ............................................................................................................ 266 11.7.4. BER Expression................................................................................................................. 269 11.7.5. Numerical Analysis ........................................................................................................... 270 11.7.6. Atmospheric Turbulence Experiment ................................................................................ 278 11.8. Differential Signalling and Pointing Errors .................................................................. 283 11.8.1. Channel Modelling ............................................................................................................ 283 11.8.2. Pointing Errors Experiment .............................................................................................. 286 11.9. Differential Signalling and Manchester Code ............................................................... 290 11.9.1. System Configuration ........................................................................................................ 290 11.9.2. Manchester Code Experiment ........................................................................................... 291 11.10. Summary ..................................................................................................................... 294 References ............................................................................................................................. 295 10 Contents 12. Fabrications of Holographic Optical Elements in Polycarbonate for Holographic Weapon Sight Application ........................................................ 301 12.1. Introduction .................................................................................................................. 301 12.2. Material and Methods................................................................................................... 302 12.3. Experimental Arrangement .......................................................................................... 304 12.3.1. Fabrication Details of Reticle HOE in AgH ..................................................................... 304 12.3.2. Direct Fabrication of HOEs in Photoresist ...................................................................... 305 12.3.3. Transfer of HOE into PR and PC ..................................................................................... 306 12.4. Result and Discussion .................................................................................................. 307 12.5. Conclusion ................................................................................................................... 311 Acknowledgment .................................................................................................................. 311 References ............................................................................................................................. 312 13. Optical Methods for the Characterization of PV Solar Concentrators ........... 315 13.1. Introduction .................................................................................................................. 315 13.2. Theoretical Aspects of SC Irradiation and Definition of New Optical Quantities ....... 323 13.2.1. Direct Collimated Irradiation ........................................................................................... 323 13.2.2. Direct Lambertian Irradiation .......................................................................................... 329 13.2.3. Inverse Lambertian Irradiation......................................................................................... 332 13.2.4. Mixed Lambertian Irradiation .......................................................................................... 335 13.3. Equivalence between DCM and ILM ........................................................................... 338 13.4. Real Prototypes of Nonimaging Solar Concentrators................................................... 342 13.5. Practical Application of the SC Characterization Methods .......................................... 346 13.5.1. Application of the DCM Method ....................................................................................... 346 13.5.2. Application of the ILM Method (Parretta-Method)........................................................... 353 13.5.3. The Application of DLCM Method .................................................................................... 364 13.5.3.1. Optical Efficiency Measurements........................................................................... 366 13.5.3.2. Beam Exit Angle Measurements ............................................................................ 366 13.5.4. The Application of the PH-Method (Parretta-Herrero Method) ....................................... 367 13.6. Experimental Results.................................................................................................... 369 13.6.1 The Truncated and Squared CPC (TS-CPC) ..................................................................... 369 13.6.1.1. Local Optical Efficiency by the Laser Method (DLCM) ........................................ 369 13.6.1.2. Beam Exit Angle Measurements ............................................................................ 373 13.6.1.3. Optical Efficiency by DCM and ILM ..................................................................... 376 13.6.1.4. Local Optical Efficiency by ILLM ......................................................................... 378 13.6.2. The (Virtual) Half-Truncated CPC (HT-CPC) ................................................................. 381 13.6.2.1. Local Optical Efficiency by ILLM ......................................................................... 381 13.6.3. The Truncated CPC (T-CPC) ........................................................................................... 383 13.6.3.1. Optical Efficiency by ILM ..................................................................................... 383 13.6.4. The Rondine Concentrators .............................................................................................. 385 13.6.4.1. Optical Efficiency by DCM and ILM ..................................................................... 385 13.6.4.2. Optical Efficiency by Parretta-Method and Parretta-Herrero Method .................... 388 13.6.5. The PhoCUS Concentrator ............................................................................................... 391 13.6.5.1. Optical Simulations ................................................................................................ 391 13.6.5.2. Experimental Measurements .................................................................................. 393 13.7. Conclusions .................................................................................................................. 394 Acknowledgements ............................................................................................................... 395 References ............................................................................................................................. 396 Appendix 13.A ...................................................................................................................... 400 Appendix 13.B ...................................................................................................................... 402 11 Advances in Optics: Reviews. Book Series, Vol. 3 Appendix 13.C ....................................................................................................................... 408 Appendix 13.D ....................................................................................................................... 411 14. Determination of the Heights of the Smoke-Plume Layering in the Vicinity of Wildfires and Prescribed Burns ....................................................................... 417 14.1. Introduction................................................................................................................... 417 14.2. Determination of the Heights of the Smoke-Plume Layering in the Vicinity of Wildfires: Theory and Data Processing Methodology............................................... 419 14.3. Some Results of Lidar Profiling of the Smoke-Polluted Atmosphere in the Vicinity of Spotted Wildfires ............................................................................... 423 14.4. Determination of the Heights of the Smoke-Plume Layering in the Vicinity of Prescribed Burns ........................................................................................................ 426 14.5. Summary ....................................................................................................................... 431 References ............................................................................................................................. 432 15. Precision Glass Molding........................................................................................ 435 15.1. Introduction................................................................................................................... 435 15.2. An Introduction to PGM ............................................................................................... 437 15.3. Selection of Glass and Preparation of Its Preform ........................................................ 438 15.4. Precision Mold Fabrication ........................................................................................... 439 15.4.1. Mold Materials in PGM .................................................................................................... 439 15.4.2. Mold Fabrications............................................................................................................. 441 15.4.2.1. Macro Mold Fabrication ......................................................................................... 441 15.4.2.2. Micro Mold Fabrication .......................................................................................... 443 15.5. PGM Process ................................................................................................................ 444 15.5.1. Stages in a PGM Process .................................................................................................. 444 15.5.2. Finite Element Analysis..................................................................................................... 445 15.5.2.1. Constitutive Modeling of Optical Glass.................................................................. 445 15.5.2.2. Mechanisms of Profile Distortion ........................................................................... 446 15.5.2.3. Residual Stresses .................................................................................................... 448 15.6. Quality Inspection Techniques ..................................................................................... 450 15.6.1. Surface Characterization .................................................................................................. 450 15.6.2. Residual Stress Characterization ...................................................................................... 451 15.6.3. Micro-Optics Characterization and Standardization ........................................................ 451 15.7. Optimization of PGM Process ...................................................................................... 453 15.7.1. Optimization Strategy........................................................................................................ 453 15.7.2. Mold Shape Optimization .................................................................................................. 454 15.7.3. Residual Stress Optimization ............................................................................................ 456 15.7.4. Multi-Objective Optimization ............................................................................................ 456 15.8. Summary ....................................................................................................................... 458 Acknowledgements................................................................................................................ 458 References ............................................................................................................................. 458 16. Deterministic Loose Abrasive Wear in Conventional Grinding-Polishing Machines ................................................................................................................. 467 16.1. Introduction................................................................................................................... 467 16.2. Abrasive Wear Theory .................................................................................................. 468 16.3. Conventional Grinding-Polishing Machines ................................................................. 469 12 Contents 16.4. Relative Velocities between an Arbitrary Pair of Points .............................................. 471 16.5. Upper Disk Relative Velocity ...................................................................................... 472 16.5.1. First Relative Velocity Contribution (V0) Approximate Calculation ................................. 472 16.5.2. Second Relative Velocity Contribution (V1) ..................................................................... 473 16.5.3. Third Relative Velocity Contribution (V2) ......................................................................... 474 16.5.4. Vector Addition of Three Relative Velocity Contributions ................................................ 476 16.6. Lower Disk Relative Velocity ...................................................................................... 476 16.6.1. Approximate Calculation of the First Relative Velocity Component (V0) ......................... 476 16.6.2. Calculation of the Relative Velocity Second Component (V1) ........................................... 477 16.6.3. General Expression for the Third Relative Velocity Component (V2) ............................... 478 16.6.4. Vector Addition of Three Relative Velocity Contributions ................................................ 480 16.7. Boundary Conditions in Abrasive Wear Process ......................................................... 481 16.8. Pressure Distribution within Disks Contact Area ......................................................... 481 16.9. Arm Stroke Adjustments (Controlling Curvature Radius and Figure) ......................... 482 16.10. Simulation and Real Optical Manufacturing .............................................................. 482 16.11. Concluding Remarks .................................................................................................. 484 Acknowledgment .................................................................................................................. 484 References ............................................................................................................................. 485 17. Quantitative Phase Microscopy and Tomography with Spatially Incoherent Light ..................................................................................................... 487 17.1. Introduction .................................................................................................................. 487 17.2. Concepts of Coherence................................................................................................. 489 17.2.1. Temporal Coherence......................................................................................................... 489 17.2.2. Spatial Coherence ............................................................................................................. 490 17.2.2.1. Transverse Spatial Coherence................................................................................. 491 17.2.2.2. Longitudinal Spatial Coherence ............................................................................. 492 17.3. Synthesis of Low Spatial and High Temporal Coherent light Source .......................... 493 17.3.1. Experimental Details ........................................................................................................ 495 17.4. Phase Retrieval Algorithm ........................................................................................... 496 17.4.1. Five Step Algorithm .......................................................................................................... 496 17.4.2. Fourier Transform Algorithm ........................................................................................... 497 17.5. Characterization of System Parameters ........................................................................ 498 17.5.1. Temporal and Spatial Phase Noise ................................................................................... 498 17.5.2. Transverse Resolution....................................................................................................... 499 17.5.3. Axial Resolution ................................................................................................................ 499 17.6. Spatial Phase Noise Comparison in Case of Direct Laser and Synthesized Light Source .................................................................................................................. 501 17.6.1. Standard Flat Mirror ........................................................................................................ 501 17.6.2. Human Red Blood Cells .................................................................................................... 502 17.7. Quantitative Phase Imaging of Industrial and Biological Cells Using Pseudo Thermal Light Source.................................................................................................... 503 17.7.1. QPI of Standard Waveguide ............................................................................................. 503 17.7.2. QPI of Human RBCs ......................................................................................................... 503 17.7.3. QPI of Onion Cells ........................................................................................................... 505 17.8. Profilometry and Optical Coherence Tomography ...................................................... 506 17.8.1. Profilometry of Standard Gauge Block and Indian Five Rupee Coin ............................... 506 17.8.2. OCT of Multilayered Onion Sample ................................................................................. 507 13 Advances in Optics: Reviews. Book Series, Vol. 3 17.9. Conclusions................................................................................................................... 508 Acknowledgements................................................................................................................ 509 References ............................................................................................................................. 509 Index ............................................................................................................................. 513 14 Contributors Contributors Mojtaba Mansour Abadi School of Engineering, University of Glasgow, Glasgow, Scotland, UK Azeem Ahmad Department of Physics, Indian Institute of Technology Delhi, Hauz Khas, New Delhi 110016, India, E-mail: ahmadazeem870@gmail.com Luis C. Alvarez-Nuñez Universidad Nacional Autónoma de México, Instituto de Astronomía, Ciudad Universitaria, Mexico City, Mexico, E-mail: lalvarez@astro.unam.mx Anatoly Babchenko Lev Academic Center, Faculty of Engineering, Department of Applied Physics/Electro-Optics Engineering, Jerusalem, Israel Igor Buzalewicz Bio-Optics Group, Department of Biomedical Engineering, Faculty of Fundamental Problems of Technology, Wroclaw University of Science and Technology, 27 Wybrzeze S. Wyspianskiego St., 50-3708, Wroclaw, Poland Oscar Chapa Universidad Nacional Autónoma de México, Instituto de Astronomía, Ciudad Universitaria, Mexico City, Mexico P.-J. Chen Dept. of Optical Science, Tokushima University, Japan Dept. of Electronic Engineering, National Taiwan University of Science and Technology, Taiwan Ilan Gadasi Lev Academic Center, Faculty of Engineering, Department of Applied Physics/Electro-Optics Engineering, Jerusalem, Israel Ying Guo Harbin Institute of Technology, China W. M. Hao Missoula Fire Sciences Laboratory, Missoula, Montana, 59808, USA B. Imran Akca Institute for Lasers, Life and Biophotonics Amsterdam, Department of Physics and Astronomy, VU University Amsterdam, De Boelelaan 1081, 1081 HV Amsterdam, The Netherlands Carolina Keiman Universidad Nacional Autónoma de México, Instituto de Astronomía, Ciudad Universitaria, Mexico City, Mexico 15 Advances in Optics: Reviews. Book Series, Vol. 3 H. Kishikawa Department of Optical Science, Tokushima University, Japan N. Korneev National Institute for Astrophysics, Optics and Electronics, A.P. 51, Puebla 72000, Mexico V. Kovalev Missoula Fire Sciences Laboratory, Missoula, Montana, 59808, USA Katarzyna Kowal Bio-Optics Group, Department of Biomedical Engineering, Faculty of Fundamental Problems of Technology, Wroclaw University of Science and Technology, 27 Wybrzeze S. Wyspianskiego St., 50-3708, Wroclaw, Poland Trung-Thanh Le International School (VNU-IS), Vietnam National University (VNU), 144 Xuan Thuy, Cau Giay, Hanoi, Vietnam S.-K. Liaw Dept. of Electronic Engineering, National Taiwan University of Science and Technology, Taiwan Mariusz Linard Bio-Optics Group, Department of Biomedical Engineering, Faculty of Fundamental Problems of Technology, Wroclaw University of Science and Technology, 27 Wybrzeze S. Wyspianskiego St., 50-3708, Wroclaw, Poland S.-S. Ling Department of Electrical Engineering, Faculty of Engineering, University of Malaya, Kuala Lumpur, 50603, Malaysia, E-mail: s.shiling@hotmail.com Cong Liu Key Laboratory of Mechanics on Disaster and Environment in Western China attached to the Ministry of Education of China, Lanzhou University, Lanzhou, Gansu 730000, PR China Department of Mechanics and Engineering Sciences, College of Civil Engineering and Mechanics, Lanzhou University, Lanzhou, Gansu 730000, PR China Weidong Liu Laboratory for Precision and Nano Processing Technologies, School of Mechanical and Manufacturing Engineering, University of New South Wales, NSW 2052 Australia Ja-Yu Lu Department of Photonics, National Cheng Kung University, Tainan 70101, Taiwan E-mail: jayu@mail.ncku.edu.tw S. Mansurova National Institute for Astrophysics, Optics and Electronics, A.P. 51, Puebla 72000, Mexico 16 Contributors Dalip Singh Mehta Department of Physics, Indian Institute of Technology Delhi, Hauz Khas, New Delhi 110016, India, E-mail: mehtads@physics.iitd.ac.in M. Okada Dept. of Optical Science, Tokushima University, Japan Antonio Parretta Physics and Earth Science Department, University of Ferrara, Italy A. Petkov Missoula Fire Sciences Laboratory, Missoula, Montana, 59808, USA Halina Podbielska Bio-Optics Group, Department of Biomedical Engineering, Faculty of Fundamental Problems of Technology, Wroclaw University of Science and Technology, 27 Wybrzeze S. Wyspianskiego St., 50-3708, Wroclaw, Poland Raveendran P. Department of Electrical Engineering, Faculty of Engineering, University of Malaya, Kuala Lumpur, 50603, Malaysia, E-mail: ravee@um.edu.my P. Rodriguez National Institute for Astrophysics, Optics and Electronics, A.P. 51, Puebla 72000, Mexico D. Sanchez de la Llave National Institute for Astrophysics, Optics and Electronics, A.P. 51, Puebla 72000, Mexico Chandar Shekar Department of Physics, KASC, G.N Mills post, Coimbatore -641029, Tamilnadu, India Agnieszka Suchwałko QUANTUP, Wrocław, Poland K. Takahashi Dept. of Optical Science, Tokushima University, Japan S. Urbanski Missoula Fire Sciences Laboratory, Missoula, Montana, 59808, USA Vadivelan V. R&D Department, Ignetta holographic (P) Ltd, Madukkari, Coimbatore – 641105, Tamilnadu, India Department of Physics, KASC, G.N Mills post, Coimbatore -641029, Tamilnadu, India C. Wold Missoula Fire Sciences Laboratory, Missoula, Montana, 59808, USA 17 Advances in Optics: Reviews. Book Series, Vol. 3 Borwen You Division of Applied Physics, Faculty of Pure and Applied Sciences, University of Tsukuba, Tennodai 1-1-1, Tsukuba, Japan Y.-L. Yu Dept. of Electronic Engineering, National Taiwan University of Science and Technology, Taiwan Y. P. Yu Department of Computer Science and Mathematics, Faculty of Computing and Information Technology, Tunku Abdul Rahman University College, 53300, Malaysia, E-mail: yuyp@acd.tarc.edu.my Ping Yuan Harbin Institute of Technology, China Ariel Zev Lev Academic Center, Faculty of Engineering, Department of Applied Physics/Electro-Optics Engineering, Jerusalem, Israel Liangchi Zhang Laboratory for Precision and Nano Processing Technologies, School of Mechanical and Manufacturing Engineering, University of New South Wales, NSW 2052 Australia Xingyi Zhang Key Laboratory of Mechanics on Disaster and Environment in Western China attached to the Ministry of Education of China, Lanzhou University, Lanzhou, Gansu 730000, PR China Department of Mechanics and Engineering Sciences, College of Civil Engineering and Mechanics, Lanzhou University, Lanzhou, Gansu 730000, PR China E-mail: zhangxingyi@lzu.edu.cn, Tel: +86-931-8914560. Yundong Zhang Harbin Institute of Technology, China Li Zhao Harbin Institute of Technology, China Youhe Zhou Key Laboratory of Mechanics on Disaster and Environment in Western China attached to the Ministry of Education of China, Lanzhou University, Lanzhou, Gansu 730000, PR China Department of Mechanics and Engineering Sciences, College of Civil Engineering and Mechanics, Lanzhou University, Lanzhou, Gansu 730000, PR China Fuxing Zhu Harbin Institute of Technology, China 18 Preface Preface It is my great pleasure to introduce the third volume of new Book Series ‘Advances in Optics: Reviews’ started by the IFSA Publishing in 2018. Three volumes were published in this year. The ‘Advances in Optics: Reviews’ Book Series is published as an Open Access Books in order to significantly increase the reach and impact of these volumes, which also published in two formats: electronic (pdf) with full-color illustrations and print (paperback). The third of three volumes of this Book Series has organized by topics of high interest. In order to offer a fast and easy reading of each topic, every chapter in this book is independent and self-contained. All chapters have the same structure: first an introduction to specific topic under study; second particular field description including sensing or/and measuring applications. Each of chapter is ending by complete list of carefully selected references with books, journals, conference proceedings and web sites. The Vol.3 is devoted to various topics of applied optics and contains 17 chapters written by 49 experts in the field from 14 countries: Australia, China, India, Israel, Italy, Japan, Malaysia, Mexico, The Netherlands, Poland, Taiwan, UK, USA and Vietnam. ‘Advances in Optics: Reviews’ Book Series is a comprehensive study of the field of optics, which provides readers with the most up-to-date coverage of optics, photonics and lasers with a good balance of practical and theoretical aspects. Directed towards both physicists and engineers this Book Series is also suitable for audiences focusing on applications of optics. A clear comprehensive presentation makes these books work well as both a teaching resources and a reference books. The book is intended for researchers and scientists in physics and optics, in academia and industry, as well as postgraduate students. I shall gratefully receive any advices, comments, suggestions and notes from readers to make the next volumes of ‘Advances in Optics: Reviews’ Book Series very interesting and useful. Dr. Sergey Y. Yurish Editor IFSA Publishing Barcelona, Spain 19 Chapter 1. Health and Wellness Fiber Optic Sensors in IoT Application Chapter 1 Health and Wellness Fiber Optic Sensors in IoT Application Anatoly Babchenko, Ariel Zev and Ilan Gadasi1 1.1. Introduction Things around us have been connected for decades. Car door openers, TV remote controls and other devices have been part of our life for generations. Industrial and medical applications of these technologies are also nothing new. In fact, even Internet connected to the physical world via smart sensors (“Internet of Things”) is not a recent invention; it was proposed around thirty years ago. However, recent developments in both sensors, especially fiber optic sensors, and networks are enabling a much greater range of connected objects and devices. The potential applications of these new systems are virtually limitless, and they have the ability to greatly improve quality of life. In this chapter, we explore a wide range of topics related to Internet of Things (IoT) and fiber optic sensors’ applications for health and wellness monitoring. Health and wellness represent one of the most attractive areas for IoT [1] and the current review is related directly to this field. The real-world example have been provided to give the reader practical insights into the successful development of IoT system for child’s wellness monitoring. The chapter is organized as follows (Fig. 1.1). Section 1.1 introduces Internet of Things systems - a reality that surrounds us and intersects with many aspects of our lives. Section 1.2 presents Fiber Optic Sensors for health and wellness monitoring that will play a central role in providing the success of IoT in these applications. Section 1.3 shows IoT with fiber optic sensing technology designed for child’s wellness monitoring. The conclusion is drawn in Section 1.4. Anatoly Babchenko Lev Academic Center, Faculty of Engineering, Department of Applied Physics/Electro-Optics Engineering Jerusalem, Israel 21 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 1.1. Overview structure. 1.2. Internet of Things A reality of Internet of Things has changed the world in which we live. The term “Internet of Things” was first used by British technology pioneer Kevin Ashton to describe a system where the Internet is connected to the physical world via sensors [2]. Ashton illustrated the power of connecting Radio-Frequency Identification tags used in corporate supply chains to the Internet in order to count and track goods without the need for human intervention. The concept of combining computers and networks to control devices has been around for decades. Early Machine to Machine (M2M) applications have typically sent all remote data from the devices to the data centre for processing [3-5]. However, as more data is generated at the edge in real-time, a greater need for real time decision making also at the edge will be required. Many of these early applications, however, were not based on Internet Protocol (IP) and Internet standards. One of the first Internet Protocol “devices” has been developed by Simon Hackett and John Romkey, becoming the hit of the 1990 Interop. They connected a Sunbeam Deluxe Automatic Radiant Control Toaster to the Internet with TCP/IP networking, and controlled with a Simple Networking Management Protocol Management Information Base (SNMP MIB). Other “thing” was the internet-connected soda vending machine that 22 Chapter 1. Health and Wellness Fiber Optic Sensors in IoT Application allows customers to check the availability status of soda remotely using a finger interface. Originally developed circa 1982 by a group of students at Carnegie Mellon University, the machine became known as one of the very first Internet appliances and inspired a number of modified versions allowing people to queue their purchases on the machine remotely via Internet, such as the Trojan Room Coffee Pot. A design student at Brunel University, UK, has developed a toaster that takes meteorological information from the internet and then browns his bread with an image of what weather to expect on the way to work. The experiment was adopted later by Electrolux to produce a more detailed weather report. All these freakish beginnings helped create the ground for today’s IoT that is rapidly becoming a part of every aspect of our lives. The IoT was added to the 2011 annual Gartner Hype Cycle (Fig. 1.2) that tracks technology life-cycles from "technology trigger" to "plateau of productivity" and has hit the Hype Cycle's "Peak of Inflated Expectations" in 2014 [6]. Fig. 1.2. Gartner Hype Cycle for Emerging Technologies, 2014. According to the McKinsey Global Institute [7] IoT has a total potential economic impact of $3.9 trillion to $11.1 trillion a year by 2025 and Morgan Stanley projects 75 billion networked devices/objects by 2020 [8]. Business-to-business applications of IoT systems certainly capture more value than consumer uses, although consumer application, such as health and wellness, attract more and more attention [9] and can create significant value in near future. The healthcare is 23 Advances in Optics: Reviews. Book Series, Vol. 3 shifting from reactive approach towards health conditions to a more proactive approach in terms of early detection of conditions, prevention and well-being management. Physical condition monitoring, their automated processing and automatic management of individual well-being based on IoT technologies [10] will play a key role in providing better healthcare services and improving the quality of life. Advances in IoT and sensors technologies have made possible the connection of more and more devices to the Internet. This is leading to a new wave of applications that have the potential to dramatically improve the way people work, learn and live. Sensors play a key role in connecting the physical world with the digital world and they are in the core of IoT [11]. For example, home-based environmental monitors allow people to track ambient air quality. They can use this data to either modify their environment or alter their behaviour in order to maintain their health and wellness [12]. Some, like the U.K.-based technology and development organization The Technology Partnership, argue that IoT should really be called the Internet of Sensors (IoS) [13]. In the IoT, things at the edge can create significantly large amounts of data. Transmitting all that data to the cloud and transmitting response data back demands wide bandwidth and requires a considerable amount of time. Rather than having to be transmitted to the cloud the data from sensors or internet can be processed locally in smart devices with a fog (fog is a cloud close to the ground) computing data processing [14-15]. Companies that adopt fog computing gain deeper and faster insights, leading to increased business agility, higher service levels, and improved safety. 1.2.1. Communication Models Used by IoT The Internet of Things is connecting various sensors and devices, based on them. The basic principle of every IoT is how devices/sensors connect and communicate. Internet Architecture Board – a group within the Internet Society that oversees the technical evolution of the Internet – defined four common communication models used by IoT [16]: Device-to-Device, Device-to-Cloud, Device-to-Gateway, and Back-End Data-Sharing. Device-to-device (Fig. 1.3) communication represents two or more devices/sensors that directly connect and communicate between one another. This model is commonly used in home IoT systems to transfer small data packets of information between sensors at a relatively low data rate. Device-to-device is used in wearable IoT devices like a heart rate or breathing efforts monitor connected to a smartwatch where data doesn’t necessarily have to be shared with multiple people. Fig. 1.3. Device-to-device model. 24 Chapter 1. Health and Wellness Fiber Optic Sensors in IoT Application Device-to-Cloud (Fig. 1.4) communication involves an IoT devices/sensors connecting directly to an Internet cloud service to exchange data and control message traffic. It often uses traditional wired Ethernet or Wi-Fi connections, but can also use mobile technology. Cloud connectivity lets the big amount of users to obtain remote access to various sensors and applications. It also gives an opportunity for big data processing and potentially supports pushing software updates to the devices. This model can be used for various health and wellness IoT systems. Fig. 1.4. Device-to-Cloud model. Device-to-Gateway (Fig. 1.5) is a model where IoT devices connect to an intermediary device to access a cloud service. The example can be a fitness device that connects to the cloud through a smartphone app like in Nike personal trainer system. Fig. 1.5. Device-to-Gateway model. Back-End Data-Sharing (Fig. 1.6) model refers to a communication architecture that enables users to export and analyze smart object data from a cloud service in combination with data from other sources. An example of this model is a fitness tracking application MapMyFitness [17]. It compiles fitness data from various devices like Fitbit, Adidas miCoach and Wahoo Bike Cadence Sensor. This means an exercise can be analyzed from the viewpoint of various sensors. 25 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 1.6. Back-End Data-Sharing model. 1.2.2. Main Existing Applications of IoT At present, a wide range of industry sectors are considering the potential for incorporating IoT technology into their products, services, and operations [18-19]. Here are the main applications of IoT devices and sensor systems: Health and wellness [20-23]. Disease management; wearables, devices attached to (including smart bed, chair, etc.) or inside the human body - to monitor and maintain human health and wellness; sport and fitness. Sensors - pressure, acceleration, shape/form, temperature, chemical, bio sensors. Wearable IoT systems have many forms, such as glasses, watches, helmet, jewelry, collar, wristband, belts, rings, gloves, body-wear clothing, shoes and socks. Smart home [24-26]. Home and building automation with controllers and monitoring devices for security, garage, fire alarm, water, light, home electronics (TV, airconditioner, refrigerator, etc.) and network. Sensors - temperature, humidity, chemical, imaging, motion, power. Transportation [27-28]. Devices inside staying or moving vehicles including cars, trucks, trains, ships and aircraft. Sensors - gyroscope, velocity, temperature, humidity, pressure, chemical. City [29]. Systems in and outside urban environments, streets, buildings, railroad tracks, autonomous vehicles (outside urban locations), flight navigation, real-time routing, connected navigation, shipment tracking. Sensors - gyroscope, environmental monitoring, chemical, light, noise, traffic control, structural health monitoring. Others: security and public safety [30]. Factories, farms, hospitals and clinics [31], stores, banks, restaurants, arenas; consumers [32]. 26 Chapter 1. Health and Wellness Fiber Optic Sensors in IoT Application 1.2.3. Leading Companies Here are the leading companies that have ventured into the IoT world. Medtronic’s Continuous Glucose Monitoring [33] is a wearable device that displays a constant reading of a diabetic’s blood glucose level. A tiny electrode is inserted under the skin, which then transmits the glucose reading via wireless radio frequency to a display device. CellNovo [34] who have developed the world’s first mobile diabetes management system. The device has a built-in insulin pump and monitoring system that records glucose level and activity monitor to track and record exercise. It can detect possible hypos and hypers before they happen, warning the user before it’s too late. GlowCap device from Vitality [35]. This is an IoT digitized medicine bottle that can be programmed so that when the user needs to take their tablet, it flashes and sounds notifications as a reminder. It also records when the bottle is opened and sends information back to the clinic to inform them. It even features a button at the bottom of the bottle, that when pressed, automatically orders the next prescription/order from the pharmacy. AliveCor [36] heart monitor. Traces standard ECG heart rhythms with a mobile device monitor. Qualcomm Life’s [37] wireless health system stores medical data in a cloud-based platform from various medical devices. Facilitates continuum of care and improves health care delivery. Fitbit Flex [38] tracks activity, calories burned, diet, and sleep. Zephyr’s BioHarmess [39] captures and transmits comprehensive physiological data. BodyMedia’s [40] on-body monitoring system collects information on temperature, moisture, and movement. Wireless blood pressure wrist monitor by iHealth [41] tracks changes in pulse rate, systolic and diastolic blood pressure. Ring [42] is a connected doorbell and home security solution used for home automation and for helping disabled or the elderly. It alerts users to motion as soon as it’s detected, so they can remotely monitor their door. A doorbell with a built-in camera, motion sensor and two-way communication allows a house owner to check on who is at the door or even nearby before ever getting near opening it. Philips’ Hue Light Bulbs and Bridge [38] provide a bridge and connected bulbs that allow the user to control their home lighting from the palm of their hand. 27 Advances in Optics: Reviews. Book Series, Vol. 3 iRobot’s Roomba [43] is a smart vacuum cleaner equipped with iAdapt technology, a system of software and sensors that enables Roomba to find its way around a home of any shape or size. Ralph Lauren’s Polo Tech Shirt [44] is a shirt with conductive threads woven into it and relays information like heart rate and breathing data to a Bluetooth-connected iPhone or iPad. Nest [38], is a smart thermostat that’s connected to the internet. The Nest learns family’s routines and will automatically adjust the temperature. Samsung SmartThings [45] system controls lights, locks, plugs, thermostats, cameras, and speakers from a central hub that can be accessed from a smartphone, as well as a wide range of sensors that can be used with the system to create a security solution that’s integrated with all of the other electronics at home. Cisco’s Connected Factory [46] is a remote monitoring and access technology to the equipment used in manufacturing. DHL’s IoT Tracking and Monitoring [47] released a report detailing some potential uses of IoT technology that includes vehicle monitoring and maintenance, real-time tracking of packages, environmental sensors in shipping containers, informationgathering on employees and tools, and a number of safety-enhancing features for vehicles and people. Ericsson Maritime ICT [48] which provides infrastructure for ships, ports and terminals. Via its 'Maritime ICT Cloud' system, the company uses sensors on its ships to monitor vessel location, speed and the temperature for heat sensitive cargo, all in real-time. 1.3. Fiber Optic Sensors for Health and Wellness Application Fiber Optic Sensors represent a technology that can be successfully applied to a multitude of sensing applications. Pressure, stress, strain, vibration, temperature, displacement, concentration, density, temperature, or chemical composition are just some of the phenomena that can be measured [49]. A huge number of fiber optic sensing devices for biomedical applications have been designed in the last years. The most reported fiber optic sensors are for biological, biomechanical, and physiological parameters. For biological purposes, several optical devices have been developed [50-51] to target cells and proteins, identify DNA markers or to measure humidity/moisture. In the field of biomechanics, sensors based on optical fibers are presented to measure displacement, force, acceleration, angle and pressure [52]. Human physiological parameters such as respiratory rate [53], temperature [54], pH [55] or heart beat [56] can be monitored with fiber optic sensors. From a biomedical engineering point of view, optical fibers and sensors based on them, possess many advantages that could be wisely used for health and wellness monitoring. Some of them are briefly explained below: 28 Chapter 1. Health and Wellness Fiber Optic Sensors in IoT Application 1. Biomedical compatibility. The material of optical fiber sensing element does not present side effects when attached to person, embedded in his/her organs or placed in biological substance [57]. 2. Small Size and Light Weight. The size and weight of the optical fiber are very small (approximately 100 μm). The fiber can be embedded, into the fabric for example, and to measure in real time various body’s parameters [58]. 3. Low Cost. Optical fibers, both polymer and silica, are cheap materials for final application. 4. Flexibility. An optical fiber can reach almost any place in the human body and measure it during surgery or other medical procedures. For instance, a guiding catheter introduced into the coronary artery, and blood pressure measured by the fiber optic sensor [59]. 5. Reliability. The issue of reliability of optical fiber and sensors becomes increasingly important, as they are more and more frequently used in applications where a failure of the sensor might have dramatic consequences on patient’s health or life. An example of such an application is integrated microfluidic chip for glucose detection [60]. 1.3.1. Working Principles and Applications It is not possible to describe all the working principles and applications of fiber optic sensors in one chapter. However fiber optic sensors for health and wellness can be easily classified into five major categories according to the location of the sensor application. 1. Sensors attached or embedded into expandable belts, patches, helmets or special garments (gloves, shoes, smart textiles) as well as affixed directly to the body, measuring pressure, angle, strain, vibration, temperature, displacement or density. 2. Sensors placed in a mouth, an ear, a nostril or oxygen mask, measuring temperature, humidity, pressure or force of inhaled/exhaled air (flow). 3. Sensors attached to chairs, placed on/in/under a bed mattress, or embedded into cushion, measuring the strain, pressure, or force caused by respiration, heart or other physiological functions. 4. In-vivo sensors placed or implanted inside a human body (heart, blood vessels etc.), measuring pressure, temperature and density. 5. Sensors used in laboratory for in-vitro tests, measuring chemical composition. The growing interest in the smart fiber optic sensors for health and well-being applications is driven by the success of IoT, as the next evolution of the Internet. These sensors allow the measurement of various biophysical parameters employing a large number of configurations and working principles [61]. Fiber optics sensors can be classified under two major categories: the sensing location and the operating principal. Depending on 29 Advances in Optics: Reviews. Book Series, Vol. 3 location of sensor, a fiber sensor can be classified to intrinsic or extrinsic. In intrinsic sensors, the internal property of the optical fiber itself converts the environmental changes into a modulation of optical signal. In contrast to intrinsic, in extrinsic sensor the modulation takes place outside the fiber. In this case the fiber merely acts as a conduit to transport light signal to and from the sensor head. Based on the operating principal and demodulation technique, fiber optic sensors can be classified into few major categories: intensity, phase, wavelength, polarisation and distributed sensors (Brillouin, Raman and Rayleigh). 1.3.1.1. Intensity-Modulated Fiber Optic Sensors An intensity-modulated sensor relies on variations of the radiant power transmitted through an optical fiber with respect to the measurand. A wide range of sensors based on intensity modulation have previously been studied and tested on patients to monitor their physiological parameters. Fig 1.7 (a, b), 8 show how intensity-based optical fiber extrinsic and intrinsic sensors monitor human seated spinal posture [62]. In one of the optical sensor for back-pain patients [63] the bending angle of spine is measured as the angle between the emitting and receiving fibers that changed. Two separate configurations of the sensor are considered: fiber longitudinal displacement loss (Fig. 1.7 (a)) and fiber tilt angle loss (Fig. 1.7 (b)). The light intensity decreases with the increase in the gap between emitting and receiving fibers as well as increases in bending angle between them. Fig. 1.7. (a) Spine-bending sensor with a fiber gap; (b) Fiber tilt angle. Another sensor (Fig. 1.8) has been designed using side-polished plastic optical fiber [64]. This work describes the evaluation of a wearable bending optical fiber sensor for monitoring seated spinal posture. It is also possible to measure the direction of the bending throughout the measurement. If the polished area is at the external side (convex) of the bending fiber, more light will escape the fiber, thus giving a higher attenuation and vice versa. 30 Chapter 1. Health and Wellness Fiber Optic Sensors in IoT Application Fig. 1.8. Side polished optical fiber for bending sensor. Other solutions utilising intensity-modulated sensors have been developed [65]. In this model, the area of polymer optical fiber is replaced by a uniform layer with a complex refractive index. The changes in the absorption characteristics of the polymer depend on the environmental properties in the sensing area and the light intensity changes as a result of light attenuation in this imperfected fiber. Micro and macrobending fiber-optic sensors In sensors measuring transmission loss caused by micro changes of the shape of an optical fiber or in fiber bends with a radius of curvature well above the fiber diameter, i.e., in socalled micro- and macro bend sensors, respectively, the intensity of the light reaching the receiver is measured. Various human body movements (caused by respiration and heart function, among other things) are a complicating factor since these movements produce micro- and macrobends of fibers with a variable radius of curvature. Thus, along the axis of a bent fiber, the layout of the mode fields continuously changes as the energy radiates; this is seen in the form of light intensity changes at the receiver. The measuring system for micro- and macrobend sensors is distinguished by a simple design for the sensors alone as well as for the associated transceiver modules (light source and photodetector). The sensors can be embedded in a cushion, mattress, armchair, on which the patient sits or lies during monitoring, or, special garments (textiles) affixed directly to the body, which are worn by the monitored person. A micro-bending fiber based sensor was reported in [66]. Authors propose a smart cushion for heart rate monitoring. The cushion consists of an integrated micro-bending fiber sensor and a new heart rate extraction algorithm. The principle of using micro-bending fiber sensor for heart rate measurement is based on Ballistocardiogram, which measures the body vibration caused by the heartbeat. The sensor contains a section of multimode fiber clamped between a pair of micro-benders, as shown in Fig. 1.9. As the displacement between two micro-benders changes, the light intensity of the clamped multimode fiber changes with subject’s body vibrations caused by respiration/heart beating, i.e. the light intensity in the micro-bending fiber is modulated by the body vibrations. This modulated signal is picked up by a detector inside the optical transceiver. The main problem associated with the use of microbending devices is the narrow range of displacement measurement. 31 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 1.9. Micro-bending fiber sensor. The macro-bending effect is used for many applications and one of them is measuring human body motions. Transmitted light power in the silica optical fiber decreases exponentially when the fiber is bended. By using several optical fiber curvature sensors, it is possible to measure two or more degrees of freedom of motion. One classic example is the use of simple silica fiber to determine angular movements of human joints. Kyoobin Lee et al. [67] presented the application for physically handicapped persons whose arms are disabled. The wearable master device is used for human shoulder motions. It has also shown that a subject can control a two-wheeled mobile robot like a wheelchair. Another example [68] demonstrates the feasibility of using flexible polymer optical fibers to measure the respiratory rate and to evaluate the type of breathing. The fibers that react to applied pressure were integrated into a carrier fabric to form a wearable sensing system. This wearable system enables to keep track of the way of breathing (diaphragmatic, upper costal and mixed) when the sensor is placed at different positions of the torso. In their study Jao-Hwa Kuang et al. [69] presented a sensitive plastic optical fiber displacement sensor based on cyclic bending. The POF sensor is pressed by cylindrical models without surface damage. Dual bending model is used to increase the sensitivity of the POF displacement sensor. The results showed that the sensitivity can be improved by regulating the number of rollers, the distance between top and bottom plates, and the interval between two rollers. Sensors based on the polymer optical fiber macrobending technique have a wide measurement range, but within this range, the resolution is very low. Polymer optical fiber with disturbance on the outer and/or inner side of a U-shaped bent fiber (Fig. 1.10) can be utilized as a displacement, high sensitive sensor [70] for different medical applications, such as respiratory monitoring [71] or smart bed [72]. Use of a graded-index fiber instead of a step-index fiber for bending measurements resulted in an increase in sensitivity [73]. The fiber deformation sensor with the highest sensitivity to bending (15 half-loops, 3 imperfection areas), had a change in radius with a resolution of 20 µm, over a range from 25 to 50 mm. Additional sensitivity to bending and optimization of the sensor can be obtained by varying the different abrasion angle, imperfection location angle, imperfection displacement and V-grove cavity depth, and by splitting the imperfections into multiple small imperfections rather than a single large imperfection [74]. 32 Chapter 1. Health and Wellness Fiber Optic Sensors in IoT Application Fig. 1.10. U-shaped fiber optic sensing element. Hetero-core optical fiber The development of a new sensitive glove is described in [75] using hetero-core fiberoptic nerve sensors, in which the glove not only works as a motion capture tool but also as a tactile sensing device and detects the angles of finger joints. Hetero-core optical fibers are fabricated by inserting a small portion of fiber with a smaller core diameter into two identical fibers with larger core diameters which is illustrated in Fig. 1.11. The cladding diameters of the fibers should be the same. The principle behind the phenomenon is the same with the evanescent wave sensors, but hetero-core optical fibers are easier to fabricate, since control of section length is easier compared to etching. Fig. 1.11. Hetero-core optical fiber structure. Hetero-core fiber optic sensors have great sensitivity because of the mode coupling taking place at the splice region. Since some of the power is coupled to the cladding, the leakage gets easier by an external effect which is to be detected. Since there is a great difference in core diameters, light can largely leak into the cladding part after the splice. Because of this structure, the light in the cladding may be affected by environmental conditions easily. 1.3.1.2. Interferometric Fiber Optic Sensors Another important type of a medical fiber optic sensor due to their very high sensitivities is sensors using in-fiber interferometry. Interferometric sensor relies on induced phase 33 Advances in Optics: Reviews. Book Series, Vol. 3 change in light propagating through the optical fiber. Fiber optic interferometers have been developed in their most popular configurations, i.e., Michelson, Mach-Zehnder, and Sagnac interferometers [76-77] and also fiber-optic speckle interferometry that, for example, have been used for a human vital signs monitoring [78]. W. B. Spillmann Jr et al. [79] presented the result aimed at the development of a smart bed to non-intrusively monitor patient respiration, heart rate and movement using spatially distributed integrating multimode fiber optic sensors. The system consists of two spatially integrating fiber optic sensors, one of which is based on inter-modal interference and the other on mode conversion. The sensing fiber is integrated into a bed. The basic concept is that any patient movement that also moved an optical fiber within the specified area would produce a change in optical signal that would indicate patient movement. The physical repetitive movement caused by respiration or heart pumping is contained within the signal as well and can be extracted via appropriate signal processing. Two different modal modulation approaches are used with 200 μm core step index silica multimode optical fibers excited by a coherent laser source. The fiber is arranged on the mattress in two sinusoidal overlapping patterns arranged orthogonal to each other so that the fiber in each pattern crossed the fiber in the other pattern at an angle of 90°. In the statistical mode sensing all guided modes of the fiber are excited and then detected by a low cost digital camera. This is shown schematically in Fig. 1.12. Fig. 1.12. Interferometric fiber optic sensor. In the interferometric fiber-optic bending sensors, the light is split and guided along two different paths, or it is coupled into different optical modes. Recombination of light from the different paths or modes generates interference pattern that is sensitive to fiber bending. Various implementations of the interferometric fiber-optic bending sensors have been proposed, including sensor based on: multimode interference [80], single mode structure [81] and photonics crystal fiber [82]. H. Qu et al. [83] demonstrated an interferometric fiber-optic bending/nano-displacement sensor based on a plastic dual-core fiber with one end coated with a silver mirror. The two fiber cores are first excited with the same laser beam, the light in each core is then back-reflected at the mirror-coated fiber-end, and, finally, the light from the two cores is made to interfere at the coupling end. Bending of the fiber leads to shifting interference fringes that can be interrogated 34 Chapter 1. Health and Wellness Fiber Optic Sensors in IoT Application with a photodetector. Experimentally it was found that the resolution of the sensor is ~3 × 10-4 m-1 for sensing of bending curvature, as well as ~70 nm for sensing of displacement of the fiber tip. 1.3.1.3. Wavelength-Modulated Sensors Fibers with replaced cladding Various techniques have been devised by which a measured modifies the spectrum of a guided light is being measured. A simple example includes glucose fiber optic sensor. A fiber based pH meter has been developed in which the cladding material is replaced with polyaniline polymer, a polymer with broad sensitivity to pH [84]. The sensor is modified by using glucose oxidase immobilised on the polyaniline polymer surface (an enzyme which converts glucose to glucuronic acid, resulting in a pH change) to predict glucose concentration [85]. The near-IR evanescent wave sensor was converted into a glucose sensor by immobilizing glucose oxidase on the surface of the polymer. The enzyme converted glucose to gluconic acid and the resulting pH change was used to predict glucose concentrations. Standard errors of prediction for the concentration range of 0 to 20 mM were 0.25 mM for distilled water and 0.8 mM for buffer solutions. Lossy Mode Resonances (LMR) One of the most relevant factors that make LMRs a good choice for optical fiber sensors (humidity sensors) development is their ability to generate an optical phenomenon that can be detected by the wavelength detection method with the same material that acts as the sensitive layer to the parameter to be measured. The structure of a LMR-based device consists of a waveguide, which allows for accessing the evanescent field, coated with a thin film of the appropriate material. The condition for LMR generation is that the real part of the thin film permittivity is positive and higher in magnitude than both its own imaginary part and the real part of the material surrounding the thin film [86]. LMRs are generated when there is a resonant coupling of light to modes guided in the external coating. The large amount of available materials enables the development of optical fiber sensors for a wide range of applications [87-88]. Fiber Bragg Grating (FBG) FBG sensors have caught attention in the last decade, due to their distinguishing advantages when compared with other sensors. First, they are not sensitive to the light source amplitude fluctuations, since the readout mechanism is based on wavelength instead of light intensity. Second, the Bragg structure is directly written into the fiber core, keeping the overall fiber structure unaffected. FBG sensors are becoming increasingly attractive for healthcare and different wellbeing application [89-90]. Fiber Bragg Gratings are fibers that reflect particular wavelengths of light and transmit all others. This is achieved by creating a periodic variation in the refractive index of the fiber core. At each periodic refraction change, a small amount of light is reflected. All the 35 Advances in Optics: Reviews. Book Series, Vol. 3 reflected light signals combine coherently to one large reflection at a particular wavelength when the grating period is approximately half the input light's wavelength. This is referred to as the Bragg condition, and the wavelength at which this reflection occurs is called the Bragg wavelength. Light signals at wavelengths other than the Bragg wavelength, which are not phase matched, are essentially transparent [91]. This principle is shown in Fig. 1.13. Fig. 1.13. Fiber Bragg Grating. For example, Kalinowski et al. [92] presented the application of FBG for the measurement of bone deformation under load. As the FBG is sensitive to both temperature and deformation, the two parameters can be measured simultaneously using two FBGs with different thermal and deformation sensitivities [93]. FBG sensors have several advantages over existing imaging modalities and measuring methods, which make them well-suited for use in a clinical environment, especially for a flexible surgical instruments. Flexible minimally invasive surgical instruments can be used to reach difficult-to-reach locations within the human body. Accurately steering these instruments requires information about the shape of the instrument. In order for FBGs to measure shape and 3D deformations in general they are used in orthogonally arranged arrays where each FBG measures one component of the 3D strain as a wavelength shift of its reflection spectrum peak [94]. 3D shape sensing could be realised with series of FBGs embedded off axis along different directions inside optical fibers [95]. A novel approach has been recently demonstrated for 3D shape sensing by using two weakly titled Bragg gratings spliced together such that their tilt plane directions are oriented ~90 degree from each other [96-97]. The transmission spectrum of this device has two resonances that respond differentially to bending along perpendicular directions, one for each of the tilted FBG. Results demonstrate that bending directions from 0 to 180 degrees and curvature magnitudes between 0 and 3 m−1 can be extracted from each pair of resonance transmission values in a single measurement using unpolarized light. The measurement sensitivities obtained from the two TFBGs range from −0.33 to + 0.21 dB/m-1, depending on orientation. Another example of FBG’s biomedical usage is a skin layer that can be easily attached to the subject’s body under monitoring. The solution relies on the development of a methodology to embed FBG sensors on a flexible polyvinyl chloride skin foil [98], using standard industrial fabrication processes. It is possible to achieve a good bonding between the sensor and the foil, which allows the sensor to track with success the applied 36 Chapter 1. Health and Wellness Fiber Optic Sensors in IoT Application displacement to the foil. A linear response of the sensor with a slope of 7.8 nm per 1 % elongation emphasizes its performance. Long-period fiber gratings (LPG) Long-period fiber gratings (LPG) consist of a periodic modification of the refractive index of the core (Fig. 1.14) of a single-mode optical fiber [99]. In opposition to FBG, which have a sub-micron period and couple light from the forward-propagating mode of the optical fiber to a backward counter-propagating mode, LPGs have a period typically in the range of 100 μm to 1 mm. This provokes in LPGs a coupling of light between the guided core mode and various co-propagating cladding modes. Coupling to the cladding modes is wavelength selective, resulting in a series of attenuation bends in the transmission spectrum. Fig. 1.14. Long-period fiber gratings. A long period grating coated with hydrogel has been developed [100] for monitoring the relative humidity (RH) level which has an important influence on several biomedical processes [101]. The response wavelength of the LPG varies with the changes of the relative humidity. The changes are based on the effect of the hydrogel thin film, the index of which increases with the increasing RH. Experiment shows that the hydrogel-coated LPG sensor is highly sensitive between 38.9 % and 100 % RH, and behaves linearly with humidity with an accuracy of ±2.3 % RH. The feasibility to monitor health condition (breathing efforts) of patients was demonstrated [102], by using silica fiber sensor woven into bandage or attached onto garment based on LPGs technique. However, the poor compatibility of the sensors with industrial textile processes limits their use for medical monitoring purposes. 1.3.2. Multicore Fiber Recently, there has been increased interest in fiber optic sensors for medical applications based on multicore fibers. These include distributed temperature [103], strain [104], and shape [105]. 37 Advances in Optics: Reviews. Book Series, Vol. 3 The complete integrated optical fiber assembly suitable for shape sensing has been developed [106]. Shape sensing using optical fibers exploit the strain sensitivity of light propagating in an optical fiber waveguide core. When such a core is offset from the center of a fiber it experiences a strain that depends on the curvature of the fiber. With more than one offset core, the direction of the bend may also be determined. Sensitivity to fiber twist can be introduced by adding a permanent twist to the outer fiber cores. In this way, when the fiber is twisted, the outer cores will all be strained in the same way, while the center remains unstrained. Several fiber designs used to satisfy this set of sensing requirements. The most straightforward design has offset cores at equal radius surrounding a central core. The module consists of a length (> 1 m) of twisted multicore optical fiber with fiber Bragg gratings inscribed along its length. The fiber has a compact 180 micron coated diameter, a twist of 50 turns per meter and grating reflectivities greater than 0.01 % per cm of array, suitable for high efficiency scatter measurements over many meters of fiber. Light signals from a series of fiber Bragg gratings (FBGs) inscribed along the length of a single-core optical fiber (SCF) can quantify the strain, temperature, and pressure experienced by that fiber. The technical team at FBGS (Jena, Germany) inscribed a high density of FBGs within a multicore fiber (MCF) using their draw-tower grating (DTG) technology to perform shape sensing [107]. To identify two- or three-dimensional spatial information, either multiple SCFs must be integrated into a multifiber bundle or, a single MCF can be used. DTG-MCF sensors are mechanically robust, lightweight and compact, immune to ionized radiation, and inert-beneficial attributes when integrated with advanced minimally invasive devices targeted for clinical diagnostic, therapeutic, or monitoring applications. Another interesting type of a fiber bending sensor is based on multicore fiber and long period fiber gratings [108]. Long period grating was UV inscribed into a multicore fiber consisting of 120 single mode cores. The multicore fiber that hosts the grating was fusion spliced into a single mode fiber at both ends. The spectral characteristics of this device were a function of fiber’s curvature. The device yielded a significant spectral sensitivity as high as 1.23 nm/m-1 and 3.57 dB/m-1 to the ultra-low curvature values from 0 to 1 m-1. 1.4. IoT Systems for the Family Based on Fiber Optic Sensor In this section, we will introduce an IoT platform, applied in a Device-to-Cloud communication model (Fig. 1.4), for family wellness based on fiber optic sensors. This IoT system combines different modules that are suitable for different family members. Each module is built from a variety of objects (Things) - sensors and electronic devices. The sensors monitor and collect in real time the different physiological parameters which are then sent to the cloud and processed there. In case any deviation from norms in the wellness/health properties is identified, an algorithm in the cloud generates a command to provide an appropriate response for correction of the deviated parameter. The response occurs automatically using the internet so that the system runs by itself without any human intervention. In that configuration, the objects 'sense' their environment, 'talk' to each other and react respectively to a given situation. 38 Chapter 1. Health and Wellness Fiber Optic Sensors in IoT Application As can be seen in Fig. 1.15, the system consists of objects, communication components and cloud. Each sensing object monitors continuously and consistently several physiological parameters of the users: breathing depth, exhalation and inhalation time, pulse rate, etc. In each module, a microprocessor samples the data from the sensors and performs an initial processing of calculating the health/wellness properties. A Wi-Fi communication component sends the processed data to the cloud for storing. Additional information is collected to the cloud from websites - the weather condition and typical physiological parameters of the person according to his/her age. The data from the sensors is analyzed according to an algorithm in the cloud along with the data from the internet and the stored data in the cloud. Then, the algorithm generates an "action parameter" – a command to execute the appropriate reaction based on the identified scenario. The action parameter activates the dedicated device in the module automatically so that each different action parameter causes a different automatic response. Fig. 1.15. IoT system diagram. The objects part is modular so that other modules with suitable sensors and devices can be connected to the system. 39 Advances in Optics: Reviews. Book Series, Vol. 3 The system objects act according to different situations and they are connected to each member of the family: child, baby, adults, or disabled and elderly, so that the system improves their life and prevents disaster from happening. The system's modules working principle is described as follows: 1. Child module - The child is wearing a respiration [53, 71] and a pulse sensors [109]. In order to ignore the measurement’s noises, the respiration sensor samples each time five cycles of breath and the processor calculates the average exhalation and inhalation time and depth of breathing. The whole health/wellness history of the child is stored and every time new information is received, it is analyzed and the action parameter is generated as follows: if, for example, the breathing depth is 30 % lower than normal, or when the breathing rate is slower or faster by 50 % than the typical, an alarm SMS message is sent to the parents. In a more unusual case, when the change in breathing depth is 70 % lower than normal, or the breathing rate is abnormal compared to the age of the child, an alert is sent immediately to an ambulance service with the child's location. 2. Disabled and elderly module - a sensor [110] that samples the temperature in the environment of the human is installed. In addition, breathing and pulse sensors are installed in the chair. The data is received in the cloud, and as explained above, it is processed with additional data. Depending on the outside weather conditions, taken from a weather site, the cloud algorithm decides how to respond. If it is a summer day and the temperature in the human environment is higher than 28 degrees Celsius, the cloud triggers the air conditioner at home. Alternatively, on a winter day when the temperature is below 21 degrees Celsius, the cloud sends a command to activate the heating system or the radiator. Additionally, a motion sensor is attached to the human body. In case of recognizing a fall of the human, the system alerts the disabled or elderly's family and calls for help automatically. 3. Baby's cradle module – breathing and temperature sensors are installed in the cradle and monitor these parameters. The optimal temperature for the baby's environment, especially in the first year, is between 20-22 degrees Celsius [111]. When the temperature in the bed area is not suitable, the cloud sends an order to activate heating or cooling of the bed in order to reach the optimal temperature. Additionally, the cloud learns to recognize when the baby is calm using a motion sensor. When the baby is restless, the cloud generates a comment to activate a relaxing music in the bed until the parent or the nanny arrives. 4. Adults module – studies have shown [112] that maintaining an optimal body temperature enhances the human performance including working memory, subjective alertness and visual attention. The adult module contains a temperature sensor, which is attached to the adult body and measures the body temperature. The adult wears a unique cloth or shoe with a built-in temperature regulator which is triggered by the cloud. The system receives the body temperature and, combined with the area temperature, determines the optimal temperature and activates the temperature regulator accordingly. This IoT system introduces how things (objects) communicate between them. The main peculiarity of such a system isn't inherent in sending the wellness/health information to 40 Chapter 1. Health and Wellness Fiber Optic Sensors in IoT Application the doctor in real time, but in the system that reacts autonomously. In addition the combination of data from physical sensors can be combined with an appropriate internet data. These kind of systems presents the real IoT system – the ability of things to speak between them and act by themselves. Moreover, the system is modular and can be combined with many more objects having a lot more capabilities for identifying different situations, act automatically and improve the family wellness. The system introduced the enormous potential inherent in wellness IoT systems particularly and in real IoT systems in general. Such systems can 'sense' their environment, use virtual sensors data by collecting it from the internet, act autonomously according to identified situations without any human intervention and bring us closer to an automatic world. 1.5. Conclusion and Future Prospect Since the discovery of IoT, there has been tremendous increase in the number of publications in the field and different applications based on these systems have been demonstrated. In recent years, IoT has slowly developed into a useful platform that has potential applications in manufacturing, healthcare, insurance, business services, airline, media and entertainment. In particular, the biomedical applications of Internet of Things have motivated leading researchers and companies to develop very efficient sensing systems based on IoT platform. Internet of Things devices such as fitness trackers, skin sensors, glucose meters and cardiac monitors have created a thriving industry that not only puts people in control of their vital health parameters, but also enables them to engage with healthcare providers in new ways. It should be mentioned that the recent developments in fiber optic sensor technology have contributed largely for the advancement of IoT based applications. It is now possible to obtain very sensitive, reliable, as well as small size and cost structures, that in conjunction with IoT health and wellness system have the ability to greatly improve health outcomes, quality of life and real-time support or intervention. Today it is becoming common practice for patients to track their physiological parameters at home and send the data wirelessly to medical centers or private doctors. It is also routine for doctors to see images in their smartphones and forward them over the Internet to be reviewed by specialists anywhere and anytime. We now already have tools like patient’s data and non-invasive sensor technologies to help us increase medication adherence and allow people to better manage their own health with reduced costs. In the near future, we hope, most of the sensing devices will be communicating on our behalf—they will be interacting with the physical and virtual worlds more than interacting with us. This will be the true realization of the idea of Internet of Things. References [1]. Z. Pang, Technologies and architectures of the Internet of Things (IoT) for health and wellbeing, MS Thesis, KTH Royal Institute of Technology, Stockholm, Sweden, Jan. 2013. 41 Advances in Optics: Reviews. Book Series, Vol. 3 [2]. [3]. [4]. [5]. [6]. [7]. [8]. [9]. [10]. [11]. [12]. [13]. [14]. [15]. [16]. [17]. [18]. [19]. [20]. [21]. [22]. [23]. 42 K. Ashton, That 'Internet of Things' thing, RFID Journal, Vol. 22, 2009, pp. 97-114. G. Wu, S. Talwar, K. Johnsson, N. Himayat, K. D. Johnson, M2M: from mobile to embedded internet, IEEE Communications Magazine, Vol. 49, No. 4, 2011, pp. 36-43. C. Anton-Haro, M. Dohler, Machine-to-Machine (M2M) Communications: Architecture, Performance and Applications, Elsevier - Technology & Engineering, 2014. K. Chen, Machine-to-machine communications for healthcare, Journal of Computing Science and Engineering, Vol. 6, No. 2, June 2012, pp. 119-126. B. Burton, D. A. Willis (analysts), Gartner's Hype Cycle Special Report for 2014, Gartner Inc., 2014. M. James, M. Chui, P. Bisson, J. Woetzel, R. Dobbs, J. Bughin, D. Aharon, The Internet of Things: Mapping the Value Beyond the Hype, McKinsey Global Institute, June 2015. T. Danova, Morgan Stanley: 75 Billion Devices Will Be Connected To The Internet Of Things By 2020, Business Insider, Oct. 2, 2013, http://www.businessinsider.com/75billion-devices-will-be-connected-to-the-internet-by-2020-2013-10 B. Alessio, W. de Donato, V. Persico, A. Pescapé, Integration of cloud computing and internet of things: A survey, Future Generation Computer Systems, Vol. 56, 2016, pp. 684-700. B. Haobijam, Y. Huang, T. Lee, Intuitive IoT-based H2U healthcare system for elderly people, in Proceedings of the IEEE 13th International Conference on Networking, Sensing and Control (ICNSC’16), México City, Mexico, 2016, pp. 1-6. M. de Fátima, F. Domingues, A. Radwan, Optical Fiber Sensors for loT and Smart Devices, Springer, Briefs in Electrical and Computer Engineering, 2017. M. McGrath, C. N. Scanaill, Sensor Technologies: Healthcare, Wellness and Environmental Applications, Apress, 2013. S. Taylor, The Internet of Sensors, not the Internet of Things, Press Release, TTP, May 8, 2015. A. Dastjerdi, R. Buyya, Fog computing: Helping the internet of things realize its potential, Computer, Vol. 49, Issue 8, 2016, pp. 112-116. B. Esposito, A. Castiglione, F. Pop, K. K. R. Choo, Challenges of Connecting Edge and Cloud Computing: A Security and Forensic Perspective, Cloud Computing, Vol. 4, 2017, pp. 13-17. H. Tschofenig, D. Thaler, Architectural Considerations in Smart Object Networking, Tech. No. RFC 7452, Internet Architecture Board, March 2015, https://www.rfceditor.org/rfc/rfc7452.txt Mapmyfitness, Mobile Apps, http://www.mapmyfitness.com/app L. Atzori, A. Iera, G. Morabito, The internet of things: A survey, Computer Networks, Vol. 54, Issue 15, 2010, pp. 2787-2805. M. Kranz, P. Holleis, A. Schmidt, Embedded interaction: interacting with the internet of things, Internet Computing, Vol. 14, Issue 2, 2010, pp. 46-53. A. Pentland, Healthwear: medical technology becomes wearable, Computer Society, Vol. 37, Issue 5, 2004, pp. 42-49. H. Alemdar, C. Ersoy, Wireless sensor networks for healthcare: A survey, Computer Networks, Vol. 54, Issue 15, 2010, pp. 2688-2710. S. Datta, C. Bonnet, A. Gyrard, R. Pedro, F. da Costa, K. Bouda, Applying Internet of Things for personalized healthcare in smart homes, in Proceedings of the Wireless and Optical Communication Conference (WOCC’15), 2015, pp. 164-169. G. Yang, L. Xie, M. Mantysalo, X. Zhou, Z. Pang, L. Da Xu, S. Kao-Walter, Q. Chen, L. Zheng, A health-IoT platform based on the integration of intelligent packaging, unobtrusive bio-sensor and intelligent medicine box, Transactions on Industrial Informatics, Vol. 10, Issue 4, 2014, pp. 2180-2191. Chapter 1. Health and Wellness Fiber Optic Sensors in IoT Application [24]. [25]. [26]. [27]. [28]. [29]. [30]. [31]. [32]. [33]. [34]. [35]. [36]. [37]. [38]. [39]. [40]. [41]. [42]. U. Bakar, H. Ghayvat, S. Hasanm, S. Mukhopadhyay, Activity and Anomaly Detection in Smart Home: A Survey, Chapter 8, in Next Generation Sensors and Systems (Mukhopadhyay, S. Chandra, Ed.), Springer, 2016, pp. 191-220. H. Ghayvat, J. Liu, M. E. Alahi, S. C. Mukhopadhyay, X. Gu, Internet of Things for smart homes and buildings: Opportunities and Challenges. Australian Journal of Telecommunications and the Digital Economy, Vol. 3, Issue 4, 2015, pp. 33-47. N. K. Suryadevara, S. C. Mukhopadhyay, S. D. T. Kelly, S. P. S. Gill, WSN-based smart sensors and actuator for power management in intelligent buildings, IEEE/ASME Transactions on Mechatronics, Vol. 20, Issue 2, 2015, pp. 564-571. J. Sherly, D. Somasundareswari, Internet of Things based smart transportation systems, International Research Journal of Engineering and Technology (IRJET), Vol. 2, Issue 7, 2015, pp. 1207-1210. L. Xiao, Z. Wang, Internet of Things: A new application for intelligent traffic monitoring system, Journal of Networks, Vol. 6, Issue 6, 2011, pp. 887-894. A. Zanella, N. Bui, A. Castellani, L. Vangelista, M. Zorzi, Internet of Things for smart cities, Internet of Things Journal, Vol. 1, 2014, pp. 22-32. P. Fraga-Lamas, T. M. Fernández-Caramés, M. Suárez-Albela, L. Casted, M. GonzálezLópez, A review on Internet of Things for defense and public safety, Sensors, Vol. 16, 2016, 1644. P. Lopez, D. Fernandez, A. J. Jara, A. F. Skarmeta, Survey of Internet of Things technologies for clinical environments, in Proceedings of the 27th International Conference on Advanced Information Networking and Applications Workshops (WAINA’13), 25-28 March 2013, pp.1349-1354. J. Groopman, S. Etlinger, Consumer Perceptions of Privacy in the Internet of Things, Altimeter Group, June 2015. D. Rodbard, Continuous glucose monitoring: A review of successes, challenges, and opportunities, Diabetes Technology & Therapeutics, Vol. 18, 2016, pp. S3-S13. P. Herrero, L. Beatriz, C. Martin, Pepper: Patient empowerment through predictive personalised decision support, in Proceedings of the ECAI Workshop on Artificial Intelligence for Diabetes, 2016, pp. 8-9. A. Naditz, Use of GlowCap e-bottle cap leads to greater medicinal compliance among patients, Telemedicine and e-Health, Vol. 16, Issue 7, 2010, pp. 759-760. H. O. Kucuk, U. Kucuk, M. Yalcin, Z. Isilak, Time to use mobile health devices to diagnose paroxysmal atrial fibrillation, International Journal of Cardiology, Vol. 222, 2016, p. 1061. D. M. West, Improving health care through mobile medical devices and sensors, Brookings Institution Policy Report, Issue 10, 2013, pp. 1-13. IoT applications spanning across industries, Internet of Things BLOG, https://www.ibm.com/blogs/internet-of-things/iot-applications-industries J. Hailstone, A. E. Kilding, Reliability and validity of the Zephyr™ BioHarness™ to measure respiratory responses to exercise, Measurement in Physical Education and Exercise Science, Vol. 15, Issue 4, 2011, pp. 293-300. M. D. Reeve, K. L. Pumpa, N. Ball, Accuracy of the SenseWear Armband Mini and the BodyMedia FIT in resistance training, Journal of Science and Medicine in Sport, Vol. 17, Issue 6, 2014, pp. 630-634. S. S. Bhunia, S. K. Dhar, N. Mukherjee, iHealth: A fuzzy approach for provisioning intelligent health-care system in smart city, in Proceedings of the IEEE 10th International Conference on Wireless and Mobile Computing, Networking and Communications (WiMob’14), Larnaca, 2014, pp. 187-193. Your Ring of Security Starts at the Front Door, Ring, https://ring.com 43 Advances in Optics: Reviews. Book Series, Vol. 3 [43]. J. L. Jones, Robots at the tipping point: the road to iRobot Roomba, Robotics & Automation Magazine, Vol. 13, Issue 1, March 2006, pp. 76-78. [44]. R. Bloss, Wearable sensors bring new benefits to continuous medical monitoring, real time physical activity assessment, baby monitoring and industrial applications, Sensor Review, Vol. 35, Issue 2, 2015, pp. 141-145. [45]. P. Dutta, I. Mohomed, A great time for makers, GetMobile: Mobile Computing and Communications, Vol. 19, Issue 4, 2016, pp. 5-7. [46]. Industrial Networking, Cisco, http://www.cisco.com/c/en/us/solutions/industries/manufacturing/industrialnetworking.html?stickynav=1 [47]. J. Macaulay, L. Buckalew, G. Chung, Internet of Things in Logistics, DHL Trend Research & Cisco Consulting Services, 2015. [48]. Industrial Internet of Things use cases: The IoT at sea, Internet of Things blog, https://www.ibm.com/blogs/internet-of-things/the-iiot-at-sea [49]. W. B. Spillman Jr., Fiber Optic Sensors: An Introduction for Engineers and Scientists, 2nd Edition, John Wiley & Sons, USA, 2011. [50]. A. Leung, P. M. Shankar, R. Mutharasan, A review of fiber-optic biosensors. Sensors and Actuators B: Chemical, Vol. 125, Issue 2, 2007, pp. 688-703. [51]. R. Y. Shah, Y. K. Agrawal, Introduction to fiber optics: Sensors for biomedical applications, Indian Journal of Pharmaceutical Sciences, Vol. 73, Issue 1, 2011, pp. 17-22. [52]. P. Roriz, L. Carvalho, O. Frazão, J. L. Santos, J. A. Simões, From conventional sensors to fibre optic sensors for strain and force measurements in biomechanics applications: A review, Journal of Biomechanics, Vol. 47, Issue 6, 2014, pp. 1251-1261. [53]. A. Babchenko, B. Khanokh, Y. Shomer, M. Nitzan. Fiber optic sensor for the measurement of respiratory chest circumference changes, Journal of Biomedical Optics, Vol. 4, Issue 2, 1999, pp. 224-229. [54]. J. Webster, Fiber-optic temperature sensors, in Medical Instrumentation Applications and Design, 4th Edition, John Wiley & Sons, NJ, USA, 2010, pp. 74-79. [55]. J. Lin, Recent development and applications of optical and fiber-optic pH sensors, TrAC Trends in Analytical Chemistry, Vol. 19, Issue 9, 2000, pp. 541-552. [56]. Z. Chen, D. Lau, J. T. Teo, S. H. Ng, X. Yang, P. L. Kei, Simultaneous measurement of breathing rate and heart rate using a microbend multimode fiber optic sensor, Journal of Biomedical Optics, Vol. 19, Issue 5, 2014, 057001. [57]. X. D. Wang, O. S. Wolfbeis, Fiber-optic chemical sensors and biosensors, Analytical Chemistry, Vol. 85, Issue 1, 2016, pp 487-508. [58]. H. Li, H. Yang, E. Li, Z. Liu, K. Wei, Wearable sensors in intelligent clothing for measuring human body temperature based on optical fiber Bragg grating, Optics Express, Vol. 20, Issue 11, May 2012, pp. 11740-11752. [59]. N. Wu, Y. Tian, X. Zou, Y. Zhai, K. Barringhaus, X. Wang, A miniature fiber optic blood pressure sensor and its application in in-vivo blood pressure measurements of a swine model, Sensors and Actuators B: Chemical, Vol. 181, May 2013, pp. 172-178. [60]. M. J. Yin, B. Huang, S. Gao, A. P. Zhang, X. Ye, Optical fiber LPG biosensor integrated microfluidic chip for ultrasensitive glucose detection, Biomedical Optics Express, Vol. 7, Issue 5, 2016, pp. 2067-2077. [61]. I. R. Matias, S. Ikezawa, J. Corres, Fiber Optic Sensors Current Status and Future Possibilities, Springer International Publishing, Vol. 21, Switzerland, 2017. [62]. L. E. Dunne, P. Walsh, S. Hermann, B. Smyth, B. Caulfield, Wearable monitoring of seated spinal posture, Transaction on Biomedical Circuits Systems, Vol. 2, Issue 2, 2008, pp. 97-105. [63]. M. A. Zawawi, S. O'Keeffe, E. Lewis, Plastic optical fibre sensor for spine bending monitoring, Journal of Physics: Conference Series, Vol. 450, Issue 1, 2013, 012004. 44 Chapter 1. Health and Wellness Fiber Optic Sensors in IoT Application [64]. K. S. Kuang, W. J. Cantwell, P. J. Scully, An evaluation of a novel plastic optical fibre sensor for axial strain and bend measurements, Measurement Science and Technology, Vol. 13, Issue 10, 2002, pp. 1523-1534. [65]. A. Babchenko, V. Chernyak, J. Maryles, Bio-chemical sensor based on imperfected plastic optical fiber, Proceedings of SPIE, Vol. 6585, 2007, 65851V. [66]. C. J. Deepu, Z. Chen, J. T. Teo, S. H. Ng, X. Yang, Y. Lian, A smart cushion for real-time heart rate monitoring, in Proceedings of the Biomedical Circuits and Systems Conference (BioCAS’12), 2012, pp. 53-56. [67]. K. Lee, D.S. Kwon, Wearable Master Device Using Optical Fiber Curvature Sensors for the Disabled, in Proceedings of the IEEE International Conference Robotics and Automation, (ICRA’01), Korea, 2001, pp. 892-896. [68]. M. Krehel, M. Schmid, R. M. Rossi, L. F. Boesel, G. L. Bona, L. J. Scherer, An optical fibre-based sensor for respiratory monitoring, Sensors, Vol. 14, Issue 7, 2014, pp. 13088-13101. [69]. J. H. Kuang, P. C. Chen, Y. C. Chen, Plastic optical fiber displacement sensor based on dual cycling bending, Sensors, Vol. 10, Issue 11, 2010, pp. 10198-10210. [70]. A. Babchenko, Z. Weinberger, N. Itzkovich, J. Maryles, Plastic optical fibre with structural imperfections as a displacement sensor, Measurements Science and Technology, Vol. 17, Issue 5, 2006, pp. 1157-1161. [71]. A. Babchenko, J. Maryles, N. Itzkovich, M. Nitzan, Respiratory sensor based on macrobending fiber with side disturbances, Proceedings of SPIE, June 2005, p. TuH21. [72]. A. Babchenko, J. Maryles, Smart bed sensor based on imperfected plastic optical fibers, in Proceedings of the World Congress on Medical Physics and Biomedical Engineering (IUPESM’06), Korea, Aug. 2006, p. 198. [73]. A. Babchenko, J. Maryles, Graded-index plastic optical fiber for deformation sensing, Optics and Lasers in Engineering, Vol. 45, Issue 7, 2007, pp. 757-760. [74]. H. Poisel, C. Weigert, A. Babchenko, M. Bloos, Imperfected POF for bending sensors, in Proceedings of the Polymer Optical Fiber Conference (POF’07), Turin, Italy, 2007, pp. 10-12. [75]. M. Nishiyama, K. Watanabe, Wearable sensing glove with embedded hetero-core fiberoptic nerves for unconstrained hand motion capture, Transactions on Instrumentation and Measurement, Vol. 58, Issue 12, 2009, pp. 3995-4000. [76]. M. Zyczkowski, B. Uzieblo-Zyczkowska, Human psychophysiological activity monitoring methods using fiber optic sensors, Proceedings of SPIE, Vol. 7838, October 12, 2010, 78380W. [77]. M. Życzkowski, M. Szustakowskia, W. Ciurapińskia, B. Uziębło-Życzkowskab, Interferometric fiber optics based sensor for monitoring of the heart activity, Acta Phys. Pol. A, Vol. 120, Issue 4, 2011, pp. 782-784. [78]. P. Podbreznik, D. Djonlagic, D. Lesnik, B. Cigale, D. Zazula, Fiber-optic speckle interferometry for unobtrusive heartbeat monitoring, Journal of the Laser and Health Academy, Vol. 2013, Issue 2, 2013, pp. 27-30. [79]. W. B. Spillman Jr., M. Mayer, J. Bennett, J. Gong, K. E. Meissner, B. Davis, R. O. Claus, A. A. Muelenaer Jr., X. Xu, A 'smart' bed for non-intrusive monitoring of patient physiological factors, Measurement Science and Technology, Vol. 15, Issue 8, 2004, pp. 1614-1620. [80]. Y. Gong, T. Zhao, J. Rao, Y. Wu, All-fiber curvature sensor based on multimode interference, Photonic Technology Letters, Vol. 23, Issue 11, 2011, pp. 679-681. [81]. S. Silva, O. Fraza, J. Viegas, L. A. Ferreira, F. M. Araujo, F. X. Malcata, J. L. Santos, Temperature and strain-independent curvature sensor based on a singlemode/multimode fiber optic structure, Measurement Science and Technology, Vol. 22, Issue 8, 2011, 085201. 45 Advances in Optics: Reviews. Book Series, Vol. 3 [82]. J. Villatoro, V. P. Minkovich, J. Zubia, Photonic crystal fiber interferometric force sensor, Photonics Technology Letters, Vol. 27, Issue 11, June 2015, pp. 1181-1184. [83]. H. Qu, G. F. Yan, M. Skorobogatiy, Interferometric fiber-optic bending/nano-displacement sensor using plastic dual-core fiber, Optics Letters, Vol. 39, Issue 16, 2014, pp. 4835-4838. [84]. Z. F. Ge, Fiber optic near and mid infrared spectroscopy and clinical applications, Diss. Abstr. Int., Vol. 55, 1995, 3855. [85]. C. W. Brown, C. S. Chen, Y. Li, Near and mid infrared chemical and biological sensors, in Proceedings of the SPIE's International Symposium on Optical Science, Engineering and Instrumentation, 1995, pp. 243-250. [86]. M. Hernáez, I. Del Villar, C. R. Zamarreño, F. J. Arregui, I. R. Matias, Optical fiber refractometers based on lossy mode resonances supported by TiO2 coatings, Applied Optics, Vol. 49, Issue 20, 2010, pp. 3980-3985. [87]. P. Sanchez, C. R. Zamarreño, M. Hernaez, I. Del Villar, C. Fernandez-Valdivielso, I. R. Matias, F. J. Arregui, Lossy mode resonances toward the fabrication of optical fiber humidity sensors, Measurement Science and Technology, Vol. 23, Issue 1, 2012, 014002. [88]. A. B Socorro, I. Del Villar, J. M. Corres, F. J. Arregui, I. R. Matias, Tapered single-mode optical fiber pH sensor based on lossy mode resonances generated by a polymeric thin-film, Sensors Journal, Vol. 10, Issue 8, 2012, pp. 2598-2603. [89]. C. Massaroni, P. Saccomandi, E. Schena, Medical smart textiles based on fiber optic technology: An overview, Journal of Functional Biomaterials, Vol. 6, Issue 2, 2015, pp. 204-221. [90]. V. Mishra, N. Singh, U. Tiwari, P. Kapur, Fiber grating sensors in medicine: Current and emerging applications, Sensors and Actuators A: Physical, Vol. 167, Issue 2, 2011, pp. 279-290. [91]. Fiber Bragg Grating (FBG) Sensor Principle, www.fbgs.com/technology/fbg-principle [92]. A. Kalinowski, L. Z. Karam, V. Pegorini, A. B. Di Renzo, C. S. R Pitta, R. Cardoso, T. S. Assmann, H. J. Kalinowski, J. C. C. da Silva, Optical fiber Bragg grating strain sensor for bone stress analysis in bovine during masticatory movements, Sensors Journal, Vol. 17, Issue 8, Feb. 2017, pp. 2385-2392. [93]. M. R. Mokhtar, T. Sun, K. T. V. Grattan, Bragg grating packages with nonuniform dimensions for strain and temperature sensing, Sensors Journal, Vol. 12, Issue 1, 2012, pp. 139-144. [94]. J. Yi, X. Zhu, H. Zhang, L. Shenand, X. Qiao, Spatial shape reconstruction using orthogonal fiber Bragg grating sensor array, Mechatronics, Vol. 22, Issue 6, 2012, pp. 679-687. [95]. C. Waltermann, A. Doering, M. Köhring, M. Angelmahr, W. Schade, Cladding waveguide gratings in standard single-mode fiber for 3D shape sensing, Optics Letters, Vol. 40, Issue 13, 2015, pp. 3109-3112. [96]. J. Albert, L. Y. Shao, C. Caucheteur, Tilted fiber Bragg grating sensors, Laser & Photonics Reviews, Vol. 7, Issue 1, 2013, pp. 83-108. [97]. D. Feng, W. Zhou, X. Qiao, J. Albert, Compact optical fiber 3D shape sensor based on a pair of orthogonal tilted fiber Bragg gratings, Scientific Reports, Vol. 5, 2015, 17415 [98]. A. F. da Silva, A. F. Gonçalves, L. A. de Almeida Ferreira, F. M. M. Araújo, P. M. Mendes, J. H. Correia, A smart skin PVC foil based on FBG sensors for monitoring strain and temperature, Transactions on Industrial Electronics, Vol. 58, Issue 7, 2011, pp. 2728-2735. [99]. S. W. James, R. P. Tatam, Optical fibre long-period grating sensors: Characteristics and application, Measurement Science and Technology, Vol. 14, Issue 5, 2003, pp. R49-R61. [100]. L. Wang, Y. Liu, M. Zhang, D. Tu, X. Mao, Y. Liao, A relative humidity sensor using a hydrogel-coated long period grating, Measurement Science and Technology, Vol. 18, Issue 10, 2007, pp. 3131-3134. 46 Chapter 1. Health and Wellness Fiber Optic Sensors in IoT Application [101]. C. Larsson, L. Davidsson, P. Lundin, G. Gustafsson, M, Vegfors, Respiratory monitoring during MR imaging: The clinical value of a new fibre-optical monitor, Acta Radiologica, Vol. 40, Issue 1, 1999, pp. 33-36. [102]. M. D. Petrović, J. Petrovic, A. Daničić, M. Vukčević, B. Bojović, L. Hadžievski, T. Allsop, G. Lloyd, D. J. Webb, Non-invasive respiratory monitoring using long-period fiber grating sensors, Biomedical Optics Express, Vol. 5, Issue 4, 2014, pp. 1136-1144. [103]. K. Gifford, B. J. Soller, M. S. Wolfe, M. E. Froggatt, Distributed fiber-optic temperature sensing using Rayleigh backscatter, in Proceedings of the 31st European Conference on IET Optical Communication (ECOC’05), Vol. 3, September 2005, pp. 511-512. [104]. M. Froggatt, J. Moore, High-spatial-resolution distributed strain measurement in optical fiber with Rayleigh scatter, Applied Optics, Vol. 37, Issue 10, 1998, pp. 1735-1740. [105]. P. S. Westbrook, K. S. Feder, T. Kremp, T. F. Taunay, E. Monberg, G. Puc, R. Ortiz, Multicore optical fiber grating array fabrication for medical sensing applications, Proceedings of SPIE, Vol. 9317, March 2015, 93170C. [106]. P. S. Westbrook, K. S. Feder, T. Kremp, T. F. Taunay, E. Monberg, J. Kelliher, R. Ortiz, K. Bradley, K. S. Abedin, D. Au, G. Puc, Integrated optical fiber shape sensor modules based on twisted multicore fiber grating arrays, Proceedings of SPIE, Vol. 8938, 2014, 89380H. [107]. B. Van Hoe, J. Van Roosbroeck, C. Voigtländer, J. Vlekken, E. Lindner, Distributed strain and curvature measurements based on tailored draw tower gratings, in Proceedings of the Avionics and Vehicle Fiber-Optics and Photonics Conference (AVFOP’16), Long Beach, CA, USA, Oct. 2016, pp. 285-286. [108]. P. Saffari, T. Allsop, A. Adebayo, D. Webb, R. Haynes, M. M. Roth, Long period grating in multicore optical fiber: an ultra-sensitive vector bending sensor for low curvatures, Optics Letters, Vol. 39, Issue 12, 2014, pp. 3508-3511. [109]. E. Suaste-Gómez, D. Hernández-Rivera, A. S. Sánchez-Sánchez, E. Villarreal-Calva, Electrically insulated sensing of respiratory rate and heartbeat using optical fibers, Sensors, Vol. 14, Issue 11, 2014, pp. 21523-21534. [110]. OMEGA, Fiber-Optic Temperature Measurement, http://www.omega.com/prodinfo/fiberoptic-temperature-measurement.html [111]. Baby Gooroo, What is the ideal temperature for my baby's room? https://babygooroo.com/articles/what-is-the-ideal-temperature-for-my-babys-room [112]. K. P. Wright, J. T. Hull, C. A Czeisler, Relationship between alertness, performance, and body temperature in humans, American Journal of Physiology-Regulatory, Integrative and Comparative Physiology, Vol. 283, Issue 6, 2002, pp. R1370-R1377. 47 Chapter 2. Multi-Point Temperature Sensor Consisting of AWG, SOA, and FBGs in Linear-Cavity Fiber Lasing System Chapter 2 Multi-Point Temperature Sensor Consisting of AWG, SOA, and FBGs in Linear-Cavity Fiber Lasing System Hiroki Kishikawa, Mao Okada, Kazuto Takahashi, Po-Jung Chen, Nobuo Goto, Yi-Lin Yu and Shien-Kuei Liaw1 2.1. Introduction Optical fiber sensing has been extensively studied in various areas, such as aging deterioration measurements of constructed buildings, seismic measurement, environmental measurement, etc. Optical fiber sensing systems are classified into two configurations. One consists of an optical source, an optical fiber transmission line, and sensing elements; the other is formed from a fiber lasing system, including sensing elements in the cavity [1-6]. Compared to the former, the latter systems have advantages such as higher resolution for wavelength-shift induced by a sensing element and higher signal-to-noise ratio (SNR). By employing a fiber Bragg grating (FBG) as the sensing element, the reflecting center wavelength can be shifted due to environmental change in temperature or in tension. In our proposed fiber sensing system, multi-point temperature or tension can be detected by employing multiple FBG elements and an arrayed waveguide grating (AWG) as an optical wavelength multi-/demultiplexer. By adopting the AWG, optical insertion loss in the cavity can be reduced. A flat-top filtering response can be easily realized by appropriate design of the integrated waveguide in the pattern of the AWG [7, 8]. In fiber lasing systems, an erbium-doped fiber amplifier (EDFA), Raman fiber amplifier, and semiconductor optical amplifier (SOA) have been employed as gain components. We consider multi-wavelength simultaneous lasing to measure temperature or tension at multiple points. To realize multi-wavelength lasing, the amplifier has to be carefully selected. When the EDFA is employed in the system, the homogeneous broadening of erbium ions limits the number of lasing wavelengths. Therefore, special lasing configurations have been proposed for multi-wavelength operation by using EDFA H. Kishikawa Dept. of Optical Science, Tokushima University, Japan 49 Advances in Optics: Reviews. Book Series, Vol. 3 [9, 10]. On the contrary, SOAs show the inhomogeneous broadening properties, which makes it possible to operate lasing at multiple wavelengths [11]. SOAs, however, have lower saturation output power compared with EDFAs, which limits the multi-wavelength lasing power. Pleros et al. reported simultaneous 38 wavelength lasing at 50 GHz spacing across a 15-nm spectral window by using two SOAs in a feedback loop [12]. Mode-locked pulse lasing at multiple wavelengths was used to identify each of the serially connected FBGs [13]. To identify each of the serial FBGs, a spectral encoding method was employed in the system [14]. Wavelength-swept pulses were also employed for multi-position fiber loop ring-down sensor array [15]. The SOAs have attracted much interest in not only optical fiber communication systems but also integrated optic devices. In optical amplification, linear amplification is required to avoid signal degradation. On the other hand, in optical signal processing using SOAs, optical nonlinearity in SOAs has been effectively used in a variety of signal processing circuits [16-18]. We consider optical lasing at multiple wavelengths. The number of wavelength channels is limited by amplifiable wavelength range, optical nonlinearity, and gain saturation in the SOA. In the proposed fiber sensing system, multi-wavelength lasing is obtained by using a single SOA [19]. Temperature at multiple points is detected by employing multiple FBG elements and an AWG. Although the SOA has a potential of multi-wavelength lasing, the number of actual lasing wavelengths is affected by physical parameters, such as the reflectivity of each FBG, cavity loss, and gain saturation of the SOA. However, there have been few studies reporting effects of such parameters in multi-point sensing systems with SOAs and FBGs. Therefore, the objective of this study is to theoretically and experimentally evaluate the influence of physical parameters with multi-point sensing characteristics by using a single SOA and multiple FBGs. Moreover, multiwavelength lasing characteristics on unequal reflecting power on each channel are especially revealed. In addition, the advantages are also clarified in the proposed sensing system configuration, employing parallel FBGs that are connected to demultiplexed ports of an AWG. By using such a star configuration, it is easy to equalize the reflecting power of each channel by using variable optical attenuators on FGBs with various reflectivities. Furthermore, it would be resilient to fiber failures [20], which means that the reliability and robustness against unexpected breakdown in a certain FBG sensing element can be improved. 2.2. Linear-Cavity Fiber Sensor Consisting of SOA, AWG, and FBGs The proposed fiber sensing system consists of multiple linear cavities lasing at different wavelengths as shown in Fig. 2.1(a). The SOA placed in the linear-cavity amplifies multiwavelength signals propagating in both directions. The AWG plays a role as a multi/demultiplexer. The passband filtering response of the AWG is schematically illustrated in Fig. 2.1(b), where a flat-top filtering passband is assumed. The base of the wavelength interval of the multi-channel lasing is lasing, which is equal to the channel interval of the AWG. The AWG demultiplexes the optical signals propagating in the right-hand direction into N signals having different wavelengths i, i = 1, … , N. The signal at wavelength i in port i is reflected by the FBG. The wavelength of each FBG is designed to be at the 50 Chapter 2. Multi-Point Temperature Sensor Consisting of AWG, SOA, and FBGs in Linear-Cavity Fiber Lasing System center of the AWG passband of each channel. The wavelength multiplexed signals propagating in the left-hand direction are amplified by the SOA. The left-end wavelengthindependent loop mirror reflects all the channel signals, where a part of the signals is coupled out for detection. The lasing spectra are also schematically shown in Fig. 2.1 (b). Since the FBG reflection wavelength depends on its environment, an environmental change results in wavelength-shift shift,i of the lasing wavelength. The maximum amount of shift,i has to be less than half of the bandwidth of each channel, AWG_FT ∕ 2. FBG 1 Loop mirror SOA 2 AWG N Detection of wavelength shift Linear cavity (a) Linear-cavity multi-channel fiber sensing system; Filtering response in AWG 1 2  shift, i N  AWG_FT   lasing (b) Lasing spectrum with AWG filtering characteristics. Fig. 2.1. Multi-point sensing system consisting of an SOA, an AWG, and FBGs. When multi-wavelength signals are amplified with the SOA, optical nonlinearity becomes an important issue. The gain saturation in the SOA limits the number of lasing channels. The induced four-wave-mixing (FWM) nonlinearity may result in degradation of sensing preciseness. When wavelengths in two neighbor lasing channels come closer each other, the amplified gain for two channels becomes unequal as discussed in Section 2.3.1. This inequality in gain might result in disappear of one of the two lasing channels. 51 Advances in Optics: Reviews. Book Series, Vol. 3 2.3. Analysis of Multi-Channel Lasing 2.3.1. Analysis for SOA Nonlinearity Optical multi-channel signal behavior through SOA amplification has been studied by many researchers. Nonlinear phenomena for multi-wavelength signals are induced by gain saturation [21] and FWM [22-24]. The wavelength range available for amplification depends on the material gain [25]. These factors cause restriction on the number of wavelength channels and sensing preciseness. The mechanism of FWM in an SOA is explained as follows [21]: interference of multiwavelength signals coupled in the SOA induces amplitude variation at the beat frequencies of the signals. Since the amplitude varying signal induces carrier variation, stimulated emission results in variation of the refractive index as well as gain variation. These gain and refractive-index variations not only modulate the incident signals but also generate sideband signals with the interval of the beat frequencies. In this section, we analyze the multi-channel signal behavior through an SOA based on the analysis reported by Connelly [26]. The analysis is composed of the traveling-wave equations for signal fields and spontaneous emission, carrier–density rate equation, and material gain modeling as discussed in Appendix. The numerical simulation was performed by using OptiSystem (Optiwave Systems Inc.). The simulated wavelength dependence of the optical gain through an SOA is shown in Fig. 2.2(a), where the SOA current I is assumed to be 130 mA. The optical input power is −10 dBm. The wavelength range for gain larger than 20 dB is evaluated to be around 50 nm. When the input optical power increases, the gain is saturated as shown in Fig. 2.2(b), where wavelength  is 1570 nm and the current I is 130 mA. The optical gain of 25 dB is expected for an input power less than −20 dBm. The gain saturation is a major factor that limits the channel number for multi-channel amplification. 22 21 Input power: -10dBm 30 130mA Gain (dB) Gain (dB) 20 19 18 17 16 1500 =1570nm 25 I=130mA 20 15 10 5 1520 1540 1560 1580 Optical wavelength (nm) (a) Gain as a function of wavelength 1600 0 -30 -20 -10 0 Input power (dBm) (b) Gain as a function of input power at  = 1570 nm Fig. 2.2. Calculated SOA gain at electric current of I = 130 mA. 52 10 Chapter 2. Multi-Point Temperature Sensor Consisting of AWG, SOA, and FBGs in Linear-Cavity Fiber Lasing System We consider simultaneous amplification for two signals at different wavelengths. When the wavelength interval decreases, the gain for the longer wavelength signal increases, whereas that for the shorter wavelength signal decreases due to FWM [23]. Fig. 2.3(a) shows the simulated gain for the longer (fixed at  = 1570 nm) and the shorter wavelength signals as a function of the wavelength interval. The input power is changed as a parameter. The induced FWM intensities are shown in Fig. 2.3(b). The generated FWM signals decrease with the wavelength interval. This gain deviation would complicate the diagnosis of its source as a result of the FWM effect or the interrogated environmental change. It would also affect the multi-channel lasing characteristics described in the following subsection. 5 20 0 FWM intensity (dBm) Gain (dB) 15 10 5 -10dBm_Long -10dBm_Short -5dBm_Long -5dBm_Short 0dBm_Long 0dBm_Short 0 -5 -10 0.0 0.5 1.0 1.5 2.0 Wavelength interval (nm) 2.5 -10dBm_Long -10dBm_Short -5dBm_Long -5dBm_Short 0dBm_Long 0dBm_Short -5 -10 -15 -20 -25 -30 3.0 0.0 (a) Gain at two wavelengths 0.5 1.0 1.5 2.0 Wavelength interval (nm) 2.5 3.0 (b) Generated FWM intensities Fig. 2.3. Simulated characteristics of two-wavelength amplification as a function of the wavelength interval, where input power is changed as a parameter. The SOA current is 130 mA. Finally, we investigate nonlinearity for amplification of multiple signals with equal interval. Fig. 2.4(a) shows output signal intensities as a function of optical frequency for cases of 4, 8, and 16 channels. The multiple signals have equal frequency separation. It is found that FWM signals are generated. The gain at the wavelengths of the input signals is shown in Fig. 2.4(b). The gain decreases with the input power due to gain saturation. The maximum number of multi-channel amplification is restricted depending on the input optical power. 25 5 4ch, 0dBm 8ch, 0dBm -15 8ch, -10dBm -20 8ch, -20dBm -25 16ch, 0dBm -30 16ch, -10dBm 16ch, -20dBm -35 -40 190 191 192 193 194 195 Optical frequency (THz) (a) Output intensities 4ch, -10dBm 15 4ch, -20dBm -10 4ch, 0dBm 20 4ch, -10dBm -5 Gain (dB) Output intensity (dBm) 0 4ch, -20dBm 8ch, 0dBm 10 8ch, -10dBm 5 8ch, -20dBm 16ch, 0dBm 0 16ch, -10dBm -5 -10 16ch, -20dBm 190 191 192 193 194 195 Optical frequency (THz) (b) Gain Fig. 2.4. Output intensities for multiple signal amplification. 53 Advances in Optics: Reviews. Book Series, Vol. 3 2.3.2. Analysis of Multi-Wavelength Lasing The proposed sensor system consists of a single SOA as the gain component for lasing. The number of channels at different wavelengths is limited due to the gain saturation in the SOA, as discussed in the previous section. The gain saturation, however, can be avoided by decreasing the incident optical power even if the channel number is large. In this section, we consider the lasing power of each channel by using a model shown in Fig. 2.5. FBG_1 FBG 1 MIRROR Loop mirror 1 AWG 2 AWG SOA FBG_N N Fig. 2.5. A model to calculate lasing powers for multi-wavelength operation. Each signal propagating in the left-hand direction at wavelength i has an optical power of Iai and Ibi, i = 1,…,N at the entrance and the exit of the SOA, respectively. The signal propagating in the right-hand direction after the SOA has an optical power of Ici. The gain of the SOA for a one-way path GS and a round-trip path GR are expressed as Ibi ∕ Iai and Ici ∕ Iai, respectively. Both are calculated numerically by using the SOA analysis model, which is described above, as shown in Fig. 2.6(a) where the optical wavelength of 1570 nm and the injection current of 130 mA are assumed. The reflectance at the left loop mirror, including the power-splitting loss for detection, is denoted by MIRROR. We assume MIRROR = 1 in this calculation. For weak incident power, the round-trip gain is higher due to the double path amplification. On the contrary, for stronger incident power, the oneway gain is higher due to the gain saturation. The total optical power propagating in the left-hand direction at the entrance of the SOA ISUM is given by ∑ (2.1) . The round-trip gain in the SOA, being denoted by GR(ISUM), depends on ISUM. The nonlinear effect of FWM is ignored in this analysis, and the gain for each wavelength is assumed to be identical. The transmittance of the AWG, including the insertion loss, is denoted by AWG; the reflectance at each FBG i is denoted by FBG_i. The total gain at i through a cavity is written as 54 _ . (2.2) Chapter 2. Multi-Point Temperature Sensor Consisting of AWG, SOA, and FBGs in Linear-Cavity Fiber Lasing System 40 Round-trip Gain (dB) 30 One-way 20 10 0 -10 -30 -20 -10 0 10 Optical input power (dBm) Total gain through linear cavity (dB) (a) Optical gain for one-way path and round-trip path through an SOA 10 N=1 N=2 N=4 N=8 N=16 N=32 N=1 N=2 N=4 N=8 N=16 N=32 5 0 -5 -10 -25 -20 -15 -10 -5 0  overall= ‐10dB ‐15dB 5 Optical power per channel (dBm) (b) Optical gain through the linear-cavity sensor as a function of optical intensity I 0 at the entrance of the SOA Fig. 2.6. Simulated results with the model of Fig. 2.5. The lasing condition in gain is given by (2.3) 1. We assume that Iai = I0, i = 1,…,N and the overall transmittance due to the losses is _ . (2.4) The total gain Gcavity is calculated as a function of I0 in Fig. 2.6(b), where the channel number N is varied as a parameter. It is found that the lasing power per channel decreases with N. As an example, the lasing power at the entrance of the SOA is around −8.5 dBm and −13.3 dBm for N = 8 with overall = −10 dB and −15 dB, respectively. The lasing power is increased by 3 dB by decreasing N to half, that is N = 4. 55 Advances in Optics: Reviews. Book Series, Vol. 3 2.4. Experimental Results 2.4.1. SOA Nonlinearity The nonlinearities in multi-channel amplification were experimentally verified with an SOA (Inphenix, IPSAD1502-214) using an experimental setup shown in Fig. 2.7. The measured wavelength dependence of the gain is shown in Fig. 2.8 (a), where a wavelength-variable laser source (Anritsu, MG9638A) and a spectrum analyzer (Anritsu, MS9710C) were used. At wavelength of 1570 nm, the maximum gain of 14.6 dB was obtained for input power of -20 dBm with injection current of 150 mA. When the input power was increased, the gain decreased due to gain saturation as shown in Fig. 2.8 (b), where the incident wavelength is 1570 nm. LD1 LD2 LDN 1 2 WDM signals Variable attenuator Combiner N SOA Optical spectrum analyzer Optical spectrum analyzer Fig. 2.7. Experimental setup for measuring multi-channel SOA amplification. 20 15 Input power ‐20 dBm =1570nm 10 0 Gain (dB) Gain (dB) 10 -10 100 mA 100 mA 0 120 mA -20 5 120 mA 150 mA 150 mA -30 1500 1510 1520 1530 1540 1550 1560 Optical wavelength (nm) 1570 (a) Gain as a function of wavelength 1580 -5 -50 -40 -30 -20 -10 Input Power (dBm) 0 10 (b) Gain as a function of input power Fig. 2.8. Measured results of optical gain characteristics for an SOA. Next, the nonlinearity due to FWM was measured for optical inputs at two wavelengths. The incident wavelengths, short and long, were varied as shown in Fig. 2.9 (a), where multi-channel laser sources (Anritsu, MU952601A, MU952602A) were used. The gain at the two wavelengths is plotted as a function of wavelength interval in (b). The gain at long was larger than that atshort as theoretically discussed in Fig. 2.3. The generated FWM intensities are shown in (c). The FWM at the shorter wavelength was larger than that at 56 Chapter 2. Multi-Point Temperature Sensor Consisting of AWG, SOA, and FBGs in Linear-Cavity Fiber Lasing System the longer wavelength. The increase of the gain at long in wider wavelength interval is considered to be caused by the wavelength dependence of the gain shown in Fig. 2.8 (a). 9 lambda1 long 1569 lambda2 short 1568 8 lambda1 long 1567 Gain (dB) Incident wavelength (nm) 1570 lambda2 short 1566 7 1565 1564 0 1 2 3 4 Wavelength interval (nm) 5 6 (a) Two incident wavelengths 6 0 1 2 3 4 Wavelength interval (nm) 5 6 (b) Gain at two wavelengths FWM intensity (dBm) -20 FWM1 FWM short -25 FWM long FWM2 -30 -35 -40 -45 0 1 2 3 4 Wavelength interval (nm) 5 6 (c) Generated FWM intensities Fig. 2.9. Measured characteristics of two-wavelength amplification as a function of the wavelength interval, where the SOA current is 150 mA. Finally, four-channel simultaneous amplification was demonstrated. The four input channels were set to have equal frequency interval of 200 GHz. The output intensities including FWM signals are shown in Fig. 2.10 (a). FWM signals were not observed for input of -20 dBm. The output intensities for larger incident power are restricted due to gain saturation. The gain at the four signal wavelengths are shown in (b). 2.4.2. ASE Spectrum and AWG Transmittance In order to assess the gain spectrum of an SOA (Inphenix, IPSAD1502-214) used in the experiment, we measured the amplified spontaneous emission (ASE) noise spectrum with no optical input at the SOA input port as shown by the dashed curve in Fig. 2.11. The current to the SOA was 120 mA. The wavelength resolution of the optical spectrum analyzer (Anritsu, MS9710C) was 0.05 nm. The maximum gain of the SOA is expected at around 1540 nm. When the ASE noise was input in the AWG (Accelink, 32-channel 57 Advances in Optics: Reviews. Book Series, Vol. 3 10 16 0 14 12 -10 0dBm -20 -20dBm -10dBm -30 Gain (dB) Output intensity (dBm) Athermal AWGMux or Demux), having 100-GHz frequency spacing, the output spectra at ports 1 and 2 were measured as shown by the solid and dotted curves, respectively, in Fig. 2.11. Although the output wavelength at port 1 of the AWG was designed to be 1535.2 nm, the other peaks were observed at 1488.0 nm and 1585.6 nm due to the periodical properties of the AWG. The frequency difference of the outputs between ports 1 and 2 was 100 GHz (≃0.8 nm in 1550-nm band). 10 8 6 4 -40 0dBm -10dBm -20dBm 2 0 1564 -50 1558 1560 1562 1564 1566 1568 1570 1572 1574 1576 1565 Optical wavelength (nm) 1566 1567 1568 1569 1570 Optical wavelength (nm) (a) Output intensities (b) Gain Fig. 2.10. Measured characteristics of multi-channel amplification as a function of optical wavelength. Input power is changed as a parameter. The SOA current is 150 mA. Optical Output (dBm) -20 ch1 ch2 ASE -30 -40 -50 -60 -70 1450 1500 1550 1600 1650 Wavelength (nm) Fig. 2.11. ASE spectrum from the SOA and the spectrum at channels 1 and 2 after passing through the AWG. 2.4.3. Multi-Wavelength Lasing We demonstrated multi-channel lasing with the setup shown in Fig. 2.12. The cavity consists of the SOA, the AWG, and fiber mirror reflectors at the end of the demultiplexed channels; a fiber loop mirror consists of a circulator, a 99:1 coupler, and a polarization controller. The AWG has a non-uniform insertion loss of 4.01–4.53 dB among ports with a 1-dB passband of 0.45 nm and 20 dB passband of 1.14 nm. The round-trip insertion loss 58 Chapter 2. Multi-Point Temperature Sensor Consisting of AWG, SOA, and FBGs in Linear-Cavity Fiber Lasing System of the AWG, the circulator, and the 99:1 coupler are approximately 8.6 dB, 1.47 dB, and 0.57 dB, respectively. Therefore, the total cavity loss is estimated to be around 10.64 dB. When the Nreflected output ports of the AWG were terminated by the fiber mirror reflectors, the lasing Nreflected wavelengths were observed at the measured port, as shown in Fig. 2.13. The number of the port terminated with the fiber mirror reflectors was changed from one to eight. The averaged power per channel decreased with the number of channels due to the gain saturation, as shown in Fig. 2.13 (b), where the lasing power was −32.6 dBm for Nreflected = 8. On the other hand, the lasing power was doubled by decreasing the lasing channel number to half, as mentioned in the simulated result shown in Fig. 2.6 (b). 1 2 Pol. controller Circulator 1:32 AWG SOA 99:1 coupler Fiber mirror reflector  Nrefl ected N Optical spectrum analyzer Fig. 2.12. Experimental setup for multi-channel lasing. -20 -40 1 -60 2 -80 3 1584 4 1586 5 1588 6 1590 7 1592 8 (a) Lasing spectrum Output Average Power (dBm) Output Power (dBm) -22 -24 -26 -28 -30 -32 -34 11 2 3 4 Number of Channels 5 6 7 8 (b) Output averaged power per channel for multi-channel lasing. Fig. 2.13. Measured results when the number of reflected channels is changed from 1 to 8. Next, we demonstrated the difference in multi-channel lasing between SOA and EDFA as the amplifier. The amplifier was placed in the left-hand side loop mirror as shown in Fig. 2.14. In this setup, the optical signal passes through the amplifier only in a single path. The eight output ports from the AWG are terminated with fiber reflector mirrors. The lasing output was observed as shown in Fig. 2.15. In the case of EDFA, lasing 59 Advances in Optics: Reviews. Book Series, Vol. 3 spectrum at one channel was observed. This is due to the homogeneous broadening in amplification with the EDFA. Therefore, SOA is superior in multi-channel lasing operation. 1 2 Pol. controller SOA or EDFA Circulator 1:32 AWG 99:1 coupler  Nrefl ected Fiber mirror reflector N Optical spectrum analyzer Fig. 2.14. Experimental setup to demonstrate the difference in lasing between SOA and EDFA. -30 SOA Optical power (dBm) Optical power (dBm) -30 -40 -50 -60 1540 1542 1544 Wavelength (nm) 1546 1548 (a) SOA EDFA -40 -50 -60 1540 1542 1544 Wavelength (nm) 1546 1548 (b) EDFA Fig. 2.15. Lasing output spectra for amplifier of (a) SOA and (b) EDFA. 2.4.4. Two-Wavelength Lasing with FBGs To verify multi-channel sensing, we demonstrated two-channel temperature sensing with two FBGs. The FBGs [channel 6: grating length (L) = 10 mm,  = 1539.872 nm, −3 dB bandwidth (BW at −3 dB) = 0.227 nm, reflectivity (R) = 97.04 %, sidelobe suppression rate (SLSR) = 15.40 dB; channel 9: L = 10 mm,  = 1542.122 nm, BW at −3 dB = 0.236 nm, R = 97.57 %, SLSR = 23.02 dB] were connected at ports 6 and 9 through variable attenuators, as shown in Fig. 2.16. Optical lasing spectra measured at the detection port at room temperature are shown in Fig. 2.17. The lasing characteristics at two channels were measured by varying the attenuator in channel 6, as shown in Fig. 2.18. When the transmittance decreased with the increase of the attenuation in channel 6, the lasing output at channel 6 decreased. On the contrary, the lasing output at channel 9 decreased when the transmittance of channel 6 increased. 60 Chapter 2. Multi-Point Temperature Sensor Consisting of AWG, SOA, and FBGs in Linear-Cavity Fiber Lasing System Simultaneous lasing with equal power was obtained with the transmittance of 0.68. Thus, the attenuation in both channels should be adjusted to achieve an equal lasing output power. The AWG used in the experiment had Gaussian-like filtering characteristics, as shown in Fig. 2.19, where the transmittance at channels from 6 to 9 were plotted. Variable attenuator Pol. controller Circulator Water bath Ch.6 1:32 AWG SOA FBG Ch.9 99:1 coupler Optical spectrum analyzer Hot plate Fig. 2.16. Experimental setup for two-channel temperature sensing. Output Power (dBm) -20 Ch.6 -30 Ch.9 -40 -50 -60 -70 1536 1538 1540 Wavelength (nm) 1542 1544 Fig. 2.17. Lasing spectra with two FBGs at room temperature. Optical Output Power (dBm) -10 -20 -30 Ch.6 -40 Ch.9 -50 -60 0 0.2 0.4 0.6 0.8 Transmission Ratio of Attenuator in Channel 6 1 Fig. 2.18. Output power at two channels with varying the attenuator in channel 6. 61 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 2.19. Transmission characteristics of the AWG at ports 6 to 9. 2.4.5. Simultaneous Temperature Sensing The temperature at the FBG in channel 9 was measured as shown in Fig. 2.20 (a), where the FBG in channel 6 was removed. The output power as a function of wavelength obtained from Fig. 2.20 (a) is plotted in Fig. 2.20 (b). Since the transmittance of the AWG in channel 9 started to decrease at the wavelength corresponding to a temperature higher than around 60 °C, the output power decreased. The ratio of the wavelength change to the temperature change was around 6.4 pm/deg. A similar result was also confirmed by using the FBG in channel 6 as shown in Fig. 2.21. The ratio of the wavelength change to the temperature change was also around 6.4 pm/deg. Note that the reason that the measured wavelength was changed almost in steps of 0.025 nm was due to the wavelength resolution of the optical spectrum analyzer. 1541.85 1541.75 -25 1541.70 Output Power Wavelength -30 1541.65 1541.60 1541.55 -35 1541.50 Channel 9 -40 10 30 50 70 Temperature (degree) 1541.45 90 1541.40 (a) Output power and the wavelength vs. temperature Optical Output Power (dBm) -20 1541.80 Wavelength (nm) Optical Output Power (dBm) -20 -25 -30 -35 -40 1541.4 1541.5 1541.6 1541.7 Wavelength (nm) 1541.8 1541.9 (b) Output power vs. wavelength Fig. 2.20. Temperature measurement result by using an FBG only in channel 9. Fig. 2.22 shows an example of simultaneous temperature measurement with two FBGs. The temperature of the FBG in channel 9 was varied while the FBG in channel 6 was kept at room temperature. Since the loss in channel 9 increased as the temperature increased more than around 50°, the lasing power in channel 9 decreased gradually with the 62 Chapter 2. Multi-Point Temperature Sensor Consisting of AWG, SOA, and FBGs in Linear-Cavity Fiber Lasing System temperature, whereas the power in channel 6 increased by about 3 dB with the temperature. A similar result of two-channel measurement also was confirmed when the temperature at the FBG in channel 6 is varied as shown in Fig. 2.23. -20 1539.50 1539.40 Output Power 1539.35 Wavelength -30 1539.30 1539.25 -35 1539.20 Channel 6 -40 10 20 30 40 50 Temperature (degree) 60 Wavelength (nm) 1539.45 -25 70 Optical Output Power (dBm) Optical Output Power (dBm) -20 -25 -30 -35 -40 1539.1 1539.15 1539.2 1539.3 1539.4 1539.5 Wavelength (nm) (a) Output power and the wavelength vs. temperature (b) Output power vs. wavelength -25 1539.45 1539.40 -30 Output Power Wavelength -35 1539.35 1539.30 -40 1539.25 -45 -50 1539.20 Channel 6 10 20 30 40 50 Temperature in Channel 9 (degree) 60 1539.15 Channel 6 (fixed temp.) Channel 9 (varied temp.) -20 1541.80 -25 1541.75 1541.70 -30 Output Power Wavelength -35 1541.65 1541.60 1541.55 -40 1541.50 -45 -50 Channel 9 10 (a) Channel 6 20 30 40 50 Temperature in Channel 9 (degree) Wavelength (nm) 1539.50 Wavelength (nm) -20 Optical Output Power (dBm) Optical Output Power (dBm) Fig. 2.21. Temperature measurement result by using an FBG only in channel 6. 1541.45 60 1541.40 (b) Channel 9 Fig. 2.22. Simultaneous temperature measurement result using FBGs in channels 6 and 9, where the temperature at channel 6 was kept fixed at room temperature and that at channel 9 was varied. 2.4.6. Increase of the Temperature Sensing Range The measured temperature range in the experimental results is mainly restricted by the facility limitations. The maximum operating temperature of the utilized FBGs is 80 °C due to the coating material of their fibers. Besides, the AWG has the non-flat-top passband profile for each wavelength channel, as shown in Fig. 2.19. By employing an FBG with wider operating temperature and a flat-top AWG having 200 GHz bandwidth with suitable central wavelengths, for example, the measured temperature range can be more than doubled. 63 Advances in Optics: Reviews. Book Series, Vol. 3 Output Power Wavelength -30 1539.40 1539.35 -35 1539.30 -40 1539.25 -45 -50 1539.20 Channel 6 10 20 30 40 Temperature (degree) (a) Channel 6 50 1539.15 Optical Output Power (dBm) 1539.45 -25 Channel 6 (varied temp.) Channel 9 (fixed temp.) -20 1541.80 -25 1541.75 Output Power -30 1541.70 Wavelength -35 1541.65 1541.60 -40 1541.55 1541.50 -45 -50 1541.45 Channel 9 10 20 30 40 Temperature (degree) Wavelength (nm) 1539.50 Wavelength (nm) Optical Output Power (dBm) -20 50 1541.40 (b) Channel 9 Fig. 2.23. Simultaneous temperature measurement result using FBGs in channels 6 and 9, where the temperature at channel 9 was kept fixed at room temperature and that at channel 6 was varied. Advantages of the proposed system configuration are that it can provide easy equalization of power of each lasing wavelength channel as well as resilience to fiber failures by exploiting the star configuration of the sensing elements. Moreover, the sensing physical parameters can be simply measured by detecting the wavelength shift of the lasing lines. 2.5. Conclusion Multi-channel amplification with an SOA was investigated for the proposed linear-cavity sensing system. The lasing condition for multi-channel operation was clarified numerically by considering the gain saturation in the SOA. The gain saturation limits the maximum number of channels and the lasing power per channel. The multi-wavelength lasing was experimentally demonstrated up to eight channels, where fiber mirror reflectors were employed instead of FBGs. The lasing power was doubled by decreasing the lasing channel number to half and by decreasing the number of the reflectors. This result agrees well with theoretical analysis. To demonstrate multi-point sensing, two FBGs were employed at two ports of the AWG. Since the filtering response of the AWG was not in a flat-top profile, the lasing power decreased when the lasing wavelength shifted to the edge of the AWG channel. It was found that equalization of loss in each channel was indispensable for multiple simultaneous sensing. By adopting a flat-top AWG, stable multi-point sensing is expected over a wide range of temperatures. Acknowledgements This research was supported in part by Collaborative Research Project between National Taiwan University of Science and Technology and Tokushima University. 64 Chapter 2. Multi-Point Temperature Sensor Consisting of AWG, SOA, and FBGs in Linear-Cavity Fiber Lasing System References [1]. S. Kim, J. Kwon, S. Kim, B. Lee, Multiplexed strain sensor using fiber grating-tuned fiber laser with a semiconductor optical amplifier, IEEE Photon. Technol. Lett., Vol. 13, Issue 4, April 2001, pp. 350-351. [2]. S.-K. Liaw, Y.-W. Lee, H.-W. Huang, W.-F. Wu, Multi-wavelength linear-cavity SOA-based laser array for long-haul sensing, IEEE Sensor J., Vol. 15, Issue 6, June 2015, pp. 3353-3358. [3]. C.-S. Shin, S.-K. Liaw, S.-W. Yang, Post-impact fatigue damage monitoring using fiber Bragg grating sensors, Sensors, Vol. 14, Issue 3, March 2014, pp. 4144-4153. [4]. S. Diaz, D. Leandro, M. Lopez-Amo, Stable multiwavelength erbium fiber ring laser with optical feedback for remote sensing, IEEE/OSA J. Lightw. Technol., Vol. 33, Issue 12, Jun. 2015, pp. 2439-2444. [5]. M. Bravo, V. DeMiguel-Soto, A. Ortigosa, M. Lopez-Amo, Fully switchable multiwavelength fiber lasers based on random distributed feedback for sensors interrogation, IEEE/OSA J. Lightw. Technol., Vol. 33, Issue 12, June 2015, pp. 2598-2604. [6]. H. D. Lee, G. H. Kim, T. J. Eom, M. Y. Jeong, and C. -S. Kim, Linearized wavelength interrogation system of fiber Bragg grating strain sensor based on wavelength-swept active mode locking fiber laser, IEEE/OSA J. Lightw. Technol., Vol. 33, Issue 12, June 2015, pp. 2617-2622. [7]. S. Pathak, M. Vanslembrouck, P. Dumon, D. Van Thourhout, W. Bogaerts, Optimized silicon AWG with flattened spectral response using an MMI aperture, IEEE/OSA J. Lightw. Technol., Vol. 31, Issue 1, January 2013, pp. 87-93. [8]. K. Okamoto, Fundamentals of Optical Waveguides, 2nd Ed., Elsevier, Amsterdam, Boston, 2006. [9]. A. Bellemare, M. Karásek, M. Rochellte, S. LaRochelle, M. Têtu, Room temperature multifrequency erbium-doped fiber lasers anchored on the ITU frequency grid, IEEE/OSA J. Lightw. Technol., Vol. 18, Issue 6, June 2000, pp. 825-831. [10]. S. Diaz, A. B. Socorro, R. M. Manuel, R. Fernandez, I. Monasterio, Stable multi-wavelength fiber lasers for temperature measurements using an optical loop mirror, Appl. Opt., Vol. 55, Issue 29, Oct. 2016, pp. 8385-8389. [11]. H.-Y. Hsu, Y.-L. Yu, C.-C. Tsai, S.-K. Liaw, N.-K. Chen, W.-F. Wu, Semiconductor optical amplifier based multi-wavelength lasers in 0.4 nm channel spacing, Laser Physics, Vol. 23, Issue 9, August 2013, pp. 95-104. [12]. N. Pleros, C. Bintjas, M. Kalyvas, G. Theophilopoulos, K. Yiannopoulos, S. Sygletos, H. Avramopoulos, Multiwavelength and power equalized SOA laser sources, IEEE Photon. Technol. Lett., Vol. 14, Issue 5, May 2002, pp. 693-695. [13]. M. Fernandez-Vallejo, D. Ardanaz, M. Lopez-Amo, Optimization of the available spectrum of a WDM sensors network using a modelocked laser, IEEE/OSA J. Lightwave Technol., Vol. 33, Issue 22, November 2015, pp. 4627-4631. [14]. A. Triana, D. Pastor, M. Varon, Overlap-proof fiber Bragg grating sensing system using spectral encoding, IEEE Photon. Technol. Lett., Vol. 28, Issue 7, Apr. 2016, pp. 744-747. [15]. R. Li, C. Tian, Y. Zhang, J. Zhang, X. Chen, S. Liu, Simultaneous measurement of force and temperature by a multi-position fiber loop ringdown sensor array, IEEE/OSA J. Lightwave Technol., Issue 33, Issue 17, September 2015, pp. 3607-3612. [16]. R. J. Manning, A. D. Ellis, A. J. Poustie, K. J. Blow, Semiconductor laser amplifiers for ultrafast all-optical signal processing, J. Opt. Soc. Am. B, Vol. 14, Issue 11, November 1997, pp. 3204-3216. [17]. K. Tajima, S. Nakamura, Y. Ueno, Ultrafast all-optical signal processing with symmetric Mach-Zehnder type all-optical switches, Opt. Quantum Electron., Vol. 33, Issue 7-10, July 2001, pp. 875-897. 65 Advances in Optics: Reviews. Book Series, Vol. 3 [18]. V. Lal, M. L. Masanovic, J. A. Summers, G. Fish, D. J. Blumenthal, Monolithic wavelength converters for high-speed packet-switched optical networks, IEEE J. of Sel. Topics in Quantum. Electron., Vol. 13, Issue 1, January-February 2007, pp. 49-57. [19]. H. Kishikawa, M. Okada, K. Takahashi, P.-J. Chen, N. Goto, Y.-L. Yu, S.-K. Liaw, MultiPoint temperature sensing using linear-cavity lasing system, Applied Optics, Vol. 56, Issue 11, April 2017, pp. 3206-3212. [20]. M. Fernandez-Vallejo, S. Díaz, R. A. Perez-Herrera, D. Passaro, S. Selleri, M. A. Quintela, J. M. López Higuera, M. Lopez-Amo, Resilient long-distance sensor system using a multiwavelength Raman laser, Meas. Sci. Technol., Vol. 21, Issue 9, September 2010, 094017. [21]. K. Inoue, Crosstalk and its power penalty in multichannel transmission due to gain saturation in a semiconductor laser amplifier, IEEE/OSA J. Lightw. Technol., Vol. 7, Issue 7, July 1989, pp. 1118-1124. [22]. K. Inoue, T. Mukai, T. Saitoh, Nearly degenerate four-wave mixing in a traveling-wave semiconductor laser amplifier, Appl. Phys. Lett., Vol. 51, Issue 14, October 1987, pp. 1051-1053. [23]. G. P. Agrawal, I. M. Habbab, Effect of four-wave mixing on multichannel amplification in semiconductor laser amplifiers, IEEE J. Quant. Electron., Vol. 26, Issue 3, March 1990, pp. 501-505. [24]. K. Kikuchi, M. Kakui, C. -E. Zah, T. -P. Lee, Observation of highly nondegenerate four-wave mixing in 1.5 μm traveling-wave semiconductor optical amplifiers and estimation of nonlinear gain coefficient, IEEE J. Quant. Electron., Vol. 28, Issue 1, January 1992, pp. 151-156. [25]. J. Leuthold, M. Mayer, J. Eckner, G. Guekos, H. Melchior, C. Zellweger, Material gain of bulk 1.55 μm InGaAsP/InP semiconductor optical amplifiers approximated by a polynomial model, J. Appl. Phys., Vol. 87, Issue 1, January 2000, pp. 618-620. [26]. M. J. Connelly, Wideband semiconductor optical amplifier steady-state numerical model, IEEE J. Quant. Electron., Vol. 37, Issue 3, March 2001, pp. 439-447. Appendix Equations Used in SOA Analysis The analysis composed of the traveling-wave equations for signal fields and spontaneous emission, carrier-density rate equation, and material gain modeling based on [26]. at The signals are assumed to be composed Ns components having power of frequencies νk, k = 1,..., Ns. The fields of each incident signal in SOA is assumed to be sum of and propagating to +z and −z directions, respectively. Photon rate per second is given by by . The traveling-wave equations for Γ Γ 66 , , and , are given (2.A1) Chapter 2. Multi-Point Temperature Sensor Consisting of AWG, SOA, and FBGs in Linear-Cavity Fiber Lasing System where βk is the propagation constant,  is optical confinement factor,  is loss coefficient, gm is material gain coefficient, and n is conduction band carrier density. To model spontaneous emission, noise photons are assumed to exist only at discrete frequencies νj, j = 0,..., Nm − 1, corresponding to integer multiples of cavity resonance. and are Traveling-wave equations for spontaneous emission of photon rates, given by Γ , Γ , , where Rsp is the spontaneous emitted noise coupled into , and (2.A2) , . The rate equation for carrier density n is given by ∑ ∑ , , (2.A3) , where Kj equals to unity for zero facet reflectivity, I is the current, d is the active region thickness, and W is the active region width. The recombination rate R(n) is given by , (2.A4) where Rrad and Rnrad are the radiative and non-radiative carrier recombination rates, respectively. These rates can be expressed as , . (2.A5) , where Arad and Brad are the linear and bimolecular radiation recombination coefficients, Anrad is the linear non-radiative recombination coefficient due to traps in the semiconductor material, Bnrad is the non-radiative bimolecular recombination, Caug is the Auger recombination coefficient, and Dleak is the recombination due to leakage effects. Material gain coefficient gm can be modeled as , √ ⁄ , (2.A6) where c is the speed of light in vacuum, n1 is the active region refractive index,  is the radiative carrier recombination lifetime, h is the Planck’s constant, me and mhh are the conductive band (CB) electron and valence band (VB) heavy hole effective masses, respectively, Eg is the bandgap energy, and fc and fv are the Fermi-Dirac distributions in CB and VB. 67 Chapter 3. Review of Fabry-Pérot Fiber Sensors Chapter 3 Review of Fabry-Pérot Fiber Sensors Yundong Zhang, Li Zhao, Fuxing Zhu, Ying Guo, Ping Yuan1 3.1. Introduction Fabry-Pérot interferometers (FPIs) have been widely used as optical fiber sensors in health monitoring of composite materials, civil engineering structures, space aircrafts, and medicine, etc. Owing to their advantages, such as compactness, simple configuration, small size, high sensitivity, fast responses, good repeatability, etc., which makes them suitable for the detection of physical and chemical parameters such as temperature, strain, refractive index (RI), transverse load, gas phase concentrations and so on. The fiber FP sensor is based on the multi-beam interference principle to detect the changes of the external parameters. The FP cavity has two parallel separated reflective surfaces which can partially reflect the lead-in optical signals. The beams reflected by the surfaces will interfere when they come back into the lead-in fiber. When the FP cavity is subjected to external parameters (strain, deformation, displacement, temperature, refractive index), there would be a phase difference change between the two reflected beams, resulting in a shift of reflection spectrum. Thus, the change of external parameters can be known. With the aim of forming FP cavities in optical fibers, two parallel separated interfaces which can partially reflect the lead-in optical signals are required to be made in optical fibers. For this purpose, considerable techniques have been developed, such as the earlier manual bonding techniques, inserting a section of hollow core fiber or hollow core photonic crystal fiber between two sections of single-mode fibers (SMFs), splicing different fibers in series, film coating techniques, direct micromachining using focused femtosecond laser beam, and using chemical etching method to form a microcavity in fiber. Different shapes of cavities have been developed according to the different fabricating method. Such as ellipsoidal cavities, spherical shape, cylindrical shape and so on. Different FP cavity structures correspond to different sensing sensitivity. How to simplify the fabricating method as well as improve the sensitivity of the FPIs is a long- Yundong Zhang Harbin Institute of Technology, China 69 Advances in Optics: Reviews. Book Series, Vol. 3 term goal. What’s more, realizing the simultaneous measurement for different parameters in a single FPI is also a focus of the study. In this chapter, we present a comprehensive overview of the FP fiber sensor technology, classified according to their applications, including FP fiber sensors for strain measurement, temperature measurement, refractive index (RI) measurement and so on. The fundamental principles of the FP fiber sensors are detailed. Each application is reviewed in turn, key recent researches that contributions to the developing of the FP fiber sensors are highlighted and discussed. Finally, we give a forward-looking perspective and discuss the outlook of the FP based fiber sensors considering how can FP fiber sensors step forward. 3.2. Basic Theory The intensity of the interference fringe of air cavity FPI in the reflection spectrum can be expressed as I =I1 +I 2 +2 I1 I 2 cos  , (3.1) where I1 and I2 are the intensities of light reflected at the two cavity interfaces, respectively, and φ is the phase different shift between the two reflected lights.  4  nL , (3.2) where λ is the wavelength of the incident light, n is the RI of the medium in cavity, is the FP cavity length. When    2m  1  , m is an integer, the minimum interference intensity occurs: m  4nL . 2m  1 (3.3) Free spectral range (FSR), namely the fringe spacing or period of the FPI spectrum at λ, can be given by FSR  2 2nL , (3.4) where λ is the wavelength of light, n is the refractive index (RI) of the medium inside the FP cavity. In experiment, L is not convenient for direct measurement, so it is usually calculated by FSR. The interference fringe contrast of the reflecting spectrum is V: 70 Chapter 3. Review of Fabry-Pérot Fiber Sensors V I max  I min , I max  I min (3.5) where Imax is the maximum of interference spectral intensity, Imin is the minimum of interference spectral intensity. 3.3. Applications 3.3.1. Strain Sensing The intrinsic optical fiber sensors based on FP cavities prove to be suitable for strain sensing and have been used successfully for health monitoring of the large civil engineering structures, composite materials, spacecraft, and so on, due to their distinct advantages such as low cost, compactness, high sensitivity and low temperature cross sensitivity. Regarding the optical fiber FPIs, varieties of FP cavities with different shapes were designed and fabricated, for instance, the prolate spheroidal shape FP cavities [1-3], quasi-spherical FP cavities [4], and FP cavities based on a tube [5-10]. Developing FPIs which have a high strain sensitivity and a low thermal sensitivity is the goal people pursuit. In this section, an overview of the accomplishments in the field of optical fiber FPIs for strain sensing is reported. According to the shape of the FP cavity, we classify the optical fiber FPIs into 3 types: spheroidal shape FP cavities, quasispherical FP cavities, and cylindrical FP cavities. 3.3.1.1. Prolate Spheroidal FP Cavity Strain Sensor What kind of fiber FP cavity shape can help to improve the strain sensitivity is of utmost importance. Besides, the relationship between the strain sensitivity and the length of the FP cavity is another key point to research. But For a long time, researchers didn’t have a systematic investigation to the above questions. Some authors even assume that the strain sensitivity of a FPI is independent of the cavity size and it can be enhanced solely by choosing longer wavelengths. Till 2012, F. C. Favero et al. investigated the relationship between the FPIs cavity shape and the strain sensitivity, and demonstrated that the strain sensitivity of FPIs with spheroidal cavities can be controlled through the dimensions of the spheroid [1]. When FPI cavities like those described in Fig. 3.1 is subjected to axial strain, it will experience both axial deformation  d , and transversal deformation  r , the shift of the interference pattern of a FPI with quasi-spherical or spheroidal cavity is proportional to  r . That is to say, the strain sensitivity is proportional to  r . d d  r 3  E   R2 2  r  1    .  d 4  K   r 2 3  d (3.6) 71 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 3.1. Diagram of a FPI with (a) spherical and (b) prolate spheroidal air cavity. In Eq. (3.6), E is the Young’s modulus, K is the bulk modulus, d is the polar radius and r is the equator radius. Plot the term  r  d as a function of the polar radius d, considering both the quasi- spherical cavity(r ≈ d) and the prolate spheroidal cavity. We can easily find that when the cavity has spherical shape, the term  r  d decrease as the cavity becomes smaller. However, in Fig. 3.2(b), the result is quite opposite to the former situation, when the cavity has spheroidal shape,  r  d increase as the cavity becomes smaller. For the same spheroid diameter, the term  r  d increases with the increase of r. That means the prolate spheroidal FP cavities corresponds to higher strain sensitivity. The theoretical analysis here provides us a way to enhance the sensitivity for different kind of FPIs. Fig. 3.2. (a) Theoretical value of δr/δd as a function of d of a quasi-spherical cavity; (b) a prolate spheroidal cavity for different values of r. In all cases r > d. By fusion splicing SMF to PCF, prolate spheroidal FP cavities are fabricated, as shown in Fig. 3.3. To verify whether the experimental results fit well with the theoretical analysis, FPIs with cavities of sizes of 10  60 µm and 29  40 µm were subjected to the strain tests. The experimental results are shown in Fig. 3.4. The strain sensitivity of the FPI with large d and small r was found to be 3.5 pm/µε, while the prolate one with small d and large r was 72 Chapter 3. Review of Fabry-Pérot Fiber Sensors 10.3 pm/µε. Thus, the results fit very well with the theoretical analysis of spheroidal cavities presented above. Fig. 3.3. Schematic diagram of the interrogation set-up. Fig. 3.4. (a) Reflection spectra at different strains observed in a FPI whose cavity had prolate spheroidal shape and dimensions of 10  60 µm; (b) Interference pattern shifts vs strain observed in FPIs with cavity size of 10  60 µm (dots) and cavity size of 29  40 µm (stars). The presented FPIs also exhibit a low temperature sensitivity (~0.95 pm/°C). This means that for the fiber FP sensor with cavity size of 10  60 µm, the temperature-induced strain error is only 0.09 µε/°C. In 2014, Yiping Wang et al. demonstrated a simple technique to create prolate spheroidal FP cavity in fiber directly [2]. They coated a liquid film on the hemispherical end surface of the SMFs, when the liquid films were overlapped, arc discharge was done, then two fiber ends were spliced with each other and a prolate spheroidal air bubble was created in the spliced joint. Using the above method, prolate spheroidal air bubbles with different cavity lengths of 79, 70, 58, 54, and 46 μm were achieved, as shown in Figs. 3.5 (a)-3.5 (e). The corresponding reflection spectra of the air-cavity-based FPIs are illustrated in Figs. 3.5(f)3.5(j), respectively, in which the FSR of the interference fringes is 14.9, 16.8, 20.8, 22.8, and 26.4 nm, respectively. As the cavity length decrease from 79 to 46 μm, the 73 Advances in Optics: Reviews. Book Series, Vol. 3 corresponding FSR around 1530 nm increases gradually from 14.9 to 26.4 nm, which is in accordance well with the FSR theory. Fig. 3.5. (a), (b), (c), (d), and (e) Microscope images of the created prolate spheroidal air bubble with a cavity length of 79, 70, 58, 54, and 46 μm, respectively; (f)–(j) the corresponding reflection spectra of the prolate spheroidal air-cavity-based FPI. (ER, extinction ratio). Then the response of the FPI with different cavity lengths to the applied tensile strain was investigated. The experiment results show that the shorter length the FP cavity is, a higher strain sensitivity the FPI has. Hence the strain sensitivity of the air-cavity-based FPI fiber sensor can be enhanced by means of shortening the cavity length. The strain sensitivity of the air-cavity-based FPI was enhanced from 2.9 pm/με to 6.0 pm/με while the cavity length was shortened from 79 to 46 μm. This is of great consistence with the conclusion proposed in [1]. A FPI sample with cavity length of 46 μm was used to investigate the temperature responses. The dip wavelength in the reflection spectrum was shifted toward a longer wavelength with a low temperature sensitivity of 1.1 pm/°C. In 2012, Yun-jiang Rao et al. reported an easy fabricated and low-cost fiber optical FPI strain sensor whose cavity is a microscopic prolate spheroidal air bubble [3]. The bubble is formed by fusion splicing together two sections of single-mode fibers (SMFs) with cleaved flat tip and arc fusion induced hemispherical tip, respectively. The schematic of the sensor system with highlighting detail structure of the proposed sensor head is shown in Fig. 3.6. The effect of strain variation on the spectrum shifts were experimentally investigated in two FPIs, both with cavity lengths ~91 μm. The experimental results show that sensor 1 has a higher strain sensitivity of 4.2 pm/με and sensor 2 has a relatively lower strain sensitivity of 4.0 pm/με. The authors also stated that since the refractive index of the medium inside the bubble does not change with strain, the spectrum shift depends solely 74 Chapter 3. Review of Fabry-Pérot Fiber Sensors on the length of the cavity L. If the FPI sensors have larger bubble diameter (cavity length) L and thinner bubble outer cladding than theirs, they are more susceptible to the axial stress, and thus have higher strain sensitivity. But for this point of view, the authors didn’t give the experimental data to support. Considering the FP cavity is formed by a prolate spheroidal air bubble, this kind of FPI maybe more susceptible to the axial stress if they have shorter bubble diameter (cavity length). And it’s certain that thinner bubble outer cladding help to enhance the strain sensitivity. Fig. 3.6. Schematic of the sensor system (top) with highlighting detail structure of the proposed sensor head. The inset in the middle is the microscope photograph of a fabricated microbubble sensor. The temperature tests show that sensitivity of this FPI is only 0.828 pm/°C for sensor 1 and 0.868 pm/°C for sensor 2 in a range of temperature from 100 °C to 950 °C. As for the fore-mentioned three works, all the FPI strain sensors are based on the prolate spheroidal FP air cavities, during the strain measure experiments, the smaller FP cavity lengths corresponds to higher strain sensitivity, conversely, as the FP cavity length become larger, the sensitivity of the FPI strain sensor decreases gradually. Thus, for the prolate spheroidal FP cavity based fiber strain sensor, the sensitivity can be increased by shortening their cavity length. In general, they all show high strain sensitivity and low temperature cross sensitivity, which may become suitable candidates as strain sensors to be used in the practical applications. 3.3.1.2. Spherical FP Cavity Strain Sensor In this section, let’s take a look at the situation when the FP cavity is a spherical one. In 2009, Joel Villatoro et al. reported a FP strain sensor based on index guiding photonic crystal fiber, whose cavity is a spherical air bubble [4]. When splicing is done, the micro air holes in the cladding collapse over a few hundred micrometers and are trapped inside the PCF, thus, forming a spherical micro bubble, the diameter of the bubbles was in a range of 20–25 μm. The cross section of the PCF and the diagram of the interrogation setup are shown in Fig. 3.7. 75 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 3.7. Diagram of the interrogation setup highlighting the zone of the splice. The 25 μm diameter spherical air cavity was subjected to strain ranging from 0 to 5000 με, and the observation was carried out at wavelength 1290 ± 40 nm. The 58 μm diameter bubble was also subjected to the same strain range, but the shift was measured at wavelength 1550 ± 30 nm. The strain sensitivity of the interferometer with the smaller cavity was 0.62 pm/με, while that of the interferometer with the larger cavity was 2.7 pm/με. So, for the spherical FP cavity, the strain sensitivity increases when the FPI has larger FP cavity length. The experimental result is opposite to the former situation when the cavity has spheroidal shape. These two types are in agreement with the theory proposed in [1]. 3.3.1.3. Cylindrical FP Cavity Strain Sensor Apart from the aforementioned FP cavities, there are also some FP cavities formed by inserting a section of hollow tube or hollow core PCF into SMFs, or just using the fs laser to fabricate a notch on the fiber. FP cavities formed by this way usually have cylindrical or rectangular shape. In 2016, Yong Zhao et al. demonstrated a kind of in-fiber rectangular air FP strain sensor with different cavity lengths [5]. They show that the shorter length the cavity has, the higher strain sensitivity the sensor obtains. The strain sensitivity of in-fiber rectangular air FP sensor with a cavity length of 35 µm can be up to 2.23 pm/με. Fig. 3.8 shows that the in-fiber air FP cavity, which is formed by splicing a section of hollow-core fiber (HCF) with a diameter of 50/125 µm (hole/outer clad) between SMFs. Three FP cavities fabricated are show in Fig. 3.8 (c), in which the lengths of the middle HCF segments are 35 µm, 50 µm, 100 µm, respectively. Through formula derivation, the sensitivity of optical microcavity tension sensor can be can be expressed as follow: K 76 dip F =dip  L , LF (3.7) Chapter 3. Review of Fabry-Pérot Fiber Sensors where K is the sensitivity of sensor, λ is the dip of the reflection spectrum, F is the axial tension, and L is the microcavity length. By formula (3.7), when F is the constant, the sensitivity of optical microcavity tension sensor mainly depends on ΔL/L. That means higher ΔL/L value corresponds to higher strain sensitivity. Fig. 3.8. Preparation process of in-fiber air FP cavity: (a) Welding process of SMF-HCF; (b) Cutting process base on HFCP, and (c) Physical map of FP platform. Fig. 3.9 shows the relationship between L and ΔL/L. When cavity length L varies, ΔL increases with the increase of cavity length L, but ΔL/L reduces with the increase of cavity length L. According to the formula (3.7), what really matters is the ΔL/L value, thus, high strain sensitivity can be obtained by reducing the cavity length L (namely, increasing the ΔL/L value). Fig. 3.9. The relationship diagram between cavity length L and ΔL/L. Fig. 3.10 shows the shift observed as a function of strain in two samples. The strain sensitivity of the 35 µm FP cavity is measured to be 2.23 pm/με, while that the strain sensitivity of the 75 µm FP cavity is 1.12 pm/με. The experimental results and the theory are in agreement, the shorter length the cavity has, the higher strain sensitivity the sensor obtains. 77 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 3.10. Reflection spectrum and strain sensitivity of in-fiber rectangular air FP sensor with different cavity length 35 µm and 75 µm. The temperature sensing properties of the in-fiber rectangular air FP sensor with cavity length of 35 µm has a low temperature sensitivity of only 0.32 pm/°C. According to the strain and temperature sensitivity of the interference fringe, the temperature-induced strain measurement error is about 0.15 με/°C. Therefore, the temperature disturbance can be ignored in the condition of small temperature fluctuations. In 2012, Marta S. Ferreira et al. proposed a FP strain sensor based on hollow-core ring photonic crystal fiber [6]. The FP cavity was constituted by splicing a small section of hollow-core ring photonic crystal fiber (HCR PCF) between two sections of SMFs. Through numerical simulation, the authors demonstrate that as the FP cavity length decreases, there is a significant increase in the normalized strain coefficient, as shown in Fig. 3.11. As the FP cavity approaches 40 μm, and decreases furthermore, there is a significant increase on the strain sensitivity. The experimental setup is shown in Fig. 3.12. The cross section image of this fiber can be seen at the lower right corner in Fig. 3.12. As expected, this figure illustrates that the strain sensitivity strongly depends on the sensing head length. Longer FP cavities exhibited lower strain sensitivity. In fact, the smaller the sensing head, the higher its sensitivity. Sensitivities of 3.12 pm/με, 3.79 pm/με, 6.16 pm/με, 15.43 pm/με were respectively obtained for the 906 μm, 207 μm, 35 μm, 13 μm sensing heads. The 207 μm long FP cavity was subjected to temperature variations from the room temperature (~26 °C) to 83 °C. The experimental result shows a very low temperature 78 Chapter 3. Review of Fabry-Pérot Fiber Sensors sensitivity of only 0.81 pm/°C, that is to say, the proposed sensing heads in this work are insensitive to temperature variations. Fig. 3.11. Theoretical response of the normalized strain coefficient with the FP cavity length, for three different single mode fibers: SMF28, SM800 and SM1500. Fig. 3.12. Scheme of the experimental setup. In 2007, Yunjiang Rao et al. first demonstrated to directly fabricate a micro in-line FPI (MFPI) by using a near-infrared femtosecond (fs) laser [7]. Fig. 3.13(a) displays the microscope picture of MFPI with 80 μm cavity length based on the SMF. Fig. 3.13 (b) displays the microscope picture of MFPI with 75 μm cavity length based on the PCF. The dependence of wavelength shift of the SMF-MFPI sensor and the PCF-MFPI sensor on the applied strain are experimentally studied and the results are presented in Fig. 3.14(a) and Fig. 3.14(b), respectively. The strain sensitivity of the SMF-MFPI sensor and the PCF-MFPI sensor are 0.006 nm/με and 0.0045 nm/με, respectively. Hence, the SMF-MFPI sensor is more sensitive to strain than the PCF-MFPI sensor. 79 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 3.13. (a) MFPI with 80 μm cavity length based on SMF; (b) MFPI with 75 μm cavity length based on PCF. Fig. 3.14. Relationship between the applied strain and wavelength-shift of the (a) SMF-MFPI; (b) the PCF-MFPI. The temperature sensitivity of the MFPI sensors are also studied. For a temperature range from 20 °C to 100 °C, the temperature sensitivity of the SMF-MFPI is -0.0021 nm/°C which is close to that of the PCF-MFPI sensor, i.e. -0.002 nm/°C. It should be noted that it is negative because the two end-faces would expand towards the cavity center with the increment of temperature, leading to such a decrease in cavity length. With the increase of the temperature, the wavelengths of both the sensors shift towards the short-wavelength direction. In 2015, Marta S. Ferreira et al. proposed a FP cavity based on a new silica tube [8]. The one shown in Fig. 3.15 is drawn under the pressure p = 2300 Pa, which was chosen to be used as the sensing element. Fig. 3.15. Micrograph of drawn fiber when the pressure p = 2300 Pa. 80 Chapter 3. Review of Fabry-Pérot Fiber Sensors The FP cavity shown in Fig. 3.16 can be produced by splicing a short section of the silica tube between SMFs. Fig. 3.16. Scheme of the experimental setup. A microscope photograph of one FP cavity is also shown. The wavelength responses of different sensing heads are exhibited in Fig. 3.17(a). As expected, the smaller the sensing head, the higher its sensitivity. For the sensing heads with lengths of 17 μm, 51 μm, 70 μm and 198 μm, the obtained strain sensitivities are 13.9 pm/με, 6.0 pm/με, 4.6 pm/με and 3.5 pm/με, respectively. Fig. 3.17. (a) FP cavity sensors response to the applied strain (b) The 198 μm long FP cavity sensor response to temperature. The temperature responses were also investigated. The 198 μm long FP sensing head was subjected to a temperature variation of ~900 °C and a sensitivity of 0.85 pm/°C was obtained, which indicates that this sensor has a cross sensitivity of only ~0.18 με/°C. In 2017, Yi Liu et al. proposed a fiber FP sensor based on micro-cavity plugged by cantilever taper with super long active-length, this fiber FP sensor exhibits ultra-high strain sensitivity [9]. The FPI based on the micro-cavity plugged by cantilever taper with ultra-long activelength is shown in Fig. 3.18, the FP interference length of the FPI based on the micro81 Advances in Optics: Reviews. Book Series, Vol. 3 cavity plugged by cantilever taper is L1, and it is less than the hollow tube length L2. The length L3 of the cantilever taper in the hollow tube can be calculated by L2-L1. Fig. 3.18. The diagram of the FPI based on the micro-cavity plugged by cantilever taper with ultra-long active-length. The stain force sensitivity of the FPI based on the micro-cavity plugged by cantilever taper can be written as: m m L2 = . F AE L1 (3.8) From Eq. (3.8) it can be seen that if the sensing wavelength is fixed, the stain force sensitivity is proportional to the value L2/L1. Therefore, the stain force sensitivity of the FPI can be improved greatly by increasing the hollow tube length L2 or decreasing the FP interference length L1. Then different lengths of cavities are fabricated, the structure A, B and C are (138 µm, 1100 µm), (26 µm, 810 µm) and (3.5 µm, 1360 µm), respectively. The strain force sensitivity of the fabricated fiber inline FP micro-cavity, plugged by cantilever taper are investigated. The changes in the reflection spectra of the three structures with different strain forces are shown in Figs. 3.5(a)–3.5(c), respectively. The results of the three structures were all linear fitted, which are shown in Fig. 3.5(d). The strain force sensitivity of the fiber inline FP micro-cavity C plugged by cantilever taper reached as high as 841.59 nm/N, which equivalents the strain sensitivity of 559 pm/µɛ. By increasing the value of L2/L1, the measured strain force sensitivity of the structure is improved greatly. The sensitivity is approximately proportional to the value L2/L1, which is in accordance with Eq. (3.8). To investigate the crosstalk of the temperature on the strain force sensing, the fiber inline FP micro-cavity C plugged by cantilever taper was placed in air environment in a tube furnace, which was heated from 25 °C to 55 °C with the interval of about 5 °C. When the temperature increases, the interference peak shifted to shorter wavelength direction as the increase of the temperature, and the temperature sensitivity is only 11 pm/°C. However, this temperature sensitivity is pretty high compared to the previous reported ones. In 2014, Shen Liu et al. demonstrated an improved technique and create a unique rectangular air bubble by means of splicing two sections of standard single mode fibers together and tapering the splicing joint [10]. 82 Chapter 3. Review of Fabry-Pérot Fiber Sensors To investigate the strain sensitivity difference between the rectangular and elliptical air bubbles, four samples are used, two rectangular air bubble sample, two elliptical air bubbles, as shown in Fig. 3.19. The fringe dips of the four samples shifted linearly toward a longer wavelength with the increased tensile strain. The strain sensitivity of S1, S2, S3 and S4 was calculated to be 3.0, 29.0, 3.5 and 43.0 pm/με, respectively, by applying a linear fitting of the experimental data. So the strain sensitivity of the rectangular air bubble samples is about nine and twelve times higher than that of the elliptical air bubbles, which indicates that the rectangular air bubble can significantly enhance the strain sensitivity of the air-cavity-based FPI. By using a commercial software, i.e. ANSYS, stress distribution and the deformation of the air cavity under an applied tensile strain is studied, the results show that under the applied tensile strain, the stress distributed on the silica wall of an infiber air bubble sharply increases with the applied tensile strain in case such an air bubble has a rectangular shape and is created in the fused taper. Fig. 3.19. Four in-fiber air bubble samples. In order to study the stress distribution and the deformation of the air cavity under an applied tensile strain, simulation models were established by use of a commercial software, i.e. ANSYS. So the stress distributed on the silica wall of an in-fiber air bubble sharply increases with the applied tensile strain in case such an air bubble has a rectangular shape and is created in the fused taper, which agrees well with the experimental results. 3.3.1.4. Analysis of Temperature Sensitivity for the Fiber FP Strain Sensor The relationship of the cavity length change and the interference pattern shift: ΔL/L= Δλ/λ. Thus, the thermal interference pattern shift sensitivity of 0.95 pm/°C at 1550 nm is equal to ΔL/L = 6.12  107 /°C, this result agrees well with the thermal expansion coefficient of pure silica 5.5 107 /°C. It should be noted that for the air-cavity based fiber FP sensor, the temperature sensitivity is all around 1 pm/°C. The wavelength shift of air cavity to temperature can be given by   L n =  T  L  T n  T          ,  (3.9) 83 Advances in Optics: Reviews. Book Series, Vol. 3 where ε = 5.5 107 and κ = 1.0  105 are the thermal expansion coefficient and the thermo-optic coefficient for pure silica respectively. For the air cavity, the thermo-optic coefficient can be neglected. So the temperature of this kind of FPI sensor is only influenced by silica thermal expansion. Taking the thermal expansion coefficient of silica at wavelength 1550 nm, we get Δλ/T ≈ 0.85 pm/°C. 3.3.1.5. Summary of this Section In this section, we summarized the application of the fiber strain sensor based on FP cavity. According to their shape, they are divided into three categories to introduce. Finally, the temperature characteristics of the FP cavity based fiber strain sensors are analyzed and discussed. Using the same principle, the FP fiber sensor can also be used to measure some other parameters, such as transversal load [11]. 3.3.2. Refractive Index Sensing Refractive index is one of the most important physical parameters. In our daily life, the refractive index (RI) of the fluid (liquid, gas) is often needed to be measured, and people already developed lots of methods to measure refractive index. The optical fiber FP sensor based on the principle of beam interference is widely used as RI sensor, which has high RI sensitivity and is temperature insensitive. There are many types of FP fiber sensors to measure the RI. In this work, they are divided into two categories, one is the method of measuring the fluid RI in the FP cavity, the other is measuring the fluid RI out of the FP cavity. 3.3.2.1. Method of Measuring Fluid RI in FP Cavity 3.3.2.1.1. The Principle of Measuring the Fluid RI in FP Cavity Here we consider a FP cavity model like this: etching a micro-hole in the end of the fiber, and the fiber tip is spliced together with another SMF tip with micro-hole. Then a microchannel is machined vertically to the FP cavity, which allows the RI liquid to flow in or out of the FP cavity, as shown in Fig. 3.20. The micro-channel is fabricated at the cladding, so the size of the micro-channel affects the time of RI liquid flow in (or out of) the FPI cavity and the physical strength of the sensor. But it will not affect the experimental results. When the beam is transmitted to the FP cavity, it will reflect at the two surfaces of the FP cavity. From the Fresnel formula, the reflectivities of the two reflecting surfaces are 2   2 n n  ; R  n1  n2 R1   0 1 , 2 n0  n1  n2  n1  where n0 and n2 are the RI of the fiber core, n1 is the RI of the medium in cavity. 84 (3.10) Chapter 3. Review of Fabry-Pérot Fiber Sensors Fig. 3.20. Diagram of a FPI with air cavity. To avoid the Fresnel reflection at the fiber end, we angle the end of the fiber >8°, [12]. Since the reflectivity of the two reflecting surfaces is very low [13], when the medium in cavity is air or water, the reflectivity is about 0.035 and 0.002, respectively, thus we ignore the multiple reflections in the cavity. The FPI can be treated as a two-beam interferometer and the interference spectrum can be expressed as Equation (3.10). For Equation (3.3), the refractive index is a function of wavelength, L is constant. Differentiating both sides of Equation (3.3) with respect to n , we can obtain n d n 1  2 0.   n    n  dn (3.11) So the RI sensitivity of the sensor can be expressed as S d  n dn   n n . (3.12) In Equation (3.12), it can be seen that the wavelength shift of the interference spectrum is linear with the change in the refractive index of the medium filled in the cavity. Using long wavelength lasers is helpful in improving the sensitivity of the sensor. The sensitivity of the RI is proportional to the wavelength. Thus, it is necessary to indicate the wavelength we use when describing the RI sensitivity. In Equation (3.2), n and L are functions of temperature,  is constant relative to temperature, simplifying both sides of Equation (3.2) with respect to T, we get the temperature sensitivity, d 4  dn dL  n ,   L dt 2m  1  dt dt  (3.13) where dn dt is the thermo-optic coefficient of medium in cavity, it is very small when the medium is air. dL is the thermal expansion coefficient of fiber material, which is also dt very small. It is 0.55  106 /°C for SMF-28 fiber. 85 Advances in Optics: Reviews. Book Series, Vol. 3 3.3.2.1.2. RI Sensing Experiments In 2012, C. R. Liao et al proposed a FP cavity sensor which was etched by laser to measure the RI of a liquid [14] as shown in Fig. 3.21. In the range of 1.31-1.39, the transmission dip wavelength shift towards longer wavelengths with the increase of the refractive index. And the refractive index has a linear relationship with the wavelength of interference dip. Increasing the temperature from 24 to 100 °C, the temperature sensitivity obtained is 4.8 pm/°C at 1554 nm, and the RI sensitivity obtained is 994 nm/RIU with extremely low temperature cross-sensitivity of 4.8 106 RIU/°C. Three different lengths of dielectric cavity are tested. The results show that the FSR decreases with the increase of cavity length, and the intensity of reflected light increases with the increase of cavity length, because the two reflection surfaces within the fiber core are nearly in parallel with a larger radius of the FP cavity, and the reflected light intensity can be enhanced. Fig. 3.21. FPI cavity with the micro-channel. In 2014, Chuang Wu et al. proposed a method for measuring the RI of a liquid in a C-shaped fiber FP cavity [13]. The structure is formed using a C-type open fiber welded two-stage single-mode fiber as shown in Fig. 3.22. In the experiment, the refractive index is changed from 1.33 to 1.36, and the sensitivity is 1368 nm/RIU at 1600 nm. The temperature is heated from room temperature to 600 °C and the sensitivity of the temperature is only 0.42 pm/°C, so temperature cross-sensitivity is 3.04  107 RIU/°C. They also point out that shorter cavities produce a wider range of measurements with greater detection limits. In 2016, ZeJin Lu et al proposed a fast response FP structure based on photonic crystal fiber to measure the RI of the liquid [15]. The structure is shown in Fig. 3.23. The micropores of the photonic crystal fiber in the structure increase its response speed. The response time of water and ethanol is measured to be less than 359 ms and 23 ms, respectively. Ethanol concentration increases from 0 to 19.11 %, the RI sensitivity is 1635.62 nm/RIU at 1500 nm. The temperature increases from 25 °C to 200 °C and the 86 Chapter 3. Review of Fabry-Pérot Fiber Sensors temperature sensitivity is 0.29 pm/°C. The temperature cross-sensitivity 1.77  107 RIU/°C. is Fig. 3.22. C-type FP cavity sensor structure. Fig. 3.23. Fast response photonic crystal FP cavity sensor. In 2012, De-wen Duan et al. proposed a method for measuring the RI of gas using a FP cavity [12], as shown in Fig. 3.24. The gas to be measured enters the FP cavity to change the RI, where the change in RI is achieved by increasing the gas pressure, meanwhile the effect of pressure on FP cavity structure is ignored. When the temperature changes from 30 °C to 300 °C, the wavelength changes by 0.06 nm, so the temperature sensitivity is 2.2 104 nm/°C. When the pressure increases by 700 kPa, the change in RI is 2.14365  103 , so the range of the sensor is limited. The sensitivity of the sensor is 1542 nm/RIU at 1550 mn. In summary, the RI sensitivity is pretty high which can reach 1635.62 nm/RIU at 1500 nm when the liquid and gas fill into the FP Cavity, and the temperature sensitivity is usually less than 1 pm/°C. The temperature cross-over sensitivity is typically at 107 RIU/°C. The sensor responds very quickly, the measurement time is basically dependent on the time of fluid flows into the FP cavity. But its production process is complex, it requires laser etching, fiber splicing and other processes. Besides, the sensor is not convenient to clean. 87 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 3.24. Gas refractive index measurement experimental structure. 3.3.2.2. Method of Measuring Fluid RI out of FP Cavity 3.3.2.2.1. The Principle of Measuring the Fluid RI out of the FP Cavity Etching a microcavity at the end of the fiber, and splicing two fibers can form an air microcavity. This air chamber is the first FP cavity. Cut the end of the fiber and keep the end face clean, so that the reflected light can be returned by the same way at surface 3, the structure is shown in Fig. 3.25. Fig. 3.25. The structure of the sensor. The reflecting surface 2 and the reflecting surface 3 form a second FP cavity. The reflection light of the two FP cavities will interfere in the fiber, and the change of the RI at the end face can be calculated by demodulating the change of the interference light. The reflection surface of the FP cavity is the interface between the fiber and the air. For the air cavity, the reflectance of the two surfaces is equal 2 n 1   0.034  1 , R1  R2   0  n  1 0   88 (3.14) Chapter 3. Review of Fabry-Pérot Fiber Sensors where n0 is the RI of the fiber. The left side of the interface 3 is optical fiber and the right side is fluid medium. The refractive index of the medium on both sides of the interface 3 is different, the beam will be transmitted and reflected at the interface 3, but we only care about the reflected light and ignore the transmitted light. The reflectance of the interface 3 is: 2  n  n'  R3   0 ' , n n  0   (3.15) where n ' is the RI of the fluid, R1 and R2 are certain, so R3 will directly affect the intensity of interference light. The total reflected electric field Er is thus given by Er  R1 Ei  1  A1 1     1  R1  R2 Ei e  j 2  L1  j  1  A1 1  A2 1     1  R1  R3 Ei e for n'  n0 ,  j  2 L1  2 L2  Er  R1 Ei  1  A1 1     1  R1  R2 Ei e j 2  L1  j  1  A1 1  A2 1     1  R1  R3 Ei e   j  2 L1  2 L2   j  for n0  n' , (3.16) where Ei is the input electric field; A1, A2, and A3 are the transmission loss factors at reflection surfaces; βis the propagation constant of the guided mode of the fiber;α is the loss factor of cavity 1. There is a πphase shift at reflection surface 2. When n ' > n0 , there is also a π phase shift at reflection surface 3, where we only talk about Er for n '  n0 . So the normalized reflection spectrum RFP    is given by: RFP     2 Er  R1  1  A1  1     1  R1  R2 2 Ei 2  1  A1  1  A2  1     1  R1  2 2 2 2 2 1  R  2 2 R3 2 R1 R3 1   1  A1 1  A2 1  R1 1  R2  cos  2 R2 R3 1    2 R1 R2 1   1  A1 1  R1  cos  2 4  L1  n0 L2   1  A  1  A 1  R  1  R  cos  4 2 L  2 1 2 2 1 1  4 L1 2    for n '  n0 .  nL      0 (3.17) 2 From Eqs. (3.16), we obtain that only the reflection coefficient R3 depends on the RI to be measured, and RFP    is independent of the power of the input light. 89 Advances in Optics: Reviews. Book Series, Vol. 3 The corresponding fringe contrast is given by R    for n'  n0 . V  10 log10  FP 2   R   1  FP  (3.18) Using Eqs. (3.16) and Eq. (3.17), we obtain that the maximum fringe contrast V (in dB) varies linearly with n ' as V 10 E  n0  n ' , n0 F log e 10 (3.19) where: F  R1  1  A1  1    1  R1  R2  2 R1 R2 1   1  A1 1  R1  ; 2 2 2 E  2 R1 1   1  A1 1  A2 1  R1 1  R2   2 R2 1    1  A1  1  A2 1  R1  1  R2  . 2 2 2 The plus and minus signs are for n '  n0 and n0  n ' , respectively. If R ,  ,  and other parameters are determined, we can obtain the RI of fluid by detecting the change of V. We obtain the RI of fluid by detecting the change of V. In the experiment, fluid does not need to flow into the FP cavity, so it will affect R3 and V. The sensor head needs to be immersed in the solution, each time after the measurement, the sensor is rinsed with Propyl alcohol carefully until the original spectrum is restored and no residual liquid is left. 3.3.2.2.2. Experimental Introduction and Result Analysis In 2008, W. J. Liu et al proposed a method for measuring the RI of liquid with an air FP cavity [16] as shown in Fig. 3.25. The experimental results show that reducing the cavity loss (  ) and transmission loss (A) can improve the refractive index sensitivity. The sensor can supply a measurement of almost any liquid which possess larger RI than air, as long as it is not very closed to the fiber index n0 , if n '  n0 , the fringe contrast will become zero. The refractive index increases from 1.0 to 1.441, the RI sensitivity is 37 dB/RIU. The refractive index increases from 1.33 to 1.441, the RI sensitivity is 27 dB/RIU. The refractive index increases from 1.45 to 1.62, the RI sensitivity is 24 dB/RIU. For different RI of the liquid, the sensitivity will be different. What’s more, the sensor is temperature insensitive. In 2013, Mingshun Jiang et al. proposed a novel TiO2 nanoparticle thin film coated optical fiber FP sensor for RI sensing [17] as shown in Fig. 3.26. The film increases the reflectivity of the interface 4. The reflectivity of interface 4 is changed with different liquids. Over the RI range from 1.333 to 1.8423, the RI sensitivity is 69.38 dB/RIU. The sensor is also temperature insensitive. 90 Chapter 3. Review of Fabry-Pérot Fiber Sensors Fig. 3.26. FP cavity sensing structure. In 2016, Xiaohui Liu et al. proposed an optical fiber FP interferometer based on hollowcore photonic crystal fiber (HCPCF) for RI sensing [18]. The sensor is formed by splicing both ends of a short section of HCPCF to SMFs. As is shown in Fig. 3.27, over a RI range of 1.312 to 1.42, the RI sensitivity is -136 dB/RIU at 1550 nm. The resolution of the sensor is 7 105 RIU. Increasing the temperature from 30°C to 90°C, the temperature sensitivity of the sensor is 10.7 pm /°C. Fig. 3.27. Hollow photonic crystal fiber FP sensor. In conclusion, the method of measuring fluid RI out of FP cavity is by analyzing the variation of the fringe contrast, rather than the wavelength shift of the interference fringe. This type of sensor does not need to fabricate micro-channels on the FP cavity, so it is robust in structure and simple to construct compared with the method of measuring fluid RI in FP cavity. However, the cleaning process of the sensor needs to be very careful to avoid damaging the reflective interface. 3.3.2.3. Summary of this Section In this section, we describe two types of FP RI fiber sensors which are insensitive to temperature. The method of measuring fluid RI in the FP cavity analyzes the wavelength shifts of a point (usually interference spectrum dip) to detect the change of fluid RI. This kind of RI sensors have high sensitivity and fast response, but the microchannel reduces its structural strength. The method of measuring fluid RI out of FP cavity analyzes the variation of the fringe contrast, which does not require the fabrication of micro-channels in the FP cavity, so that its structural strength is higher and easy to construct. 91 Advances in Optics: Reviews. Book Series, Vol. 3 Although lots of ways have been explored to measure the RI, and the sensitivity is pretty high, the study of FP cavity based fiber sensors is still in process. Only by making prominent FP fiber RI sensors, which possess features such as high sensitivity, strong structural strength, low interference, compactness and so on, can it be used widely in practical applications. 3.3.3. Temperature Sensing Temperature is one of the most basic physical quantities in science and technology. Physics, chemistry, thermodynamics, flight mechanics, hydrodynamics and other disciplines are inseparable from temperature measurement. It is also one of the most common and important parameters in industrial production. The quality of the product is closely related to the temperature in many industries. Fiber FP cavity temperature sensor has been widely used, because it has a small size, light weight, anti-electromagnetic interference ability, resistance to harsh environments, high sensitivity, resist harsh environments, simple structure, low production cost advantages. 3.3.3.1. Theory Fiber optical FP sensor is characterized by a single fiber using multi-beam interference to detect. In general, the round-trip optical path length (OPL) of an FP cavity is simply given by: lOPL  2 nL , (3.20) where n is the refractive index and L is the physical length of the cavity, the OPL depends on temperature. For a FPI, the relationship between the free space range (FSR) and the interferometer length L is:   2 2 neff L , (3.21) where λ is the wavelength of light, neff is the effective refraction index (ERI). The light in the fiber first reflect at the first mirror, the reflection light intensity is I1, then reflect at the second mirror, the reflection light intensity is I2, the two reflected light will interference after encountering, the interference between I1 and I2 can be expressed as: I  I1  I 2  2 I  I cos( 1 2 4 neff L   0 ) , (3.22) where neff is the effective refractive index neff of the fundamental mode, L is the length of the FP cavity, λ is the free space wavelength, and φ0 is initial phase. Assuming φ0 = 0, for a certain spectrum peak, the phase difference value of the optical phase shift Δϕ as a function of temperature is given by: 92 Chapter 3. Review of Fabry-Pérot Fiber Sensors   = 4 neff L   2m , 4 nL  1 dn 1 dL   T ,   n dT L dT  (3.23) where m is the integer, φ is the phase. Thus, the wavelength of the peak km is given by: m  2 neff L m . (3.24) When the FPI is subjected to temperature variation, the effective refractive index neff and the fiber length L will change due to thermo-optic effect and thermal-expansion, which can be expressed as: ( neff neff  L 1 )   , L T (3.25) where the thermo-optic coefficient δ = 6.45  10 6 /°C and thermal expansion coefficient  = 5.5x10 7 /°C for silica, respectively. The temperature sensitivity of the FPI is: L  m (   ) . T (3.26) Therefore, the temperature variation can be measured by detecting the output light spectrum of the FPI. 3.3.3.2. Applications of Temperature Sensor Based on FPI In this part, the principle, application and development trend of the FP fiber temperature sensor are introduced. The temperature sensing head are classified into two types according to their structure. The first type is by splicing one fiber section to another, thus forming the FP cavity, for simplicity, we call it splicing type FP cavity. The second type is by coating a thin film at the end of the fiber and forming the FP cavity, we call this kind of sensing head coating type FP cavity. 3.3.3.2.1. Splicing Type FP Fiber Optic Sensor The extrinsic FP fiber temperature sensors, the FP cavity is formed between the two end faces of the fiber, the air cavity has a small thermal expansion coefficient. Although the external FPI is relatively simple and low cost, its drawbacks are obvious, like low coupling efficiency, requiring calibration and tight packaging etc., all of these things limit the development of extrinsic FPIs. To overcome this problem, special fiber is adopted by 93 Advances in Optics: Reviews. Book Series, Vol. 3 many researcher. By splicing a small section of the special fiber into two SMFs, high temperature sensitivities can be achieved. What’s more, it is low cost, robust and easy to fabricate. In 2015, Peng Zhang et al. demonstrated a FPI-based high temperature fiber sensor fabricated by splicing two single mode fibers (SMFs) to both ends of a section of simplified hollow core fiber (SHCF) with a certain length and cleaving one of the two SMFs to a designed length [19]. The highest temperature sensitivity is 1.019 nm/°C for the envelope under temperature range from 250 to 300 °C. The fiber sensor can be operated in the temperature measurement range from 20 to 1050 °C. The configuration of the SHCF-based FPI fiber-optic sensor head is shown in Fig. 3.28. The use of this fiber can increase the effective reflected power at the splice point, which is different from the use of solid PCF that has a large refractive index. Fig. 3.28. Configuration of the SHCF-based FPI sensor (a) and cross section of SHCF (b). In 2008, Hae Young Choi et al. demonstrated a compact FPI fiber sensor suitable for hightemperature sensing [20]. The sensor head consists of two FP cavities formed by fusion splicing a short segment of hollow-core fiber and single-mode fiber at the end of a PCF. The schematic of the proposed sensor system is presented in Fig. 3.29. The SMF part is used as the main sensing area, and the HOF air cavity part is used as the auxiliary sensing cavity. The temperature response of the proposed sensor is measured up to 1000 °C and analyzed in the spatial frequency domain of the reflection spectrum. Fig. 3.29. Schematic of the sensor system (top) and the detail structure of the proposed sensor head (bottom). The inset in the middle is the microscope photograph of a fabricated sensor head. 94 Chapter 3. Review of Fabry-Pérot Fiber Sensors In 2015, the temperature sensor proposed by Xinghu Fu et al. exhibits a relatively high sensitivity [21]. The fiber FP temperature sensor is fabricated by just splicing photonic crystal fiber (PCF) and a section of single-mode fiber (SMF) together. Fig. 3.30 shows the schematic design. With the temperature increasing, due to the thermal-optic effect and the thermal-expansion effect of the fiber, both the RI of the core and the distance between the two reflectors increase, thus leading to the wavelengths shift toward longer wavelength. Four wavelength dips are selected as the observing points, in the temperature range of 30-80°C, the highest sensitivity is 11.12 pm/°C. Fig. 3.30. (a) Schematic diagram of the FPI sensor (b) Microscope images of the two reflection surfaces. In 2013 Wei Peng et al. constructed a miniature FPI temperature sensor based on a dualcore photonic crystal fiber (DCPCF) [22]. Fig. 3.31 shows a schematic design of the proposed DCPCF-based FPI sensor. The DCPFC has two symmetric silica cores that separated by one air hole in the center point of the fiber, thus the light guided by SMF will partially reflect at the fiber-air interface at the splicing point because of the Fresnel reflection. The rest of the light will be guided through the core and reflected back at the cleave ends of the DCPCF, finally recoupled into the SMF again. These two beams will interfere and the interferometric spectrum can be used for sensing. The temperature response of two sensors with PCF lengths of 164 and 953 μm has been characterized, respectively. The temperature sensitivities are 13.17 pm/°C for the 164 μm long FPI and 13.18 pm/°C for the 953 μm long FPI in a temperature range of 40-480 °C. Fig. 3.31. Schematics of DCPCF-based fiber optic FPI sensor system: (a) Sensor structure and operating mechanism; (b) Microscope photograph of the DCPCF and splicing processing. 95 Advances in Optics: Reviews. Book Series, Vol. 3 In 2011, D. W. Duan et al. designed a novel compact FPI which can measure temperature up to 1000 [23]. The sensor is formed by fusion splicing two sections of SMFs with a large lateral offset. The temperature responses of the FPI based on larger lateral offset splicing induced by two parallel separated mirrors are demonstrated. The sensor has been tested under high temperature, showing a high sensitivity of 41 nm/°C. The schematic of sensor system and detailed structure of proposed sensor head is shown in Fig. 3.32. This larger lateral offset splicing can excite higher order cladding modes, thus increasing the sensitivity of sensor. Fig. 3.32. Schematic of sensor system (top) and detailed structure of proposed sensor head (bottom). In 2015, Chen Pengfei et al. proposed a FP air cavity which can be created by splicing a special sapphire-derived fiber with SMF directly and cleaving the pigtail of the sapphirederived fiber subsequently [24]. The microphotography of the FP cavity fabrication process is shown in Fig. 3.33. This FP interferometer shows a high sensitivity to temperature of about 15.7 pm/°C within the temperature range up to 1000 °C. Fig. 3.33. The microphotography of the FP cavity fabrication by using sapphire-derived fiber, (a) well aligned fibers; (b) air cavity created at the splicing point, and (c) Fabry-Perot fabricated in the sapphire-derived fiber. 96 Chapter 3. Review of Fabry-Pérot Fiber Sensors 3.3.3.2.2. Coating Type FP Fiber Optic Sensor In this section, the FP cavity is formed by thin films made in different material, the outer temperature have great influences on the length of the FP cavity. Thus, it can possess extremely high temperature sensitivity. Besides, some material can be operated under high temperature, so they appeared to be suitable for high temperature sensing. In 2015 Jinesh Mathew et al. presented an in-fiber FP fiber sensor for high-temperature sensing [25]. The sensor has been demonstrated for high-temperature sensing up to 1100 °C. As shown in Fig. 3.34 (a), the FP cavity is formed between a reflective in-fiber metallic splice and the air-fiber boundary at the end of the sensor head, to make it more compact, a second FPI is fabricated where the sensor is in a reduced diameter. The sensitivity increases with longer FP cavity length. Fig. 3.34. Basic structure of the optical fibre sensor; (a) Image of the fabricated 125 μm diameter sensor; (b) Image of the fabricated 50 μm sensor. Similarly, they demonstrated two types of FP fiber sensors, SMF-Cr-SMF, and SMF-CrPCF. Both are tested over a temperature range from room temperature up to 1100 °C with good repeatability. In 2016, Iván Hernández-Romano et al. proposed a similar structure used for low temperature sensing [26]. The sensor head consists of FP micro-cavity formed by an internal mirror made of a thin titanium dioxide (TiO2) film and a segment of SMF covered with Poly (dimethylsiloxane) (PDMS). The structure of the sensing head is shown in Fig. 3.35. In this design, the RI of PDMS varies according to the temperature, the temperature change can be detected by the amplitude variation of the interference pattern. By monitoring the extinction ratio of the interference pattern, a temperature sensitivity of 0.13 dB/℃ was observed in the temperature range of 22 °C - 60 °C. In 2014, M. J˛edrzejewska-Szczerska et al. demonstrated a novel optical fiber sensor of temperature using a thin ZnO layer [27]. The ZnO layer was coated on SMF and thus a FP cavity was built. The ZnO thin films can be used in two basic extrinsic FPI types, as shown in Fig. 3.36. The sensitivity of temperature measurement is measured to be 0.05 nm/°C in the temperature range from 50 to 300 °C. 97 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 3.35. Schematic diagram of the sensing head. Fig. 3.36. FP interferometer with the cavity made of a ZnO layer: (a) Symmetric configuration; (b) Asymmetric configuration; h is the thickness of the cavity, M1, M2 are the cavity mirrors. 3.3.3.2.3. Other Types In addition to the above mentioned splicing and coating method, there are some other special methods using FP cavity for temperature sensing. Such as a sapphire fiber extrinsic FP interferometer for ultrahigh temperature [28], a U-shaped optical FPI is constructed for high temperature sensing [29], and fiber FPI assisted with iron V-groove for temperature measurement [30]. 3.3.3.3. Summary of this Section In summary, we divided the FP temperature sensor into two types, splicing type and coating type. The measuring temperature range of the splicing type is lower than coated type, but it possesses higher sensitivity. The coating type FPI can measure higher temperature, because the thin films which constructed the FP is made from special materials can work in high temperature. 3.4. Concluding Remarks and Perspectives In the past few decades, optical fiber sensors have been widely used in sensing applications of various fields. Different fiber sensors based on various principles have 98 Chapter 3. Review of Fabry-Pérot Fiber Sensors been proposed [31, 32]. Among which, FP cavity based fiber sensors possess the advantages of compactness, simple configuration, small size, high sensitivity, fast responses etc., which make them attractive for sensing applications in industry. This chapter reviewed the typical intrinsic FP fiber sensors. The fundamental principles of the FP fiber sensors are discussed in detail. Each application is reviewed in turn, key recent researches and their contributions for the development of the FP fiber sensors are highlighted and discussed. Some methods for fabricating FP fiber sensors are described according to their operating principles, fabrication methods, and application fields. However, due to temperature crosstalk and other factors in the FP fiber sensors, there are still some application limitations. And there is still long way to go to make the FP fiber sensors into more compact, integrated devices, can adjust to extreme environment, and meet different parameter sensing requirements. In order to make it more perfect, solving these problems is of utmost importance. References [1]. F. C. Favero, L. Araujo, G. Bouwmans, V. Finazzi, J. Villatoro, V. Pruneri, Spheroidal FabryPerot microcavities in optical fibers for high-sensitivity sensing, Opt. Express, Vol. 20, 2012, pp. 7112-7118. [2]. S. Liu, Y. P. Wang, C. R. Liao, G. J. Wang, Z. Y. Li, Q. Wang, J. T. Zhou, K. M. Yang, X. Y. Zhong, J. Zhao, J. Tang, High sensitivity strain sensor based on in-fiber improved Fabry-Perot interferometer, Opt. Lett., Vol. 39, 2014, pp. 2121-2124. [3]. D. W. Duan, Y. J. Rao, Y. S. Hou, T. Zhu, Microbubble based fiber-optic Fabry-Perot interferometer formed by fusion splicing single-mode fibers for strain measurement, Appl. Opt., Vol. 51, 2012, pp. 1033-1036. [4]. J. Villatoro, V. Finazzi, G. Coviello, V. Pruneri, Photonic-crystal-fiber-enabled micro-Fabry– Perot interferometer, Opt. Lett., Vol. 34, 2009, pp. 2441-2443. [5]. Y. Zhao, M. Q. Chen, R. Q. Lv, F. Xia, In-fiber rectangular air Fabry-Perot strain sensor based on high-precision fiber cutting platform, Opt. Commun., Vol. 384, 2017, pp. 107-110. [6]. M. S. Ferreira, J. Bierlich, J. Kobelke, K. Schuster, J. L. Santos, O. Frazão, Towards the control of highly sensitive Fabry-Pérot strain sensor based on hollow-core ring photonic crystal fiber, Opt. Express, Vol. 20, 2012, pp. 21946-21952. [7]. Y. J. Rao, M. Deng, D. W. Duan, X. C. Yang, T. Zhu, G. H. Cheng, Micro Fabry-Perot interferometers in silica fibers machined by femtosecond laser, Opt. Express, Vol. 15, 2007, pp. 14123-14128. [8]. M. S. Ferreira, P. Roriz, J. Bierlich, J. Kobelke, K. Wondraczek, C. Aichele, K. Schuster, J. L. Santos, O. Frazão, Fabry-Perot cavity based on silica tube for strain sensing at high temperatures, Opt. Express, Vol. 23, 2015, pp. 16063-16070. [9]. Y. Liu, C. P. Lang, X. C. Wei, S. L. Qu, Strain force sensor with ultra-high sensitivity based on fiber inline Fabry-Perot micro-cavity plugged by cantilever taper, Opt. Express, Vol. 25, 2017, pp. 7797-7806. [10]. S. Liu, K. M. Yang, Y. P. Wang, J. L. Qu, C. R. Liao, J. He, Z. Y. Li, G. L. Yin, B. Sun, J. T. Zhou, G. J. Wang, J. Tang, J. Zhao, High-sensitivity strain sensor based on in-fiber rectangular air bubble, Sci. Rep., Vol. 5, 2015, 7624. [11]. Y. F. Wu, Y. D. Zhang, J. Wu, P. Yuan, Temperature-insensitive fiber optic Fabry-Perot interferometer based on special air cavity for transverse load and strain measurements, Opt. Express, Vol. 25, 2017, pp. 9443-9448. 99 Advances in Optics: Reviews. Book Series, Vol. 3 [12]. D. W. Duan, Y. J. Rao, T. Zhu, High sensitivity gas refractometer based on all-fiber opencavity Fabry-Perot interferometer formed by large lateral offset splicing, Opt. Soc., Vol. 29, 2012, pp. 912-915. [13]. C. Wu, Z. Y. Liu, A. P. Zhang, B.-O. Guan, H.-Y. Tam, In-line open-cavity Fabry-Pérot interferometer formed by C-shaped fiber for temperature insensitive refractive index sensing, Opt. Express, Vol. 22, 2014, pp. 21757-21766. [14]. C. R. Liao, T. Y. Hu, D. N. Wang, Optical fiber Fabry-Perot interferometer cavity fabricated by femtosecond laser micromachining and fusion splicing for refractive index sensing, Opt. Express, Vol. 20, 2013, pp. 22813-22818. [15]. J. J. Tian, Z. J. Lu, M. R. Quan, Y. Z. Jiao, Y. Yao, Fast response Fabry-Perot interferometer microfluidic refractive index fiber sensor based on concave-core photonic crystal fiber, Opt. Express, Vol. 24, 2016, pp. 20132-20142. [16]. Z. L. Ran, Y. J. Rao, W. J. Liu, X. Liao, K. S. Chiang, Laser-micromachined Fabry-Perot optical fiber tip sensor for high-resolution temperature-independent measurement of refractive index, Opt. Express, Vol. 16, 2008, pp. 2252-2263. [17]. M. S. Jiang, Q. S. Li, J. N. Wang, Z. W. Jin, Q. M. Sui, TiO2 nanoparticle thin film-coated optical fiber Fabry-Perot sensor, Opt. Express, Vol. 21, 2013, pp. 3083-3090. [18]. X. H. Liu, M. S. Jiang, Q. M. Sui, S. Y. Luo, X. Y. Geng, Optical fiber Fabry–Perot interferometer for microorganism growth detection, Opt. Fiber Technol., Vol. 30, 2016, pp. 32-37. [19]. P. Zhang, M. Tang, F. Gao, B. P. Zhu, Z. Y. Zhao, L. Duan, S. N. Fu, J. O. Yang, H. F. Wei, P. P. Shum, D. M. Liu, Simplified hollow-core fiber-based Fabry-Perot interferometer with modified vernier effect for highly sensitive high-temperature measurement, IEEE Photonics J., Vol. 7, 2015, pp. 1-10. [20]. H. Y. Choi, K. S. Park, S. J. Park, U. C. Paek, B. H. Lee, E. S. Choi, Miniature fiber-optic high temperature sensor based on a hybrid structured Fabry-Perot interferometer, Opt. Lett., Vol. 33, 2008, pp. 2455-2457. [21]. X. H. Fu, H. Y. Xie, F. Wang, P. Jiang, G. W. Fu, W. H. Bi, Miniature Fabry-Perot interferometric sensor for temperature measurement based on photonic crystal fiber, Optoelectron. Lett., Vol. 11, 2015, pp. 382-385. [22]. W. Peng, X. P. Zhang, Y. Liu, X. L. Zhou, Q. X. Yu, Temperature characteristics of a coremode interferometer based on a dual-core photonic crystal fiber, Appl. Phys. B, Vol. 116, 2014, pp. 21-26. [23]. D. W. Duan, Y. J. Rao, W. P. Wen, J. Yao, D. Wu, L. C. Xu, T. Zhu, In-line all-fibre FabryPerot interferometer high temperature sensor formed by large lateral offset splicing, Electron. Lett., Vol. 47, 2011, pp. 401-403. [24]. P. F. Chen, F. F. Pang, Z. W. Zhao, L. Hong, N. C. Chen, Z. Y. Chen, T. Y. Wang, FabryPerot cavity based on sapphire-derived fiber for high temperature sensor, in Proceedings of the International Conference on Optical Fibre Sensors (OFS’15), Society of Photo-Optical Instrumentation Engineers, Curitiba, Brazil, 2015, 96347T. [25]. J. Mathew, O. Schneller, D. Polyzos, D. Havermann, R. M. Carter, W. N. MacPherson, D. P. Hand, R. J. Maier, In-fiber Fabry-Perot cavity sensor for high-temperature applications, J. Lightw. Technol., Vol. 33, 2015, pp. 2419-2425. [26]. I. Hernández-Romano, M. A. Cruz-Garcia, C. Morenohernández, D. Monzónhernández, E. O. Lópezfigueroa, Optical fiber temperature sensor based on a microcavity with polymer overlay, Opt. Express, Vol. 24, 2016, pp. 5654-5661. [27]. M. Jędrzejewska-Szczerska, P. Wierzba, A. A. Chaaya, M. Bechelany, P. Miele, R. Viter, A. Mazikowski, K. Karpienko, M. Wróbel, ALD thin ZnO layer as an active medium in a fiber-optic Fabry-Perot interferometer, Sens. & Actuators A Phys., Vol. 221, 2015, pp. 88-94. [28]. Z. P. Tian, Z. H. Yu, B. Liu, A. B. Wang, Sourceless optical fiber high temperature sensor, Opt. Lett., Vol. 41, 2016, pp. 195-198. 100 Chapter 3. Review of Fabry-Pérot Fiber Sensors [29]. L. Yuan, X. Lan, J. Huang, H. Z. Wang, B. K. Cheng, J. Liu, H. Xiao, Miniaturized optical fiber Fabry-Perot interferometer fabricated by femtosecond laser irradiation and selective chemical etching, in Proceedings of the Conference Advanced Fabrication Technologies for Micro/Nano Optics and Photonics VII, Society of Photo-Optical Instrumentation Engineers, San Francisco, CA, USA, 2014, 89741A. [30]. X. D. Wen, T. G. Ning, Y. Bai, C. Li, J. Li, C. B. Zhang, Ultrasensitive temperature fiber sensor based on Fabry-Pérot interferometer assisted with iron V-groove, Opt. Express, Vol. 23, 2015, pp. 11526-11536. [31]. J. F. Wang, Y. D. Zhang, X. N. Zhang, H. Tian, H. Wu, Y. X. Cai, J. Zhang, P. Yuan, Enhancing the sensitivity of fiber Mach–Zehnder interferometers using slow and fast light, Opt. Lett., Vol. 36, 2011, pp. 3173-3175. [32]. J. F. Wang, Y. D. Zhang, J. Zhang, Y. X. Cai, X. N. Zhang, P. Yuan, Simultaneous observation of superluminal and slow light propagation in a nested fiber ring resonator, Opt. Express, Vol. 18, 2010, pp. 13180-13186. 101 Chapter 4. Multi-Parameter Integrated Optical Sensor Based on Multimode Interference and Microring Resonator Structures Chapter 4 Multi-Parameter Integrated Optical Sensor Based on Multimode Interference and Microring Resonator Structures Trung-Thanh Le1 4.1. Introduction Current approaches to the real time analysis of chemical and biological sensing applications utilize systematic approaches such as mass spectrometry for detection. Such systems are expensive, heavy and cannot monolithically integrated in one single chip [1]. Electronic sensors use metallic probes which produces electro-magnetic noise, which can disturb the electro-magnetic field being measured. This can be avoided in the case of using integrated optical sensors. Integrated optical sensors are very attractive due to their advantages of high sensitivity and ultra-wide bandwidth, low detection limit, compactness and immunity to electromagnetic interference [2, 3]. Optical sensors have been used widely in many applications such as biomedical research, healthcare and environmental monitoring. Typically, detection can be made by the optical absorption of the analytes, optic spectroscopy or the refractive index change [1]. The two former methods can be directly obtained by measuring optical intensity. The third method is to monitor various chemical and biological systems via sensing of the change in refractive index [4]. Optical waveguide devices can perform as refractive index sensors particularly when the analyte becomes a physical part of the device, such as waveguide cladding. In this case, the evanescent portion of the guided mode within the cladding will overlap and interact with the analyte. The measurement of the refractive index change of the guided mode of the optical waveguides requires a special structure to convert the refractive index change into detectable signals. A number of refractive index sensors based on optical waveguide structures have been reported, including Bragg grating sensors, directional coupler Trung-Thanh Le International School (VNU-IS), Vietnam National University (VNU), Cau Giay, Hanoi, Vietnam 103 Advances in Optics: Reviews. Book Series, Vol. 3 sensors, Mach- Zehnder interferometer (MZI) sensors, microring resonator sensors and surface plasmon resonance sensors [1, 4-7]. Recently, the use of optical microring resonators as sensors [2, 6] is becoming one of the most attractive candidates for optical sensing applications because of its ultra-compact size and easy to realize an array of sensors with a large scale integration [8-10]. When detecting target chemicals by using microring resonator sensors, one can use a certain chemical binding on the surface. There are two ways to measure the presence of the target chemicals. One is to measure the shift of the resonant wavelength and the other is to measure the optical intensity with a fixed wavelength. In the literature, some highly sensitive resonator sensors based on polymer and silicon microring and disk resonators have been developed [11-14]. However, multichannel sensors based on silicon waveguides and MMI structures, which have ultra-small bends due to the high refractive index contrast and are compatible with the existing CMOS fabrication technologies, are not presented much. In order to achieve multichannel capability, multiplexed single microring resonators must be used. This leads to large footprint area and low sensitivity. For example, recent results on using single microring resonators for glucose and ethanol detection showed that sensitivity of 108 nm/RIU [2, 15], 200 nm/RIU [16] or using microfluidics with grating for ethanol sensor with a sensitivity of 50 nm/RIU [17]. Silicon waveguide based sensors has attracted much attention for realizing ultra-compact and cheap optical sensors. In addition, the reported sensors can be capable of determining only one chemical or biological element. The sensing structures based on one microring resonator or Mach Zender interferometer can only provide a small sensitivity and single anylate detection [13]. Therefore, in this study, we present new structures for achieving a highly sensitive and multichannel sensor. Our structures are based on only 4×4, 6×6 and 8×8 multimode interference (MMI) coupler assisted microring resonators for two, three and four parameter sensors. The proposed sensors provide very high sensitivity compared with the conventional MZI sensor. In addition, it can measure multi-parameter target chemicals and biological elements simultaneously. 4.2. Multimode Interference Structures The conventional MMI coupler has a structure consisting of a homogeneous planar multimode waveguide region connected to a number of single mode access waveguides. The MMI region is sufficiently wide to support a large number of lateral modes. There are three main interference mechanisms. These mechanisms depend upon the locations of the access waveguides [18]. The first is the general interference (GI) mechanism which is independent of the modal excitation. The second is the restricted interference (RI) mechanism, in which excitation inputs are placed at some special positions so that certain modes are not excited. The last mechanism is the symmetric interference (SI), in which the excitation input is located at the centre of the multimode section. 104 Chapter 4. Multi-Parameter Integrated Optical Sensor Based on Multimode Interference and Microring Resonator Structures The characteristics of an MMI device can be described by a transfer matrix [19-21]. This transfer matrix is a very useful tool for analyzing cascaded MMI structures. The phase ij associated with imaging an input i to an output j in an MMI coupler. These phases ij form a matrix  , with i representing the row number, and j representing the column number. Then the transfer matrix of the MMI coupler  is directly related to  , and the output field distribution emerging from the MMI coupler can be written as b  Ma , (4.1) where a  [a1 a 2 . . . a N ]T , b  [b1 b 2 . . . b N ]T and M  [mij ]NxN . The superscript T indicates the transpose of a matrix. a i (I = 1,..,N) is the complex field amplitude at input waveguide i and b j (j = 1,..,N) is the complex field amplitude at output waveguide j. Elements of the transfer matrix M are mij  m ji  Aije jij , where A ij is the field amplitude transfer coefficient and ij is the phase shift when imaging from input i to output j. 4.3. Microring Resonator Consider a curved waveguide having a radius R connected to an MMI coupler to form a single microresonator as shown in Fig. 4.1. Fig. 4.1. The structure of a microresonator using a 2×2 MMI coupler. If the common phase factor 0 of the MMI coupler is factored out for simplicity, then the complex amplitudes of the input signals a i (i=1, 2) and output signals b j (j=1, 2) are related through the transfer matrix of the 2×2 MMI coupler [22] b = Ma , (4.2) 105 Advances in Optics: Reviews. Book Series, Vol. 3  τ where M =  *  -κ κ , a  [a1 a 2 ]T and b  [b1 b2 ]T . * τ  (4.3) Here  and  are the amplitude transmission and coupling coefficients of the coupler, respectively. The superscripts * and T denote the complex conjugate and the transpose of 2 2 a matrix, respectively. For a lossless coupler,     1 . A plot of the transmission characteristics as a function of microresonator loss factor (  ), with transmission coefficient  as parameter, is presented in Fig. 4.2. The transmission loss factor  is   exp(0 LR ) , where L R is the total length of the racetrack (or ring) waveguide and 0 (dB / cm) is the transmission loss coefficient. Fig. 4.2. The transmission characteristic of a single microresonator based on a 2×2 MMI. By rapidly changing the loss/gain or the coupling coefficient of the coupler, optical modulators and optical switches can be created. In addition, a single microresonator can be used as an optical notch filter. The spectral response of the microresonator is shown in Fig. 4.3, for a loss factor of α = 0.7. Here,  is the phase accumulated inside the microresonator,   0 (2R  L') , where 0 is the propagation constant, L ' is the length shown in Fig. 4.3 and R is the radius of the curved waveguide. The simulations show that the largest extinction ratio can be achieved with critical coupling that is when the loss factor  equals the transmission coefficient  (    ). 4.4. Two-Parameter Sensor Based on 4×4 MMI and Resonator Structure We present a structure for achieving a highly sensitive and multichannel sensor. Our structure is based on only one 4×4 multimode interference (MMI) coupler assisted microring resonators [23, 24]. The proposed sensors provide very high sensitivity compared with the conventional MZI sensors. In addition, it can measure two different 106 Chapter 4. Multi-Parameter Integrated Optical Sensor Based on Multimode Interference and Microring Resonator Structures and independent target chemicals and biological elements simultaneously. We investigate the use of our proposed structure to glucose and ethanol sensing at the same time. The proposed sensor based on 4×4 multimode interference and microring resonator structures is shown in Fig. 4.4. The two MMI couplers are identical. The two 4×4 MMI couplers have the same width WMMI and length LMMI . Fig. 4.3. Transmission characteristic of a single microresonator. Fig. 4.4. Schematic of the new sensor using 4×4 MMI couplers and microring resonators. In this structure, there are two sensing windows having lengths Larm1 , Larm2 . As with the conventional MZI sensor device, segments of two MZI arms overlap with the flow channel, forming two separate sensing regions. The other two MZI arms isolated from the analyte by the micro fluidic substrate. The MMI coupler consists of a multimode optical waveguide that can support a number of modes [25]. In order to launch and extract light from the multimode region, a number of single mode access waveguides are placed at the input and output planes. If there are N input waveguides and M output waveguides, then the device is called an NxM MMI coupler. 107 Advances in Optics: Reviews. Book Series, Vol. 3 In this study, the access waveguides are identical single mode waveguides with width Wa . The input and output waveguides are located at [18] 1 W x i  (i  ) MMI , (i = 0, 1,…, N-1). 2 N (4.4) The electrical field inside the MMI coupler can be expressed by [19] E(x,z)  exp( jkz) M  m 1 Em exp( j m2  m z)sin( x). 4 WMMI (4.5) 3L  , where L  is the beat 2 length of the MMI coupler [26]. One can prove that the normalized optical powers transmitted through the proposed sensor at wavelengths on resonance with the microring resonators are given by [9] If we choose the MMI coupler having a length of L MMI  2  1   1  cos( 2 )  T1    ,  1   cos( 1 )  1  2  (4.6) 2  2    2  cos( 2 )  T2    , 1   cos( 2 )  2 2   (4.7) 1  ) 2  sin( 2 ), and  2  cos( 2 ) ; 1 , 2 are the 2 2 2 2 phase differences between two arms of the MZI, respectively; 1 ,  2 are round trip transmissions of light propagation through the two microring resonators [27]. where 1  sin( 1 ) , 1  cos( In this study, the locations of input, output waveguides, MMI width and length are carefully designed, so the desired characteristics of the MMI coupler can be achieved. It is now shown that the proposed sensor can be realized using silicon nanowire waveguides [28, 29]. By using the numerical method, the optimal width of the MMI is calculated to be WMMI  6 m for high performance and compact device. The core thickness is h co = 220 nm. The access waveguide is tapered from a width of 500 nm to a width of 800 nm to improve device performance. It is assumed that the designs are for the transverse electric (TE) polarization at a central optical wavelength   1550 nm . The FDTD simulations for sensing operation when input signal is at port 1 and port 2 for 108 Chapter 4. Multi-Parameter Integrated Optical Sensor Based on Multimode Interference and Microring Resonator Structures glucose and ethanol sensing are shown in Fig. 4.5 (a) and 4.5 (b), respectively. The mask design for the whole sensor structure using CMOS technology is shown in Fig. 4.5 (c). (a) Input 1, glucose sensing (b) Input 2, Ethanol sensing (c) Mask design Fig. 4.5. FDTD simulations for two-channel sensors (a) glucose; (b) Ethanol and (c) mask design. The proposed structure can be viewed as a sensor with two channel sensing windows, which are separated with two power transmission characteristics T1 , T2 and sensitivities S1 , S2 . When the analyte is presented, the resonance wavelengths are shifted. As the result, the proposed sensors are able to monitor two target chemicals simultaneously and their sensitivities can be expressed by: S1  1  , S2  2 , n c n c (4.8) where 1 and 2 are resonance wavelengths of the transmissions at output 1 and 2, respectively. For the conventional sensor based on MZI structure, the relative phase shift  between two MZI arms and the optical power transmitted through the MZI can be made a function of the environmental refractive index, via the modal effective index n eff . The transmission at the bar port of the MZI structure can be given by [1] 109 Advances in Optics: Reviews. Book Series, Vol. 3 TMZI  cos 2 (  ), 2 (4.9) where   2Larm (n eff ,a  n eff ,0 ) /  , Larm is the interaction length of the MZI arm, n eff ,a is effective refractive index in the interaction arm when the ambient analyte is presented and n eff ,0 is effective refractive index of the reference arm. The sensitivity S MZI of the MZI sensor is defined as a change in normalized transmission per unit change in the refractive index and can be expressed as SMZI  TMZI , n c (4.10) where n c is the cover medium refractive index or the refractive index of the analyte. The sensitivity of the MZI sensor can be rewritten by SMZI  TMZI TMZI n eff ,a .  n c n eff ,a n c The waveguide sensitivity parameter n eff ,a n c theorem for optical waveguides [1]: n eff ,a n c  nc n eff ,a  can be calculated using the variation 2 E a (x, y) dxdy analyte  (4.11) 2 E a (x, y) dxdy , (4.12)  where E a (x, y) is the transverse field profile of the optical mode within the sensing region, calculated assuming a dielectric material with index n c occupies the appropriate part of the cross-section. The integral in the numerator is carried out over the fraction of the waveguide cross-section occupied by the analyte and the integral in the denominator is carried out over the whole cross-section. For sensing applications, sensor should have steeper slopes on the transmission and phase shift curve for higher sensitivity. From (4.9) and (4.10), we see that the sensitivity of the MZI sensor is maximized at phase shift   0.5 . Therefore, the sensitivity of the MZI sensor can be enhanced by increasing the sensing window length L a or increasing the n eff ,a , which can be obtained by properly designing optical waveguide sensitivity factor n c waveguide structure. In this chapter, we present a new sensor structure based on microring resonators for very high sensitive and multi-channel sensing applications. 110 Chapter 4. Multi-Parameter Integrated Optical Sensor Based on Multimode Interference and Microring Resonator Structures S1 / SMZI From equations (4.8) and (4.10), the ratio of the sensitivities of the proposed sensor and the conventional MZI sensor can be numerically evaluated. The sensitivity enhancement factor S1 / SMZI can be calculated for values of 1 between 0 and 1 is plotted in Fig. 4.6. For 1  0.99 , an enhancement factor of approximately 10 is obtained. The similar results can be achieved for other sensing arms. Round trip 1 Fig. 4.6. Sensitivity enhancement factor for the proposed sensor, calculated with the first sensing arm. In general, our proposed structure can be used for detection of chemical and biological elements by using both surface and homogeneous mechanisms. Without loss of generality, we applied our structure to detection of glucose and ethanol sensing as an example. The refractive indexes of the glucose ( n glucose ) and ethanol ( n EtOH ) can be calculated from the concentration (C %) based on experimental results at wavelength 1550 nm by [30-32] n glucose  0.2015  [C]  1.3292, (4.13) n EtOH  1.3292  a[C]  b[C]2 , (4.14) where a  (8.4535  104 ) and b  (4.8294  106 ) . The refractive indexes of the glucose and EtOH at different concentrations are shown in Fig. 4.7. In our design, the silicon waveguide with a height of 220 nm, width of 500 nm is used for single mode operation. The wavelength is at 1550 nm. It is assumed that the interaction lengths for glucose and ethanol sensing arms are 100 m . By using the finite difference method (FDM), the effective refractive indexes of the waveguide at different concentration is shown in Fig. 4.8. The glucose solutions with concentrations of 0 %, 0.2 % and 0.4 % and Ethanol concentrations of 0 %, 3 % and 6 % are induced to the device. The resonance wavelength shifts corresponding to the concentrations can be measured by the optical spectrometer as 111 Advances in Optics: Reviews. Book Series, Vol. 3 shown in Fig. 4.9 for glucose and Fig. 4.10 for ethanol. For each 0.2 % increment of the glucose concentration, the resonance wavelength shifts of about 105 pm is achieved. This is a greatly higher order than that of the recent conventional sensor based on single microring resonator [31, 33]. For each 3 % increment of the ethanol concentration, the resonance wavelength shifts of about 1.5×104 pm is achieved. Fig. 4.7. Refractive indexes of the glucose and ethanol vs. concentations. Fig. 4.8. Effective refractive indexes of the waveguide with glucose and ethanol cover at different concentrations. 112 Chapter 4. Multi-Parameter Integrated Optical Sensor Based on Multimode Interference and Microring Resonator Structures Fig. 4.9. Resonance wavelength shift at different glucose concentrations. Fig. 4.10. Resonance wavelength shift at different ethanol concentrations. By measuring the resonance wavelength shift (  ), the glucose concentration is detected. The sensitivity of the glucose sensor can be calculated by Sglu cos e    9000(nm/ RIU). n (4.15) Our sensor provides the sensitivity of 9000 nm/RIU compared with a sensitivity of 170 nm/RIU [33]. In addition to the sensitivity, the detection limit (DL) is another important parameter. For the refractive index sensing, the DL presents for the smallest ambient refractive index change, which can be accurately measured. The Detection limit (DL) can be calculated as the ratio of the resonance wavelength resolution  to the sensitivity Sglu cose by [34] 113 Advances in Optics: Reviews. Book Series, Vol. 3 DL   Sglu cose , (4.16) 2 2 2 where   amp  noise   temp induced  spec  res , amp  noise is the standard deviation of the spectral variation which is determined by the Q factor and extinction ratio, tempinduced is the standard deviation induced by noises in the sensing systems and spec  res is resulted from the spectral resolution of the optical spectrometer. In our sensor design, we use the optical refractometer with a resolution of 20 pm, the detection limit of our sensor is calculated to be 2×10-4, compared with a detection limit of 1.78×10-5 of single microring resonator sensor [35]. The sensitivity of the ethanol sensor is calculated to be SEtOH  6000 (nm/ RIU) and detection limit is 1.3×10-5. It is noted that silicon waveguides are highly sensitive to temperature fluctuations due to the high thermo-optic coefficient (TOC) of silicon ( TOCSi  1.86 104 K 1 ). As a result, the sensing performance will be affected due to the phase drift. In order to overcome the effect of the temperature and phase fluctuations, we can use some approaches including of both active and passive methods. For example, the local heating of silicon itself to dynamically compensate for any temperature fluctuations [36], material cladding with negative thermo-optic coefficient [37-40], MZI cascading intensity interrogation [14], control of the thermal drift by tailoring the degree of optical confinement in silicon waveguides with different waveguide widths [41], ultra-thin silicon waveguides [42] can be used for reducing the thermal drift. 4.5. Three-Parameter Sensor Based on 6×6 MMI and Resonator Structure The proposed sensor based on 6×6 multimode interference and microring resonator structures is shown in Fig. 4.11 [9]. The two MMI couplers are identical. The two 6×6 MMI couplers have the same width WMMI and length L MMI . In this structure, there are three sensing windows having lengths L a1 , L a 2 , L a 3 . As with the conventional MZI sensor device, segments of four MZI arms having lengths L a1 , L a 2 , L a 3 overlap with the flow channel, forming three separate sensing regions. The other three MZI arms isolated from the analyte by the micro fluidic circuit’s substrate. If we choose the MMI coupler having a length of LMMI  3L , where L  is the beat 2  ; the MMI coupler is characterized by a transfer 1  2 matrix M. We can prove that the overall transfer matrix S of both the MMI coupler and combiner in Fig. 4.11 is expressed by length of the MMI coupler, L   114 Chapter 4. Multi-Parameter Integrated Optical Sensor Based on Multimode Interference and Microring Resonator Structures Fig. 4.11. Schematic of the new sensor using 6×6 MMI couplers and microring resonators. Four arms of the MZI is exposed to the analyte within the interaction regions  j e 4   0   1  0 S=  2 0    0  3 e j 4  0 e j 0  4 0 0 0 0 0  j e4 3 j e 4 j e 3 4 e j  4 0 e j 3 4 e j 3 4 0 0 0 0 0 0 3 j e 4 0 0  j e4 0 0 0 0 e j  4        .        (4.17) This matrix can be considered as consisting of four separate sub-matrices which describe four 2×2 3 dB MMI couplers, both having the transfer matrix  j 1 e 4 M2   2  j 3 4 e e j e 3 4 j  4    1 j 4 1 j  e  j 1 .  2     (4.18) Relations between the complex amplitudes a 1 , a 2 ,..., a 6 at the input ports and d1 , d 2 ,..., d 6 at the output ports can be expressed in terms of the transfer matrices of the 3 dB MMI couplers and the phase shifters as follows 115 Advances in Optics: Reviews. Book Series, Vol. 3 1 2  1  *  1 1   a1    , 1*   a 6  (4.19) 2 2  2  *  2  2  a 2    , *2   a 5  (4.20) 1 2  3  *  3 3   a 3    , *3   a 4  (4.21) j  d1   d   je  6 j d 2   d   je  5 j  d3   d   je  4     1  ), 1  cos( 1 ) ; 2  sin( 2 ), 2  cos( 2 ) ; 3  sin( 3 ), 3  cos( 3 ) ; 2 2 2 2 2 2 1 ,  2 ,  3 are the phase differences between two arms of the MZI, respectively. where 1  sin( One can prove that the normalized optical powers transmitted through the proposed sensor at wavelengths on resonance with the microring resonators are given by 2 T1  T2  T3  d1 a1 d2 a2 d3 a3 2 2 2  1   1  cos( 2 )   ,  1   1  1 cos( 2 )    (4.22) 2  2   2  cos( 2 )  2   1   2 cos( 2 )     ,     3   3  cos( 2 )  3   1   3 cos( 2 )     ,    (4.23) 2 (4.24) where 1 ,  2 , and  3 are round trip transmissions of light propagation through the four microring resonators [27] depending the losses of light propagation from output ports d 4, d5 , d 6 back to input ports a 4, a 5 , a 5 ; for a lossless resonator   1 . The proposed structure can be viewed as a sensor with four channel sensing windows, which are separated with four power transmission characteristics T1 , T2 , and T3 and four sensitivities S1 , S2 and S3 . This means that the proposed sensor is able to monitor four target chemicals simultaneously. Their sensitivities can be expressed by: 116 Chapter 4. Multi-Parameter Integrated Optical Sensor Based on Multimode Interference and Microring Resonator Structures S1  T T1 T , S2  2 , S3  3 . n c n c n c (4.25) Fig. 4.12 compares the normalized transmission for the proposed sensor with 1  0.99 and 0.90 to that for the conventional MZI, as functions of the total relative phase  . Given that the sensitivity is linearly proportional to the slope of the power transfer characteristics. Fig. 4.12 shows that the proposed sensor should have a higher sensitivity to a change in the refractive index of the analyte than the conventional MZI, when biased for operation with the region of large slope near 1  0 . Fig. 4.12. Normalized optical transmissions as functions of total relative phase for the proposed sensor with 1  0.99 and 0.90 and conventional MZI sensor. It is now shown that the proposed sensor can be realized using silicon nanowire waveguides. The width of the MMI is WMMI  8.4 m and the core thickness is h co = 220 nm. The access waveguide is tapered to a width of 0.8 µm to improve device performance. It is assumed that the designs are for the transverse electric (TE) polarization at a central optical wavelength   1550 nm . The first 6×6 MMI coupler is now optimized by using the 3D BPM. Fig. 4.13 (a) shows the normalized output powers at the bar and cross ports at different MMI lengths for a signal presented at input port 1 of the MMI coupler. From this simulation result, the optimized length of MMI calculated to be L MMI  273.5 m . The field propagation through the 6×6 MMI coupler at this optimized length is plotted in Fig. 4.13 (b). The relation between the effective index n eff ,a and the ambient index or cladding index n analyte  n c is achieved by using the beam propagation method (BPM). From this 117 Advances in Optics: Reviews. Book Series, Vol. 3 relationship, we achieve the waveguide sensitivity factor n eff ,a n c . Fig. 4.14 shows the effective index change n eff ,a due to the ambient change for silicon nanowire waveguides having a width of 500 nm. We can see that effective index n eff ,a increases almost linearly in the change in the refractive index of ambient material, i.e., the waveguide sensitivity factor is almost a constant. (a) Normalized output powers vs MMI length. (b) Field propagation. Fig. 4.13. BPM simulation results: (a) Normalized output powers vs the length of the 6×6 MMI coupler, and (b) field propagation at the optimized MMI length. From the simulation results of Fig. 4.14, the sensitivities of the proposed sensor and the conventional MZI with the active region length of L a  100 m and L a  500 m are plotted in Fig. 4.15. The simulations obviously show that the sensitivity of the proposed sensor is much higher than the sensitivity of the conventional MZI sensor. 4.6. Four-Parameter Sensor Based on 8×8 MMI and Resonator Structure The proposed sensor based on 8×8 multimode interference and microring resonator structures is shown in Fig. 4.16 [43]. The two 8×8 MMI couplers have the same width WMMI and length L MMI . There are four sensing windows having lengths L a1 , L a 2 , L a 3 , L a 4 . As with the conventional MZI sensor device, segments of four MZI arms having lengths L a1 , L a 2 , L a 3 , L a 4 overlap with the flow channel, forming four separate sensing regions. As a result, this structure can be used to detect four chemical or analyates at the same time. If we choose the MMI coupler having a length of L 2  3L  / 4 , the overall transfer matrix S of both the MMI coupler and combiner of length L MMI  3L  / 2 is expressed by [43] 118 Chapter 4. Multi-Parameter Integrated Optical Sensor Based on Multimode Interference and Microring Resonator Structures (a) (b) (c) Fig. 4.14. (a) The change of the effective index as the increase of refractive index of the analyte for silicon nanowire waveguides; (b) optical field profile for n analyte  1.33 and (c) optical field profile for n analyte  1.34 . Fig. 4.15. Sensitivity of the proposed sensor for sensing window S1 and the conventional MZI sensor versus the round trip loss of the first microring resonator. 119 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 4.16. Schematic of the new sensor using 8×8 MMI couplers and microring resonators.  j e 4   0    0   1  0 S= 2  0   0    0  3  j4 e 0 e j  4 0 3 4 0 0 0 0 0 0 0 0 0 e j 3 4 e j 0 0  j e4 0 0  j e4 0 0  j e4 3 j e 4 0 0 0 0 0 e 0 0 0 0 3 j e 4  j e4 0 0 0 0 j 3 4 0 e j  4 0 3 j e 4 0 0 0 0  j e4 0 0 0 0 0 0 e j  4          .           (4.26) The 3D-BPM simulations for optimised designs of 8×8 MMI structures based on an SOI channel waveguide having a width of WMMI  9 m are shown in Fig. 4.17. The optimised length calculated to be L MMI  382 m . It is note that the complete device is also equivalent to four separate 2×2 MMI-based microresonators. Each microresonator may have different transmission characteristics such as different quality factor (Q), different free spectral range (FSR) and different bandwidth. The 3D-BPM simulations show that the device performs the functions as predicted by the theory. However, when the signal is applied to input port 1, then 3D-BPM simulations show that at the optimised length of L MMI  382 m , the computed excess loss is 1.08 dB and the imbalance is 0.11 dB. The normalized output powers at the bar and cross ports at different MMI lengths for a signal at input port 1 are shown in Fig. 4.18. 120 Chapter 4. Multi-Parameter Integrated Optical Sensor Based on Multimode Interference and Microring Resonator Structures (a) (b) Fig. 4.17. 3D-BPM simulations of an 8×8 MMI structure used in a microresonator for two cases (a) the signal entered at input port 1, and (b) signal entered at input port 2. Fig. 4.18. Normalized output powers at the bar and cross ports as functions of the MMI length for the signal at input port 1. The complex amplitudes at the output ports of the sensor structure in Fig. 4.16 can be expressed by    j  1 1   a1   c1  1 j 4 1 j e j 0  1 j 4 1 j  a1  2 e e je   (4.27)  *  c  ,  j 1   j 1 a  * 2    0 1 2   8  8  1 1  a 8  1 1  c2  1 j 4 1 j e j e  c   j 1  2   0  7 2  j 1 j j  a 2  4 e   j 1 a   je 1 2   7 0 1 2 2  2  *  2  2  a 2    , *2  a 7  (4.28) 121 Advances in Optics: Reviews. Book Series, Vol. 3   c3  1 j 4 1 j e j e  c   j 1  2   0  6   j 1 j 1  3 j  a 3  4 2 e je   *   j 1 a  1 2   6  3  c4  1 j 4 1 j e j e  c   j 1  2   0  5  j 1 j j  a 4  4 e   j 1  a   je 1 2   5 3 0 1 0 1 4 2 2  4  *  4 3   a 3    , *3  a 6   4  a 4    , *4   a 5  (4.29) (4.30) where a  [a1 a 2 a 3 a 4 a 5 a 6 a 7 a 8 ]T is the input field and c  [c1 c 2 c3 c 4 c5 c 6 c 7 c8 ]T 1    ), 1  cos( 1 ); 2  sin( 2 ), 2  cos( 2 ); 2 2 2 2 3 3 4 4 3  sin( ), 3  cos( ) ; 4  sin( ), 4  cos( ) . 1 ,  2 , 3 and  4 are the 2 2 2 2 is the output field and 1  sin( phase differences between two arms of the MZI, respectively. The normalized optical powers transmitted through the proposed sensor at wavelengths on resonance with the microring resonators are given by T1  c1 a1 c T2  2 a2 c T3  3 a3 T4  c4 a4 2 2 2 2 2  1  1  cos( 2 )  1   1  1 cos( 2 )     ,     2   2  cos( 2 )  2   1   2 cos( 2 )     ,     3   3  cos( 2 )  3  1   3 cos( 2 )     ,     4   4  cos( 2 )  4   1   4 cos( 2 )     ,    (4.31) 2 (4.32) 2 (4.33) 2 (4.34) where 1 ,  2 ,  3 , and  4 are round trip transmissions of light propagation through the four microring resonators [27] depending the losses of light propagation from output ports c5, c6 , c7 , c8 back to input ports a 5, a 6 , a 7 , a 8 ; for a lossless resonator   1 . The proposed structure can be viewed as a sensor with four channel sensing windows, which 122 Chapter 4. Multi-Parameter Integrated Optical Sensor Based on Multimode Interference and Microring Resonator Structures are separated with four power transmission characteristics T1 , T2 , T3 , T4 and four sensitivities S1 , S2 , S3 , S4 . This means that the proposed sensor is able to monitor four target chemicals simultaneously. Their sensitivities can be expressed by: S1  T T1 T T , S2  2 , S3  3 , S4  4 . n c n c n c n c (4.35) Fig. 4.19 compares the normalized transmission for the proposed sensor with 1  0.99, 0.98, 0.97 and 0.90 to that for the conventional MZI, as functions of the total relative phase  . Given that the sensitivity is linearly proportional to the slope of the power transfer characteristics. Fig. 4.3 shows that the proposed sensor should have a higher sensitivity to a change in the refractive index of the analyte than the conventional MZI, when biased for operation with the region of large slope near 1  0 . Fig. 4.19. Normalized optical transmissions as functions of total relative phase for the proposed sensor with 1  0.99, 0.98, 0.97 and 0.90 and conventional MZI sensor. Fig. 4.20 shows the effective index change n eff ,a due to the ambient change for silicon nanowire waveguides having a width of 500 nm. From this simulation, one can see that the effective index n eff ,a increases almost linearly in the change in the refractive index of ambient material, i.e., the waveguide sensitivity factor is almost a constant. From the simulation results of Fig. 4.20, the sensitivities of the proposed sensor and the conventional MZI with the active region length of La  50 m , La  100 m and La  500 m are plotted in Fig. 4.21. The simulations obviously show that the sensitivity of the proposed sensor is much higher than the sensitivity of the conventional MZI sensor. 123 n eff ,a Advances in Optics: Reviews. Book Series, Vol. 3 n analyte (a) (b) (c) Fig. 4.20. (a) The change of the effective index as the increase of refractive index of the analyte for silicon nanowire waveguides, (b) optical field profile for n analyte  1.33 and (c) optical field profile for n analyte  1.335 . Fig. 4.21. Sensitivity of the proposed sensor for sensing window S1 and the conventional MZI sensor versus the round trip transmissivity of the first microring resonator. 124 Chapter 4. Multi-Parameter Integrated Optical Sensor Based on Multimode Interference and Microring Resonator Structures 4.7. Conclusions We have presented novel sensor structures based on the integration of 4×4, 6×6 and 8×8 multimode interference structure and microring resonators. The proposed sensor structures can detect two, three and four chemical or biological elements simultaneously. Our sensor structure can be realized on silicon photonics that has advantages of compatibility with CMOS fabrication technology and compactness. It has been shown that our proposed sensors can provide a very high sensitivity compared with the conventional MZI sensor. References [1]. V. M. N. Passaro, F. Dell’Olio, B. Casamassima, et al., Guided-wave optical biosensors, Sensors, Vol. 7, 2007, pp. 508-536. [2]. C. Ciminelli, C. M. Campanella, F. Dell’Olio, et al., Label-free optical resonant sensors for biochemical applications, Progress in Quantum Electronics, Vol. 37, 2013, pp. 51-107. [3]. W. Wang (Ed.), Advances in Chemical Sensors, InTech, 2012. [4]. L. Shi, Y. Xu, W. Tan, et al., Simulation of optical microfiber loop resonators for ambient refractive index sensing, Sensors, Vol. 7, 2007, pp. 689-696. [5]. H. Yi, D. S. Citrin, Z. Zhou, Highly sensitive silicon microring sensor with sharp asymmetrical resonance, Optics Express, Vol. 18, 2010, pp. 2967-2972. [6]. Z. Xia, Y. Chen, Z. Zhou, Dual waveguide coupled microring resonator sensor based on intensity detection, IEEE Journal of Quantum Electronics, Vol. 44, 2008, pp. 100-107. [7]. V. M. Passaro, F. Dell’Olio, F. Leonardis, Ammonia optical sensing by microring resonators, Sensors, Vol. 7, 2007, pp. 2741-2749. [8]. C. Lerma Arce, K. De Vos, T. Claes et al., Silicon-on-insulator microring resonator sensor integrated on an optical fiber facet, IEEE Photonics Technology Letters, Vol. 23, 2011, pp. 890-892. [9]. T.-T. Le, Realization of a multichannel chemical and biological sensor using 6×6 Multimode interference structures, International Journal of Information and Electronics Engineering, Singapore, Vol. 2, 2011, pp. 240-244. [10]. T.-T. Le, Microring resonator based on 3×3 general multimode interference structures using silicon waveguides for highly sensitive sensing and optical communication applications, International Journal of Applied Science and Engineering, Vol. 11, 2013, pp. 31-39. [11]. K. De Vos, J. Girones, T. Claes, et al., Multiplexed antibody detection with an array of Silicon-on-Insulator microring resonators, IEEE Photonics Journal, Vol. 1, 2009, pp. 225-235. [12]. D. Dai, Highly sensitive digital optical sensor based on cascaded high-Q ring-resonators, Optics Express, Vol. 17, 2009, pp. 23817-23822. [13]. A. F. Gavela, D. G. García, C. Jhonattan Ramirez, et al., Last advances in silicon-based optical biosensors, Sensors, 2016, Vol. 16, 285 [14]. X. Han, Y. Shao, X. Han, et al., Athermal optical waveguide microring biosensor with intensity interrogation, Optics Communications, Vol. 356, 2015, pp. 41-48. [15]. Y. Chen, Z. Li, H. Yi, et al., Microring resonator for glucose sensing applications, Frontiers of Optoelectronics in China, Vol. 2, 2009, pp. 304-307. [16]. G.-D. Kim, G.-S. Son, H.-S. Lee, et al., Integrated photonic glucose biosensor using a vertically coupled microring resonator in polymers, Optics Communications, Vol. 281, 2008, pp. 4644-4647. 125 Advances in Optics: Reviews. Book Series, Vol. 3 [17]. C. Errando-Herranz, F. Saharil, A. M. Romero, et al., Integration of microfluidics with grating coupled silicon photonic sensors by one-step combined photopatterning and molding of OSTE, Optics Express, Vol. 21, 2013, pp. 21293-21298. [18]. T.-T. Le, Multimode Interference Structures for Photonic Signal Processing, LAP Lambert Academic Publishing, 2010. [19]. J. M. Heaton, R. M. Jenkins, General matrix theory of self-imaging in multimode interference (MMI) couplers, IEEE Photonics Technology Letters, Vol. 11, 1999, pp. 212-214. [20]. N. S Lagali, The general Mach-Zehnder interferometer using multimode interference coupler for optical communication network, PhD Thesis, University of Alberta, Canada, 2000. [21]. T.-T. Le, L. Cahill, All optical signal processing circuits using multimode interference structures on silicon waveguides, in 4G Wireless Communication Networks: Design Planning and Applications (J. I. Agbinya, M. C. Aguayo-Torres, et al., Eds.), River Publisher, 2013. [22]. A. Yariv, Universal relations for coupling of optical power between microresonators and dielectric waveguides, Electronics Letters, Vol. 36, 2000, pp. 321-322. [23]. D.-T. Le, T.-T. Le, Coupled Resonator Induced Transparency (CRIT) based on interference effect in 4×4 MMI coupler, International Journal of Computer Systems (IJCS), Vol. 4, 2017, pp. 95-98. [24]. T.-T. Le, All-optical Karhunen-Loeve transform using multimode interference structures on silicon nanowires, Journal of Optical Communications, Vol. 32, 2011, pp. 217-220. [25]. L. B. Soldano, E. C. M. Pennings, Optical multi-mode interference devices based on selfimaging:principles and applications, IEEE Journal of Lightwave Technology, Vol. 13, 1995, pp. 615-627. [26]. T.-T. Le, L. Cahill, Generation of two Fano resonances using 4×4 multimode interference structures on silicon waveguides, Optics Communications, Vol. 301-302, 2013, pp. 100-105. [27]. W. Green, R. Lee, G. DeRose, et al., Hybrid InGaAsP-InP Mach-Zehnder racetrack resonator for thermooptic switching and coupling control, Optics Express, Vol. 13, 2005, pp. 1651-1659. [28]. T.-T. Le, L. Cahill, The design of 4×4 multimode interference coupler based microring resonators on an SOI platform, Journal of Telecommunications and Information Technology, Poland, 2009, pp. 98-102. [29]. D.-T. Le, M.-C. Nguyen, T.-T. Le, Fast and slow light enhancement using cascaded microring resonators with the Sagnac reflector, Optik - International Journal for Light and Electron Optics, Vol. 131, 2017, pp. 292–301. [30]. X. Liang, Q. Zhang, H. Jiang, Quantitative reconstruction of refractive index distribution and imaging of glucose concentration by using diffusing light, Applied Optics, Vol. 45, 2006, pp. 8360-8365. [31]. C. Ciminelli, F. Dell’Olio, D. Conteduca, et al., High performance SOI microring resonator for biochemical sensing, Optics & Laser Technology, Vol. 59, 2014, pp. 60-67. [32]. T.-T. Le, Two-channel highly sensitive sensors based on 4×4 multimode interference couplers, Photonic Sensors, 2017, pp. 1-8. [33]. O. A. Marsh, Y. Xiong, W. N. Ye, Slot waveguide ring-assisted Mach–Zehnder interferometer for sensing applications, IEEE Journal of Selected Topics in Quantum Electronics, Vol. 23, 2017, pp. 440-443. [34]. J. Hu, X. Sun, A. Agarwal, et al., Design guidelines for optical resonator biochemical sensors, Journal of the Optical Society of America B, Vol. 26, 2009, pp. 1032-1041. [35]. Y. Chen, Y. L. Ding, and Z. Y. Li, Ethanol sensor based on microring resonator, Advanced Materials Research, Vol. 655-657, 2013, pp. 669-672. [36]. S. Manipatruni, R. K. Dokania, B. Schmidt, et al., Wide temperature range operation of micrometer-scale silicon electro-optic modulators, Optics Letters, Vol. 33, 2008, pp. 2185-2187. 126 Chapter 4. Multi-Parameter Integrated Optical Sensor Based on Multimode Interference and Microring Resonator Structures [37]. M. Han, A. Wang, Temperature compensation of optical microresonators using a surface layer with negative thermo-optic coefficient, Optics Letters, Vol. 32, 2007, pp. 1800-1802. [38]. K. B. Gylfason, A. M. Romero, H. Sohlström, Reducing the temperature sensitivity of SOI waveguide-based biosensors, Proceedings of SPIE, Vol. 8431, 2012, pp. 84310F (1-15). [39]. C.-T. Wang, C.-Y. Wang, J.-H. Yu, et al., Highly sensitive optical temperature sensor based on a SiN micro-ring resonator with liquid crystal cladding, Optics Express, Vol. 24, 2016, pp. 1002-1007. [40]. F. Qiu, F. Yu, A. M. Spring, et al., Athermal silicon nitride ring resonator by photobleaching of Disperse Red 1-doped poly (methyl methacrylate) polymer, Optics Letters, Vol. 37, 2012, pp. 4086-4088. [41]. B. Guha, B. B. C. Kyotoku, M. Lipson, CMOS-compatible athermal silicon microring resonators, Optics Express, Vol. 18, 2010, pp. 3487-3493. [42]. S. T. Fard, V. Donzella, S. A. Schmidt, et al., Performance of ultra-thin SOI-based resonators for sensing applications, Optics Express, Vol. 22, 2014, pp. 14166-14179. [43]. T.-T. Le, The design of microresonators based on multimode interference couplers on an SOI platform, in Proceedings of the 3rd International Conference on Communications and Electronics (ICCE’10), Nha Trang, Vietnam, 11-13 August, 2010. 127 Chapter 5. Coherent Gradient Sensor for Curvature Measurement in Extreme Environments Chapter 5 Coherent Gradient Sensor for Curvature Measurement in Extreme Environments Cong Liu, Xingyi Zhang and Youhe Zhou1 5.1. Introduction Due to the merits of the real-time, full-field, non-contact and non-intrusive, the coherent gradient sensor (CGS), as a lateral shearing interferometry, has a wide application in measurements of gradients [1] and curvatures [2-5], especially in the deformation state of field of crack tip [6-8] at room temperature or thin solid film-substrate structure in high temperature with an ignored air refraction [9]. On the contrary, with decrease of the environment temperature, the influences of air refraction cannot be ignored, so the investigation on the CGS at cryogenic temperature [10] is necessary and important for its practical applications [11, 12]. In the first section of this chapter, we will briefly introduce the measurement theory of the CGS method. And in the second section, the error factor dependent on air refraction at low temperature is discussed, as well as the effects of transparent interface reflection and refraction on the interferogram. Some thin filmstructures always work in the multi-media, thus in the third section, we present the study of the effect of refraction of multi-layer media and we suggest a modification factor to eliminate the difference between the experimental measurement and the actual value. Subsequently, the modification factor is verified through gradients and curvatures measurements of specimen immerged into water and silicone oil, respectively. Finally, to process the sparse fringes [13], often encountered in measurements of small gradients or curvatures, an arbitrary integral-multiple fringe multiplication method is proposed and testified by optical-elastic experimental results, this is accounted in the fourth last section of this chapter. Cong Liu Department of Mechanics and Engineering Sciences, College of Civil Engineering and Mechanics, Key Laboratory of Mechanics on Disaster and Environment in Western China attached to the Ministry of Education of China, Lanzhou University, Lanzhou, Gansu, PR China 129 Advances in Optics: Reviews. Book Series, Vol. 3 5.2. The CGS System Fig. 5.1 displays the schematic of the CGS system. The surface of film is illuminated by collimated laser beam through the reflection of beam splitter (Fig. 5.1(a)). The reflective plane wave which includes information of deformation state of film is passed into the CGS system (Fig. 5.1(b)). The surface can be expressed as a function of (x, y). z  f ( x, y ) or F ( x, y, z )  z  f ( x, y )  0 . (5.1) Fig. 5.1. Schematic of the CGS. The normal vector of the reflective surface is equal to the following equation: N F  f x ex  f y e y  e z  . F 1  fx2  f y2 The reflective plane wave propagation vector of wave plane d can be described: 130 (5.2) Chapter 5. Coherent Gradient Sensor for Curvature Measurement in Extreme Environments d  (2e z  N ) N  e z   e x   e y   e z  where:   2(  f x e x  f y e y  e z ) 1  fx2  f y2  ez , (5.3) 1  f x2  f y2 2 f y 2 f x   , , .  1  f x2  f y2 1  f x2  f y2 1  f x2  f y2 The propagation vector of wave plane d becomes d 1 , d0 , and d 1 after diffraction from grating G1. At this situation, d0  d , d1  1d0 , (5.4) where  1 is the rotation tensor whose components are given by: 0 1  1   0 cos 0  sin  0   sin   . cos  (5.5) Form the equation (5.3) to (5.5), one can gain: d1  [ e x  ( cos   sin  )e y  ( cos   sin  )e z ] , (5.6) OA  d1  e z  OA ( cos   sin  )   , (5.7) OB  d0  e z  OB    , (5.8) OC  d1  e z  OC ( cos   sin  )   . (5.9) At the front the second grating G2, the wave equation is displayed as following: E1  a1 exp[i ( kd1  OA  kd1  x)]  a1 exp[i ( k (  )  kd1  x)] ,  cos   sin  E0  a0 exp[i (kd 0  OB  kd 0  x)]  a0 exp[i (k E1  a1 exp[i ( kd 1  OC  kd 1  x)]  a1 exp[i ( k    kd 0  x)] ,   cos    sin   kd 1  x)] . (5.10) (5.11) (5.12) Then the second diffraction happens after passing through the grating G2. The light becomes E(1,1) , E(1,0) , E(1,1) , E(0,1) , E(0,0) , E(0,1) , E( 1,1) , E( 1,0) , and E( 1,1) (Fig. 5.1(b)). The propagation vectors of E(1,0) and E(0,1) are equal to d1 . These E(1,1) , E(0,0) and 131 Advances in Optics: Reviews. Book Series, Vol. 3 E( 1,1) are equal to d0 , and these E(0,1) and E(1,0) are d 1 . E(1,0) , E(0,1) and E(1,1) , E(0,0) , E( 1,1) and E(0,1) , E(1,0) can make interference fringes respectively. However, only interferometric fringes of E(1,0) , E(0,1) and E(0,1) , E( 1,0) which are considered. Then we can separate these situations into A and B: A. Interference fringes made by E(1,0) and E(0,1) E(1,0)  a1 exp[i ( kd1  OA  kd1  x)]  a1 exp[i ( k (   cos    sin  )  kd1  x)] ,  E(0,1)  a0 exp[i (kd 0  OB  kd1  x)]  a0 exp[i ( k ( )  kd1  x)] . (5.13) (5.14)  The intensity function in image plane is I1 , I 2 , respectively. I1    E(1,0) 2  , I 2    E(0,1) 2  . (5.15) The intensity of E(1,0) , E(0,1) together is shown as the following equation: 2 2 I  I 1  I 2  I 12   [ a1 2  a0 2  a1 a 2 cos( kd 1  x  kd 1  x  k   k  cos    sin  )] , (5.16) which can be written as: I  I c   a1a2  cos{k [  (cos   1)   sin  ]} ,  ( cos    sin  ) (5.17) a12  a22 . Considering the small  , we can use approximation   sin  . 2 The above equation becomes: where I c   I  I c   a1a2 cos( k  2 ). (5.18) Then the fringe distribution is gained while: k  2  2n , n  0, 1, 2, ... B. Interference fringes made by E( 1,0) and E(0,1) : 132 (5.19) Chapter 5. Coherent Gradient Sensor for Curvature Measurement in Extreme Environments E(0,1)  a0 exp[i (kd 0  OB  kd 1  x)]  a0 exp[i ( k E( 1,0)  a1 exp[i ( kd 1  OC  kd 1  x)]  a1 exp[i ( k    kd 1  x)] ,   cos    sin  (5.20)  kd 1  x)] , (5.21) also one can get: 2 2 I  I 1  I 2  I 12   [ a1 2  a0 2  a1 a 2 cos( kd 1  x  kd 1  x  k   k  cos    sin  )] , (5.22) which can be written as I  I c   a1a2  cos{k [  (cos   1)   sin  ]} ,  ( cos    sin  ) (5.23) a12  a22 . Same as case A, we gain the equation (5.18) once again. Then we 2 prove the equivalence of these two situations. where I c   Substituting  ,  ,  of the Eq. (5.3) into the Eq. (5.19) and considering that 2  k , p  , one can get:   2 (1   f ) 2 np , n  0, 1, 2,...... . fy  [ ] 2 2 1  f (5.24) When the principal direction is oriented coinciding with the x-axis, one can gain: 2 fx  [ (1   f ) 2 mp , m  0, 1, 2,...... . ] 2 2 1  f (5.25) The curvatures of film can be expressed as k xx  k yy  k xy  f xx 2 x 1 f  f 2 y f yy 2 x 1 f  f 2 y f xy 2 x 1 f  f 2 y    f xx 1  f 2 f yy 1  f 2 f xy 1  f 2 , (5.26) , (5.27) . (5.28) 133 Advances in Optics: Reviews. Book Series, Vol. 3 2 Assuming f  1 , substituting Eq. (5.24) and (5.25) into (5.26)-(5.28), the relationships between fringes and curvatures can be showed as followings: kxx  k xx   2 f ( x, y ) p n ( x )  ( ), x 2 2 x (5.29) kyy  2 f (x, y) p n( y) ),  ( 2 y y2 (5.30) p n( y ) 2 f ( x, y) p n( x) ) ( ).  ( 2 y 2 x xy (5.31) 5.3. Curvature Measurements in Cryogenic Medium 5.3.1. Governing Equations When measurement of the curvature of film-substrate structure is operated at low temperature, the variation of refractive index caused by temperature change should be considered. Fig. 5.2 displays the interface refraction at different temperature. Generally, the relationship between the refractive index and temperature can be displayed as: n1 (T )  1  n0  1 , 1  T where  is the constant of 0.00367C 1 . Fig. 5.2. Schematic of the refraction in the interface with different temperature. The propagation vector of the light from the film's surface becomes 134 (5.32) Chapter 5. Coherent Gradient Sensor for Curvature Measurement in Extreme Environments d '  (2ez  N)N  ez  2( f xex  f ye y  ez ) 1  f x2  f y2  ez . (5.33) The propagation vector changes after the light passing though the interface between the air at room temperature (the refractive index is equal to n 2 ) and low temperature (the refractive index is equal to n1). According to the refraction law, one can get: n1  sin 1  n2  sin  2 . (5.34) The incidence vector and refraction vector are coplanar and related by: n1 n d ' (cos  2  1 cos 1 )e z , n2 n2 d where cos 1  d ' e z  1  f 2 1  f 2 sin 1  1  cos12  (5.35) , (5.36) 2 f 1  f 2 . (5.37) From Eq. (5.34) one can gain: sin  2  n1 2 f  n2 1  f 2 , (5.38) 2 cos2  1  sin 2  2 n22 (1  f )2  n12  4 f 2 n2 (1  f ) 2 . (5.39) Substituting Eq. (5.33), Eq. (5.36) and Eq. (5.39) into Eq. (5.35): d   ex   ey   ez  n1 n2 2 d ' ( n 22 (1   f ) 2  n12  4  f 2 n 2 (1   f ) 2  n1 1   f  n2 1   f 2 2 2 )e z , 2 f, y n22 (1  f )2  n12  4 f n1 n1         , , where 2 1  f, x2  f, 2y n2 1  f, x2  f, y2 n2 n2 (1  f ) 2 f, x (5.40) 2 Then substitute  ,  ,  into Eq. (5.19), one can obtain a new description between gradients and fringes: 135 Advances in Optics: Reviews. Book Series, Vol. 3 fy  fx  2 2 2 2 2 2 2 np n2 (1   f )  4 n1  f  2 2 n1 n2 (1   f ) 2 2 2 mp n2 (1  f )  4n1 f  2 2 n1n2 (1  f ) , n  0, 1, 2,... , (5.41) , m  0, 1, 2,... . (5.42) Same as above, we substitute Eq. (5.41) and (5.42) into Eqs. (5.26)-(5.28) with assumption 2 f  1 , the new relationship between fringes and curvatures can be given as k xy  k xx  n2 p n( x ) ( ) , n  0, 1, 2,... ,  n1 2 x (5.43) k yy  n2 p n ( y )  ( ) , n  0, 1, 2,... , n1 2 y (5.44) n2 p n ( y ) n p n ( x ) ( ) 2  ( ) , n  0, 1, 2,... .  n1 2 x n1 2 y (5.45) 5.3.2. Error Analysis 2 Define   max( f ) in the whole plane, then substitute Eq. (5.41) and (5.42) into Eqs. (5.26)-(5.28), one can see: k xx  n22 (1   )2  4n12  p n( x ) ( ), 2 x (5.46)  p n( y ) ( ), 2 y (5.47) n2 (1   )2  4n12 p n( y ) p n( x )  ( ) 2 ( ). 3 2 y 2 x 2 n1n2 (1   ) (5.48) n1n2 (1   ) k yy  n22 (1   )2  4n12 n1n2 (1   ) k xy  n22 (1   )2  4n12 n1n2 (1   ) 3 2 3 2  A Taylor expansion is made on the factor 3 2 n22 (1   )2  4n12 2 n1n2 (1   ) n22 (1   )2  4n12 2 n1n2 (1   ) 136 3 2  3 2 , n2 n22  8n12     ( 2 ) . n1 2n1n2 (5.49) Chapter 5. Coherent Gradient Sensor for Curvature Measurement in Extreme Environments The systemic absolute error limits are described as: ekxx  k xx  k xx  8n12  n22 p n( x ) ( ),   2n1n2 2 x (5.50) ekyy  k yy  k yy  8n12  n22 p n ( y ) ( ),   2n1n2 2 y (5.51) 8n12  n22 8n 2  n22 p n( x ) p n ( y ) ( ) 1 ( ),     2n1n2 2 y 2n1n2 2 x (5.52) ekxy  k xy  k xy  and systemic relative error limit is described by: er*kxx  er*kyy  er*kxy  k xx  k xx k xx  k yy  k yy k xx  k xy  k xy k xx  (4 n12 1  ) . n22 2 (5.53) According to the Eq. (5.53), one can see that if the CGS is used in the same medium, that 7 is to say n1  n2 , the Eq. (5.53) is equal to  . The change of the relative error limit 2 2 2 factor 4n1 n2  1 2 with temperature is displayed in Fig. 5.3. One can see that with the decrease of the ambient temperature, the systematic relative error limit increases with a negative exponential law. In addition, based on the Eqs. (5.43)-(5.45), the curvature of the film, which includes the modifying factor n2 n1 , there are valuable for the two mediums measurements even temperature is not change. 3.508 3.507 3.506 Error Factor 3.505 3.504 3.503 3.502 3.501 3.500 3.499 3.498 50 100 150 200 250 300 350 400 450 500 Temperature (K) Fig. 5.3. The relationship between the temperature and the relative error limit factor. In this figure, the horizontal ordinate is temperature, and the vertical ordinate is equal to the factor 4n1 n2  1 2 , 2 2 where n2 is the constant and equal to the refractive index at 300K of the light, n1 varies with temperature change. 137 Advances in Optics: Reviews. Book Series, Vol. 3 5.3.3. Curvature Measurement in Cryogenic Vacuum Chamber In practice, CGS needs to be utilized in a vacuum chamber with temperature variation as showed in Fig. 5.4. Here we take emphasis on the technique how to eliminate the influences of interface on the interferogram. At first, in order to eliminate the disturbance of the reflected beam from the upper surface of the transparent window, a tilt of the transparent window will be conducted, which is illustrated in Fig. 5.5. Fig. 5.4. (a) Schematic of the CGS for cryogenic temperature; (b) photo of the measurement system, in which the number 1 denotes the closed cycle refrigerator (G-M). Fig. 5.5. (a) schematic of the lean of transparent windows,  is the angle between the window and the horizontal plane, l denotes the width of the beam splitter, and h is the distance between the project plane of the transparent window and the bottom surface of the beam splitter, and while the thickness of quartz window is neglected; (b) Schematic of the effects of the quartz window on the interferogram, the dotted lines denote the normal of the window’s surface. In Fig. 5.5 (a), one can see that in order to avoid the reflected light of the transparent window into the CGS system, the angle  between the project plane of the transparent 1  tan 2  h window and the bottom surface of the beam splitter should be satisfied by  . 2 tan  l 138 Chapter 5. Coherent Gradient Sensor for Curvature Measurement in Extreme Environments One can get that the two roots as tan  h h2  2  1 , respectively. Considering the small l l  and h  l , we can obtain that tan  is equal to approximation   tan , thus,  is equal to h h2  2  1 , we can use l l h h2  2  1 approximately. According to l l the law of refraction, one can easily see that there is no effect of the quartz window on the light propagation vector, which has effect on the location of the interferogram only. Thus, the change of the incidence vector will be decided by the air (its refractive index is equal to n 2 ) and the vacuum (the refractive index is equal to n1 ). Based on our previous derivation, the propagation vector of the light from the window's surface is satisfied by n1  n2 , we can obtain d  d' without any modification. 5.4. Curvature Measurements in Multiple Media 5.4.1. Refraction Analysis Taking account of multilayer mediums condition as shown in Fig. 5.6 (a), CGS system is developed to measure the gradient or curvature of a sample in multilayer of mediums. We can derive the equivalent situation by the refraction law:   sinsin sin  m sin  2 sin  3    sin  1 sin  1 sin  2 m m 1  n1 , ( m  2) . nm (5.54) Fig. 5.6. (a) Schematic of the effects of the multilayer medium on the reflection, N denotes the normal line of the specimen surface; (b) the situation which is equal to (a). 139 Advances in Optics: Reviews. Book Series, Vol. 3 It is noted that only the first and last layer of medium have effects on refraction of the light reflecting from the sample surface. Thus we can simplify this situation into just two types of mediums as shown in Fig. 5.6 (b). Without loss generality, The CGS for two types of mediums measurement is illustrated in Fig. 5.7 (a). A collimated laser beam passes through a beam splitter (medium 2) and then directly arrives at the reflecting specimen surface (medium 1). The reflected beam from the specimen is further reflected by the beam splitter and then passes through two Ronchi gratings, G1 and G2 with the same density (50 lines/mm) separated by a distance  . The diffracted beams from the two gratings are converged to interfere using a lens. Either of the ±1 diffraction orders is filtered by the filtering aperture to obtain the interferogram recorded by a CCD camera. Fig. 5.7 (b) is displayed schematic of the refraction, for medium 1, its refractive index is equal to n1 , and for the medium 2, the refractive index equals to n2 . In this scenario, the modified gradients and curvatures can be acquired as same as Eqs. (5.41)-(5.45). Fig. 5.7. (a) Schematic of the CGS system for two types of medium; (b). Schematic of the effects of the different mediums on the interferogram, for the medium 1, refractive index is equal to n1 , and for medium 2, the refractive index is n2 . N denotes the normal line of the specimen surface. 5.4.2. Experiment Verification In order to verify the modified factor experimentally, a standard spherical mirror with diameter of 16 mm and radius of curvature of 8 m is used in the experiments. The CGS system is located in the air (medium 2) and the spherical mirror is immerged into the liquid (medium 1). In this work, the liquid is selected by water and silicone oil. The refractive indexes of the air, water and silicone oil are 1, 1.3 and 1.4, respectively. 140 Chapter 5. Coherent Gradient Sensor for Curvature Measurement in Extreme Environments For the spherical 2 2 mirror, the function of surface can be expressed as 2 f ( x, y )  R  x  y , the gradient (x component) of center line (along x direction) has a theoretical solution as following: f x ( x, y ) y 0  x y 0 64  x 2  y 2 , ( 0.008  x  0.008 ). (5.55) 2 According to the CGS system (single medium) and with the assumption f  1 , the gradient (x component) of center line of the spherical mirror can be obtained by: f x ( x, y ) y 0  p  ( x, y) 4 y 0 , (5.56) where   x, y  denotes the phase angle, and based on the above theoretical analysis, when the spherical mirror is immerged into the liquid whose refractive index is not equal to that of the air, its gradient (x component) should be expressed as: n1  f x ( x, y ) n2 y 0  p  ( x, y ) 4 y 0 , (5.57) where n1 , n2 denote the refractive indexes of the liquid and air, respectively. The distances  between the two Ronchi gratings G1 and G2 in the CGS system are selected by 25, 30, 35, and 40 mm, respectively. When  is equal to 25 mm, the red fringes in Figs. 5.8 (a), 5.8 (b) and 5.8 (c) represent the contour curves of the specimen surface slope in x direction within three different mediums. Fig. 5.8 (a) is obtained when the specimen is placed in the air, and Figs. 5.8 (b) and 5.8 (c) are obtained when the specimen is immerged into water and silicone oil, while the CGS system is located in the air. The wrapped phase map is calculated by FFT method and shown in Figs. 5.8 (d), 5.8 (e) and 5.8 (f). The theoretical deformation gradients (left part of Eq. (5.57)) and experimental measurements (right part of Eq. (5.57)) are displayed in Fig. 5.9. One can see that the experimental results in three conditions have a favorable comparison with the theoretical analysis except points located at the vicinity of the specimen edge. According to these results, we come to a conclusion that the experimental results gain good agreement with the gradient enlarged by the factor n1 n2 . For this reason, we need to add a modified factor n2 n1 if we want to get the real gradient of the surface. Now, we will verify the modification factor of curvatures by the same experiments. Table 5.1 show the experimental curvatures when the standard spherical mirror is immerged into the air, water and silicone oil. For the spherical mirror used in this paper, the curvature of  xx is equal to 0.125 m-1. One can see that the modified curvatures have little difference comparing with the standard value when the spherical mirror is immerged into water and silicone oil. 141 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 5.8. Interferogram fringes in x direction and their wrapped phase maps: (a) obtained in the air, (b) obtained in the water, (c) obtained in the silicone oil; (d) wrapped phase map for (a), (e) wrapped phase map for (b), (f) wrapped phase map for (c). Fig. 5.9. Comparison of the experimental results with different  and theoretical calculations of the deformation gradients of the center line (x direction) for the standard spherical mirror. (a) Experiment is carried out in the air; (b) Experiment is carried out in the water, (c) Experiment is carried out in the silicone oil. 142 Chapter 5. Coherent Gradient Sensor for Curvature Measurement in Extreme Environments Table 5.1. List of the experimental results and their modification by n2 n1 in different mediums. Grating space ( ) Experimental curvature (in air) Experimental curvature (in water) Corrected curvature (in water) 25 mm 30 mm 35 mm 40 mm 0.1245 m-1 0.1228 m-1 0.1263 m-1 0.1221 m-1 0.1600 m-1 0.1623 m-1 0.1633 m-1 0.1609 m-1 0.1231 m-1 0.1248 m-1 0.1257 m-1 0.1238 m-1 Experimental Curvature (in silicone oil) 0.1776 m-1 0.1772 m-1 0.1794 m-1 0.1786 m-1 Corrected curvature (in silicone oil) 0.1269 m-1 0.1265 m-1 0.1281 m-1 0.1275 m-1 If the CGS system is applied in single medium, wave length of the incident laser has no effect on the curvature measurements. However, the wave length has effect on the refractive index, which can be expressed as: na b  2  c 4 , (5.58) where n,  denote the refractive index and wave length, respectively, a, b and c are the constant related to medium as in the water condition (a = 1.32, b = 3300, c = 3.2e7). Submitting Eq. (5.58) into the Eq. (5.53), one can obtain the influences of the wave length on the factor of system error. One can see that with decrease of the wave length, the system error factor increases as showed in Fig. 5.10. That is to say, when the CGS system is used in two types of mediums even multilayer of mediums, the lager the wave length of incident laser, the weaker is the system error. 6.75 Error factor 6.70 6.65 6.60 6.55 390 420 450 480 510 540 570 600 630 660 690 720 750 Wave length(nm) Fig. 5.10. The relationship between the wave length and the relative error limit factor. In this figure, the horizontal ordinate is wave length, and the vertical ordinate is equal to the factor 4n12 n22  1 2 , where n2 is refractive index of the air, n1 varies with wave length. 143 Advances in Optics: Reviews. Book Series, Vol. 3 5.5. The Multiplication Method for Sparse Interferometric Fringes For an optical ideal interferometric fringe such as CGS fringe, the form is I (r )  A[1  cos( (r ))] , (5.59) where I (r ), r,  (r ) and A denotes the light intensity of the fringe pattern, coordinate vector, phase and amplitude of the fringe variation, respectively. The phase  (r ) contains the desired information, e. g. strain, stress or deformation gradient. The light intensity of the N-fold order multiplied fringe can be expressed as: I N (r )  A[1  cos(N (r ))] , (5.60) where N (r) is the phase distribution of the multiplied pattern. The phase  (r ) before multiplication and N (r) after multiplication can be considered as two parameters in the algorithm, where the relationship between them gives: N (r)  N   (r)  N  arccos( I (r )  1) . A (5.61) This equation constructs a direct correlation between the Eq. (5.59) and Eq. (5.60). Substituting Eq. (5.61) into Eq. (5.60), one finds: I N (r)  A[1  cos( N  arccos( I (r)  1))] . A (5.62) For the non-ideal fringe pattern, the intensity of which can be expressed as I (r )  I1 (r ) 1  cos  (r )   I b (r )  I random , (5.63) where I (r ) is the intensity of the recorded image, Ib (r) is the intensity of background, I1 (r ) is the amplitude of the fringe variation, and I random is the random noise caused by light source and /or digital recording system. According to the description in the second paragraph, the random item can be reduced by using the filter technique, thus the Eq. (5.63) becomes I (r )  I1 (r ) 1  cos  (r )   I b (r ) . (5.64) When the fringe pattern is normalized, the Eq. (5.64) can be rewritten by I (r )  I1  1  cos  (r )   , 144 (5.65) Chapter 5. Coherent Gradient Sensor for Curvature Measurement in Extreme Environments where I1 denotes the constant amplitude that independent from the point position. The form of the Eq. (5.65) is consistent with the Eq. (5.59), thus our method can also be applicable for the no-ideal fringe pattern. Phase determination by Fourier-transform method is an effective way to obtain the phase information in the whole field. This method is applicable for the multiplied fringe pattern, the Fourier transform of the Eq. (5.59) can be written: I (r)  C0  C(r)  C (r) , (5.66) where C0 is the Fourier transform of the constant I1 , C (r) is the Fourier transform of 1 j (r )  I1e and C (r) is the conjugate of C (r) . So we can filter out the DC term C0 and 2  either the components C (r) or C (r) in frequency domain, if we leave the term C (r) , then we can get the wrapped phase with values between  and  :  (r )  arctan Im C (r )  Re C (r )  . (5.67) After that, phase unwrapping can be proceeds iteratively in x and y-direction: n( x1 , y1 )  0 n( x1 , yi 1 ) if  ( x1 , yi )   ( x1 , yi 1 )    n( x1 , yi )  n( x1 , yi 1 )  1 if ( x1 , yi )   ( x1 , yi 1 )   n( x , y )  1 if ( x , y )   ( x , y )   1 i 1 i 1  1 i 1 i  2,3,...,  n( x j 1 , yi ) if  ( x j , yi )   ( x j 1 , yi )    n( x j , yi )   n( x j 1 , yi )  1 if ( x j , yi )   ( x j 1 , yi )     n( x j 1 , yi )  1 if ( x j , yi )   ( x j 1 , yi )   j  2,3,..., (5.68) unwrep ( x j , yi )   ( x j , yi )  2 n( x j , yi ), i,j=1,2,... . Figs. 5.12(a) and 5.12(b) display the frequency domain of the images obtained by the Fourier transform for the original and processed images as showed in Figs. 5.11(a) and 5.11(b), respectively. The plots of amplitude of corresponding frequency along the centerline in the Figs. 5.12(a) and 5.12(b) are shown in Figs. 5.12(c) and 5.12(d), respectively. We can find that the main frequencies of the multiplied fringe image are separated from the DC component, which allow us to use band-pass filtering and phase unwrapping technique to obtain the phase values in the whole field. Fig. 5.13(a) shows the wrapped phase of the 7-fold multiplied fringe. Based on the relationship between the 145 Advances in Optics: Reviews. Book Series, Vol. 3  (r ) and N (r) in the Eq. (5.61), the distribution of the retrieved relative phase is demonstrated in Fig. 5.13 (b), where the reference phase point is located at the center of the image, and the corresponding phase can be set to zero. Fig. 5.13 (c) shows the information for phase error of the whole field, and also indicates that there is no more than 0.07 phase error besides the margin of the picture. Fig. 5.11. (a) The simulated digital image; (b) the 7-fold multiplied fringe pattern; (c) and (d) show the skeleton lines of the (a) and (b), respectively; (e) the intensity of centerline (seen the red line in (a) and 9 (b)). In summary, the presented method for ideal fringe multiplication has clear mathematics algorithm, which can not only provide a way to extract the skeleton lines, but also separate the main frequency form the DC component in the Fourier transform. Furthermore, the presented method has been realized experimentally. As illustrated in Fig. 5.14, the typical photoelasticity experiment has been arranged. In this study, the light source is replaced by a 17-inch LCD in order to provide nearly uniform illumination, and a photoelastic disk composed by the epoxy resin with diameter of 50 mm and thickness of 5.4 mm is used. When the normal loading of 68.6 N is applied, we obtained a sparse fringe pattern with zero and one order fringes. Fig. 5.15 shows our experimental result. 146 Chapter 5. Coherent Gradient Sensor for Curvature Measurement in Extreme Environments Fig. 5.12. (a) and (b) are the frequency domain of the original and the multiplied image respectively ( To increase the contrast, the displayed intensity I  ln 1  I F  is used, where I F is the amplitude of frequency, however, the intensity in center of (b) appears lower than the two neighboring lobes, this optical illusion is mainly caused by picture zooming out in the text and the displaying function I  ln 1  I F  we used. ); (c) and (d) are the frequency amplitude distribution along centerline of frequency map of the original and the multiplied image respectively. Fig. 5.13. (a) The wrapped phase of multiplied fringes pattern; (b) The unwrapped phase N N distribution of the simulated image; (c) The phase error distribution. 147 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 5.14. The arrangement of photoelasticity experiment. Fig. 5.15. The original photoelastic fringe obtained in this experiment. For this kind of sparse fringe pattern, the intensity of the recording fringe image can be expressed as I (r)  I1  1  cos( (r))  I b  I random . (5.69) We assume that I1 and I b in Eq. (5.69) are uniform (independent of the coordinate), and they are only affected by the response of the CCD camera. Moreover, we reduce the item of random noise by smooth filtering method. In order to normalize the fringe pattern, we removed two 1/4 wave plates and the specimen at the first, and rotated the analyzer with unvaried interval of 10 degree while recording the intensity of image, which can be written as I  I (r )  I1  1  cos(2   rot   )  I b , (5.70) where  rot denotes the rotation angle of analyzer. The intensity dependent on rotation angle is displayed by the black curve in Fig. 5.16 (a). According to the experimental curve, we can propose a theoretic equation with the constant background I b  I min and constant amplitude I1 =  I max -I min  2 , where I max and I min are on behalf of the maximum and minimum intensity of the recording image, respectively. Based on Eq. (5.70), results are shown as the red curve in Fig. 5.16(a). It is found that there are some or little differences between the experimental and theoretic results. For eliminating these differences, the 148 Chapter 5. Coherent Gradient Sensor for Curvature Measurement in Extreme Environments experimental intensity can be calibrated by using a polynomial fitting function with the maximum order of two. The comparison between the fitted results and the experimental results is displayed in Fig. 5.16 (b). One can see that the fitting polynomial function adequately capture the characteristics of the calibration results. Fig. 5.16. The process of calibration (a) the experimental intensity dependent on rotation degree of analyzer, (b) the fitting function of calibration. In order to obtain the multiplied image for the sparse photoelastic fringe, following steps have been applied. First, the fringe pattern should be calibrated by using the fitting polynomial function. Second, the background intensity I b can be subtracted directly. Third, the processed fringe pattern is multiplied by the method presented. The normalized fringe patterns without and with calibration are shown in Figs. 5.17(a) and 5.17(d), respectively. Figs. 5.17(b) and 5.17(c) show the 15-fold and 23-fold multiplication results of normalized fringe pattern without calibration while the Figs. 5.17(f) and 5.17(g) depict the 15-fold and 23-fold multiplication results of the one with calibration. As shown in the Figs. 5.17(b) and 5.17(c), the fringes are diseased at the left and right sides of the multiplication image, which can be understood as the contribution of the slow change in the illumination intensity. Moreover, other possible reasons for the diseased fringes after multiplication can be considered as that the fringes at the top and bottom of all the multiplication results are not normalized. The dependence of phase on the resultant difference between two principal stresses for the stress-optical law has the form as (1   2 )r   (r)  f , 2 h (5.71) where  (r ) denotes the real phase distribution of the fringe pattern, namely multiplied (r) N , f is the material fringe value (equals to 18.4 kN / m in this work), and h is the thickness of the photoelastic slice. Fig. 5.18 shows the calculated stresses of the 15fold and 23-fold multiplied fringes of the normalization results with and without calibration (along the white line in Figs. 5.17(b), 5.17(c), 5.17(e) and 5.17(f)). The blue line is on behalf of the theoretical results. We can see that the calculated results based on 149 Advances in Optics: Reviews. Book Series, Vol. 3 the multiplication fringes under calibration (dark cyan line and red line) have a good agreement with the analytical results except the margin of the sample. In comparison, the calculated results based on the multiplication fringes without calibration (orange line and dark yellow line) are much bigger than theoretical results, which can cause many errors of analysis. Fig. 5.17. The multiplication result of normalized fringe pattern with and without calibration, (a) the normalized fringe without calibration (only the area interested is shown with mask); (b) the 15-fold multiplication result of (a); (c) the 23-fold multiplication result of (a); (d) the normalized fringe with calibration; (e) the 15-fold multiplication result of (d); (f) the 23-fold multiplication result of (d). 6 1x10 Theory 15-MC 23-MC 15-MNC 23-MNC 5 8x10 5 Stress(Pa) 6x10 5 4x10 5 2x10 0 5 -2x10 -30 -25 -20 -15 -10 -5 0 5 x(mm) 10 15 20 25 30 Fig. 5.18. Comparison the stress distributions along the white line in the Figs. 5.17 (b), 5.17 (c), 5.17 (e) and 5.17 (f) with theory result, in the figure the curve 15-MC stands for 15-fold multiplication of calibrated normalization, 15-MNC stands for 15-fold multiplication of noncalibrated normalization, so do the 23-MC and 23-MNC. 150 Chapter 5. Coherent Gradient Sensor for Curvature Measurement in Extreme Environments Acknowledgements This work is supported by the Fund of Natural Science Foundation of China (No. 11622217 , 11372121), Innovative Research Group of the National Natural Science Foundation of China (Grant No. 11421062), the National Key Project of Scientific Instrument and Equipment Development (11327802), National Program for Special Support of Top-Notch Young Professionals. This work is also supported by the Fundamental Research Funds for the Central Universities (lzujbky-2017-ot18, lzujbky2017-k18, lzujbky-2016-228) References [1]. H. V. Tippur, S. Krishnaswamy, A. J. Rosakis, A coherent gradient sensor for crack tip deformation measurements: analysis and experimental results, International Journal of Fracture, Vol. 48, 1991, pp. 193-204. [2]. A. Rosakis, R. Singh, Y. Tsuji, E. Kolawa, N. Moore, Full field measurements of curvature using coherent gradient sensing: application to thin film characterization, Thin Solid Films, Vol. 325, 1998, pp. 42-54. [3]. H. Lee, A. J. Rosakis, L. Freund, Full-field optical measurement of curvatures in ultra-thinfilm-substrate systems in the range of geometrically nonlinear deformations, Journal of Applied Physics, Vol. 89, 2001, pp. 6116-6129. [4]. M. A. Brown, T. -S. Park, A. Rosakis, E. Ustundag, Y. Huang, N. Tamura, B. Valek, A comparison of X-ray microdiffraction and coherent gradient sensing in measuring discontinuous curvatures in thin film: substrate systems, Journal of Applied Mechanics, Vol. 73, 2006, pp. 723-729. [5]. M. Budyansky, C. Madormo, J. L. Maciaszek, G. Lykotrafitis, Coherent gradient sensing microscopy (micro-CGS): A microscale curvature detection technique, Optics and Lasers in Engineering, Vol. 49, 2011, pp. 874-879. [6]. S. L. Kramer, M. Mello, G. Ravichandran, K. Bhattacharya, Phase shifting full-field interferometric methods for determination of in-plane tensorial stress, Experimental Mechanics, Vol. 49, 2009, pp. 303-315. [7]. M. Mello, S. Hong, A. Rosakis, Extension of the coherent gradient sensor (CGS) to the combined measurement of in-plane and out-of-plane displacement field gradients, Experimental Mechanics, Vol. 49, 2009, pp. 277-289. [8]. X. Yao, H. Yeh, W. Xu, Fracture investigation at V-notch tip using coherent gradient sensing (CGS), International Journal of Solids and Structures, Vol. 43, 2006, pp. 1189-1200. [9]. X. Dong, X. Feng, K.-C. Hwang, S. Ma, Q. Ma, Full-field measurement of nonuniform stresses of thin films at high temperature, Optics Express, Vol. 19, 2011, pp. 13201-13208. [10]. C. Liu, X. Zhang, J. Zhou, Y. Zhou, A general coherent gradient sensor for film curvature measurements: error analysis without temperature constraint, Optics and Lasers in Engineering, Vol. 51, 2013, pp. 808-812. [11]. C. Liu, X. Zhang, J. Zhou, Y. Zhou, X. Feng, The coherent gradient sensor for film curvature measurements at cryogenic temperature, Optics Express, Vol. 21, 2013, pp. 26352-26362. [12]. X. Zhang, C. Liu, J. Zhou, Y. Zhou, Nonuniform magnetic stresses in high temperature superconducting thin films, Journal of Applied Physics, Vol. 115, 2014, 043911. [13]. C. Liu, X. Zhang, Y. Zhou, Multiplication method for sparse interferometric fringes, Optics Express, Vol. 24, 2016, pp. 7693-7702. 151 Chapter 6. Photo-Emf Sensors and Talbot Effect: Measurement of Displacement and Vibration Chapter 6 Photo-Emf Sensors and Talbot Effect: Measurement of Displacement and Vibration P. Rodriguez, S. Mansurova, N. Korneev and D. Sanchez de la Llave1 6.1. Introduction The near-field diffraction of a periodic object has the special property of repeating itself in intensity at certain propagation distances; this effect is widely known as the Talbot effect [1-3]. In particular, when the object with an amplitude transmittance that is periodic along one axis is illuminated by a monochromatic plane wave (Fig. 6.1), the field distribution at the transmittance plane repeats itself at multiples of the so called Talbot distance. The Talbot distance depends only on two parameters, namely, the illuminating wavelength (λ) and the transmittance spatial period (d). More precisely, for plane wave illumination, the Talbot distance is given by ZT = 2d2∕λ. Grating 1st self‐ image 2st self‐ image Planes of uniform illumination Fig. 6.1. Optical arrangement to observe the Talbot effect or self-imaging phenomenon. P. Rodriguez National Institute for Astrophysics, Optics and Electronics, Puebla, Mexico 153 Advances in Optics: Reviews. Book Series, Vol. 3 It can be observed that between any two consecutive self-images a replica of the grating is also formed but with the contrast inverted, that is, the grating is formed with a lateral displacement of half of the period (d/2). For gratings with an opening ratio of ½ (e. g. Ronchi gratings) there are planes of uniform illumination (i.e. with a visibility equal to zero); the position of these latter planes are given by: Z k  (2k  1) ZT , k  0, 1, 2, ... . 4 (6.1) Hence, by measuring the fringe visibility at an observation plane of interest, the distance between the sinusoidal amplitude grating and the desired observation plane can be unambiguously determined in a distance range of one fourth the Talbot distance, e.g., from z = 0 to z = ZT ∕4. In recent years the Talbot effect and Talbot self-images localization have received much attention, both as a fundamental optical phenomenon and because of its optical applications such as interferometry [4], nanolithography [5], [6] and spectrometry [7, 8]. The Talbot effect also has been exploited for several metrological applications, among them: the measurement of the refractive index [9], the measurement of temperature [10], contouring [11], the measurement of focal length [12], collimation testing [13], wavefront sensors [14, 15] and the measurement of distance and displacement [16-20]. For this latter application, the first proposal [16] relied on measuring the fringe contrast or visibility of the diffracted field intensity and realizing temporal processing of the detected images. Afterward, Schirripa, Spagnolo and Ambrosini proposed a measurement method that used either a cosine or a Ronchi [18] grating and then realized numerical Fourier processing of the detected intensity pattern to determine the distance between the grating and the observation plane. Finally, methods for measuring discrete distances using the Talbot effect were proposed by Metha et al. [19] and Dubey et al. [20]. In [19], two wavelengths were employed to measure a step-height that coincides with the difference between the Talbot distances associated with each of the wavelengths utilized. In [20] a superluminescent diode was employed to provide several wavelengths instead of two. In all of the proposals mentioned above, the fields of interest were detected by a CCD camera and numerical processing of the detected images was required in order to determine its visibility and hence, the propagation distance. Among the image processing methods used to determine the light pattern visibility the root mean square (RMS) method [21], histogram-based method [22, 23] and semivariogram-based method [24, 25] can be mentioned. Despite of their simplicity, all CCD–based methods share common weaknesses, namely, necessity of a filter for avoid the saturation of the camera, sensitivity to the environmental vibrations, and need for additional processing which makes difficult sensing fast changing processes. Recently, the photodetector based on the non-steady-state photo-electromotive force (photo-emf) effect for measuring the visibility and for localizing the Fresnel diffraction patterns (Talbot self-images) generated by a Ronchi grating [26] has been proposed. Photo-emf effect [27] reveals itself as alternating current (ac) induced in a short-circuited 154 Chapter 6. Photo-Emf Sensors and Talbot Effect: Measurement of Displacement and Vibration photoconductive sample by a vibrating non-uniform light distribution. Standard theory developed for the simplest case of sinusoidal light pattern created by the interference of two plane waves predicts, that the output current amplitude is proportional to the square of the light pattern visibility V, which makes the photo-emf based sensors suitable for the direct measurements of the changes on the light pattern contrast. Due to its temporal adaptability to the slow phase drifts [27], this detector possesses additional robustness to the environmental vibrations. For these outstanding properties photo-emf based detector has already been proposed for a number of practical applications. In particular, they are used for detecting vibrations of diffusely scattering objects [28], for measuring the coherent length of several light sources [29], for sensing laser-generated ultrasonic displacements [30], for phase locking of lasers [31], for characterizing femtosecond pulses [32], for the detection of Doppler frequency shift [33] etc. In this chapter we present a comprehensive review of a detection method to measure the displacement and out of plane vibrations of a mirror-like object which combines the Talbot effect with the photo-emf-based detector. The information about displacement is codified in the visibility of the diffracted field; however, in contrast to the previous proposals, the decodification is achieved by direct measurement of the photo-emf current without any image processing. 6.2. Non-Steady-State Photo-Electromotive Force Effect Non-steady-state photo-electromotive force (photo-emf) effect was first described by Petrov et al. [34] in 1990’s. The photo-emf effect manifests itself as an alternating current (ac) through a short-circuited photoconductive sample illuminated by an oscillating nonuniform light distribution. A detailed analysis of photo-emf effect for sinusoidal light distribution can be found in [27] and here it is only briefly discussed. The standard configuration for observing the photo-emf effect is shown in Fig. 6.2. Under illumination by two coherent plane waves, one of which is periodically modulated in phase with the frequency f and amplitude  an oscillating sinusoidal light distribution is created inside photoconductive material: I ( x , t )  I 0 1  V cos Kx   sin(  t )  . (6.2) Here I0 is the average light intensity, K =  is the spatial frequency of the interference pattern, and  is the fringe spacing. This intensity distribution, in its turn, generates a spatially non-uniform distribution of the mobile carriers n(x,t). Mobile charge redistribution due to the diffusion/drift and its subsequent trapping gives rise to a periodical distribution of space charged electric field Esc. The expression for the total current density flowing through the sample is: 155 Advances in Optics: Reviews. Book Series, Vol. 3 J   x , t   e  n  x , t  E sc  x , t   eD  J  n  x, t   E sc  x , t  ,   0 x x (6.3) Lx Ly  EOM V  I2 I1  RL Fig. 6.2. Standard configuration for photo-emf effect. Here I1 and I2 are the intensities of each wave, EOM is the electro-optical modulator, J is the photo-emf current, V is the photo-emf voltage, and RL is the load resistor. where e is the electronic charge,  the mobility, n(x,t) is the density of the mobile photoelectrons in the conduction band, D is the diffusion coefficient,  is the permittivity, and  is the vacuum permittivity. If the Eq. (6.3) is integrated in the interval from 0 to L (the interelectrode spacing), due to the boundary conditions n(0,t) = n(L,t) the averaged diffusion current [the second term of the Eq. (6.3)] caused by the nonuniform photocarriers density, has to be equal to zero:  n  x, t  dx  n  L, t   n  0, t   0. x 0 L  (6.4) Because of the potential nature of the electric field, the averaged displacement current through the sample represented by the third term of the Eq. (6.3) is equal to zero also. Therefore, the density of the total current thought the sample is equal to the drift current:  e J  ( x, t )  E ( x, t ) n( x, t ) dx .  0  (6.5) Under the steady state conditions, the diffusion-driven sinusoidal distribution of Esc is shifted by /4 with respect to the photoconductivity distribution (x), (see Fig. 6.3), therefore the photo-emf current is zero. However, if the intensity distribution shifts, the situation is different. It is assumed that the photoconductivity (x) follows the vibrations of the illuminating intensity distribution 156 Chapter 6. Photo-Emf Sensors and Talbot Effect: Measurement of Displacement and Vibration I(x,t) almost instantaneously, while the distribution Esc(x,t) possesses certain inertia with a response time sc. As a result, the phase shift between mobile carriers concentration distributions n(x,t) and space charge field Esc(x,t) is different from /4 and the non-steadystate photo-electromotive force is created giving rise to alternating current through the short-circuited semiconductor. Photoconductive Material (a) Interference pattern (b) Spatial distributions of light intensity I(x) + + (c) Space charge + + +  (x) - (d) Photoconductivity - - -  (x) - - + x x x (e) Space-charge electric field E (x) sc x Fig. 6.3. Diagram of space-charge electric field Esc generation in a photoconductive sample. To calculate the photo-emf current amplitude, first the complex amplitudes of space charge field distribution, as well as the carriers concentration should be found, solving the standard set of equations, described e.g. in [27]. The solution of Eq. (6.5) can be obtained in the approximation of low amplitude of oscillations (<< 1) and low contrast (V << 1) of the illuminating pattern. As result of this solution, the amplitude of the first harmonic of photo-emf current can be expressed as: J   0  i / 0 V 2 E , D 1  i / 0 4 1  K 2 L2D (6.6) here 0 is the average photoconductivity, LD is the diffusion length of the photocarriers, ED=KkBT/e is the diffusion electric field, and the characteristic cutoff frequency  is defined as:   0   sc1   di 1  K 2 L2D  1 , (6.7) 157 Advances in Optics: Reviews. Book Series, Vol. 3 where sc is the recording/erasure time of the electric field grating, and di = is the dielectric relaxation time of the photoconductive material. Note, that the frequency transfer function of photo-EMF current is similar to the transfer function of RC circuit, i.e. at low frequency the signal grows up to the cutoff frequency , and then has a constant level at frequencies higher than this cutoff frequency (see Fig. 6.4).  Photo‐emf signal J (a.u.) 10 1 0,1 0,01 0,01 0,1 1 10 100 Modulation frequency  (a.u.) Fig. 6.4. Dependence of the first harmonic of photo-emf current J versus the modulation frequency . This kind of transfer function is responsible for the inherent adaptive properties of the photo-emf effect for detection of phase-modulated signals. By the nature of the photoemf, the sensors based on this effect have the ability to compensate slow environmental phase drifts, i.e. possess additional robustness to the environmental pertubations. In addition, these photo-emf sensors possess spatial adaptability to the wavefront irregularities in the interfering beams, since the space charge field and concentration distribution are the exact replica of the intensity distributions, i.e. the detection can be performed in presence of speckle of the illuminating light pattern or aberrations of the interfering beams. For these outstanding properties, the photo-emf based sensors have already been proposed for a number of practical applications as an adaptive detector of phase modulated signal [28-35]. For the experiments presented here, the photo-emf sensor was fabricated from a piece of GaAs:Cr crystal with the dimensions 8 mm x 5 mm x 0.5 mm glueded to a plastic substrate. Two electrodes were deposited on its front surface with silver paint, in such a way that they present an effective interelectrode surface with the dimensions Lx = L ≈ 5 mm and Ly = 5 mm (see Fig. 6.2). The photo-emf signal was measured by connecting the silver electrodes to a lock-in amplifier by means of a coaxial cable. 158 Chapter 6. Photo-Emf Sensors and Talbot Effect: Measurement of Displacement and Vibration 6.3. Applications 6.3.1. Measurements of Displacements A novel approach for measuring distances or displacements using detectors based on the photo-emf sensors was recently demonstrated [36]. Similar to previous techniques, the information about the displacement of the test object is codified in the visibility of the diffracted light by a diffraction grating. The decodification, however, is carried out by direct measurement of the photo-emf electrical current, which is proportional to the square of the visibility of the light pattern [see Eq. (6.6)]; unlike the previous techniques there is no any image processing involved. Analysis of this novel proposal is carried out as follows [26]. Consider a one-dimensional sinusoidal amplitude grating, with the transmission described by following expression: t( x)  1 2   2x  1  m cos d  ,    (6.8) where d is the grating period and m is the grating modulation index. For a plane wave illumination, the near-field intensity distribution at axial distance z from the grating position is given by [37]: I ( x)    2z   2x  1 2 2  2x   cos   m cos   . 1  2m cos 4   d   ZT   d  (6.9) For low values of the grating modulation index (m << 1), it is straightforward to show that the visibility of the diffracted light pattern is  2 z   . V ( z )  2m cos  ZT  (6.10) Therefore by measuring the visibility of the diffracted light pattern, the distance from the grating to the plane of interest (z) can be unambiguously determined in a range of ZT/4. On the other hand, as stated in Section 6.2, the photo-emf current amplitude [Eq. (6.6)] at angular frequency =2f can be written as: J   CI0  V 2 , (6.11) where the factor C groups together the factors that depend on electro-optical parameters of the sample and on spatial (K) and temporal frequencies () of the illuminating pattern, and I0 is the average intensity of the light pattern. 159 Advances in Optics: Reviews. Book Series, Vol. 3 If the sinusoidal amplitude grating described by Eq. (6.8) is set to oscillate (at frequency f) on the direction of its grating vector, the axial dependence of the photo-emf current induced in the photo-emf sensor by the light diffracted on the oscillating grating, is obtained by substitution of Eq. (6.10) into Eq. (6.11):  1 1  2 z   J   4 m 2 CI 0    cos   ,  2 2  ( Z T / 2)   (6.12) where  is the amplitude of the oscillations of the grating which is assumed to be smaller than the period of the grating (<< d). It follows from this latter equation that by measuring this photo-emf current, the distance between the oscillating sinusoidal grating and the photo-emf sensor can be determined in a range of ZT/4 in a linear arrangement. To determine the displacement of a mirror-like object using the photo-emf sensors an optical setup depicted in Fig. 6.5 has been proposed. Translation stage Grating B.S. Collimated beam Lock‐in amplifier  V Piezo electric  S.G. Photo‐emf sensor Fig. 6.5. Experimental setup to measure displacements of a test object with mirror-like surface employing the photo-emf sensor. The test object is mounted on a translation stage. In the setup, a sinusoidal amplitude grating is illuminated by a collimated He-Ne laser beam. The light diffracted by the grating is directed to the test object by a beam splitter (BS) and the light reflected from the mirror-like surface of the object is brought to the surface of the photo-emf sensor. Oscillations are induced to the grating by gluing it to a piezoelectric transducer excited by a signal generator. The period of the sinusoidal amplitude grating used in the experiment was d = 0.1 mm, which in combination with the wavelength of the laser (632.8 nm) results in a Talbot distance ZT = 31.60 mm. The frequency of the oscillations was set at f = 600 Hz, and the excitation voltage was such that the produced amplitude of oscillations was 15 m, much lower than the period of the 160 Chapter 6. Photo-Emf Sensors and Talbot Effect: Measurement of Displacement and Vibration grating. A second He-Ne laser beam was used to provide a background illumination to further decrease the visibility at the surface of the photo-emf sensor described in Section 6.2. The electrical current generated by the photo-emf sensor was measured as voltage drop across the input impedance of the lock-in amplifier, so the measured photoemf signal V can be expressed as:  2 z   , V   V0  VC cos   ZT / 4  (6.13) where V0 is the offset signal, VC is the amplitude of the sinusoidal component. The axial period has been changed accordingly due to the “folded” geometry of the experimental setup. Fig. 6.6 shows the signal from the photo-emf sensor as a function of the displacement of the test object in steps of 0.25 mm. Photo_emf signal, V 1.000 800 600 400 200 0 0 2 4 6 8 10 12 14 Displacement , mm Fig. 6.6. Photo-emf signal as a function of the displacement of the test object. The positions are indicated by the scale of the translational stage. The dots are the experimental values and the solid line is the fit to Eq. (6.13), the fitting values are V0 = 510 V, VC = 416 V and ZT/4 = 8 mm. The light power impinging on the photo-emf sensor from the test object is 0.52 mW and from the background is 0.36 mW. As predicted by the theory [Eq. (6.13)] the photo-emf signal varies as a cosine function of the displacement of the test object. The maximum value of the photo-emf signal corresponds to a test object’s position such that the total distance traveled by the light from the grating to the photo-emf sensor is equal to a multiple of ZT/2, where the visibility of the diffracted light pattern is maximal (11*ZT in this particular configuration). The positions of minimal photo-emf signal correspond to positions of minimal visibility, i.e. to planes of uniform illumination. Note, that as shown in Fig. 6.6 and indicated in Eq. (6.13), around the total propagation distances from the grating to photo-emf sensor equal to z= (2k+1)*(ZT/8), with 161 Advances in Optics: Reviews. Book Series, Vol. 3 k = 0,1,2,…, (the quadrature position), there is a region in which the output signal from the photo-emf sensor is linearly proportional to the displacement of the test object. Fig. 6.7 shows an extended view of the photo-emf signal as a function of 20 m displacements of the test object around the position = 5.25 mm. Experimentally, this relationship holds in a range of 1.5 mm. As far as the power of laser, frequency and amplitude of oscillation of the piezoelectric remain constant, these measurements for determination of the displacement of the test object are quite reproducible, yielding a device with high precision. 6.3.2. Determination of Low Frequency, Out-of-Plane Vibrations Vibration detection systems are widely used in several scientific and technological areas. In the region of high frequency vibrations (greater than 20 kHz) the detection of ultrasound stands out [38]. In the low frequency region (in the order of tens of Hertz) the diagnosis of large civil structures, evaluation of mechanical components and the detection and monitoring of telluric movements are highlighted. Photo‐emf signal,  V 650 600 550 500 450 5,1 5,2 5,3 5,4 5,5 5,6 5,7 5,8 Displacement , mm Fig. 6.7. Photo-emf signal as a function of the displacement of the test object around the position where a linear relation is hold. Optical methods provide different techniques and devices for measuring contactless (i.e. remote) vibrations. Among many others we can mention interferometric and holographic techniques [39], techniques based on self-mixing in laser diodes [40], speckle-based techniques and techniques based in the moire method [41]. In these techniques, the detection and decoding of the signals or images containing the vibration amplitude and frequency of the test object are recorded and stored in a computer for later processing. In interferometric techniques and self-mixing in laser diodes, the use of a photodiode is required; while speckle-based and moire-based techniques require the use of CCD cameras. 162 Chapter 6. Photo-Emf Sensors and Talbot Effect: Measurement of Displacement and Vibration In the following we demonstrate the measuring of low frequency, out-of-plane vibrations of a test object with amplitudes of few microns approximately, using the photo-emf sensors and the Talbot effect. For measuring the amplitude of vibration of a test object with mirror-like surface we used a setup similar that on Fig. 6.5 except for two modifications: the test object (a mirror) was attached to a loudspeaker and the amplified photo-emf signal detected by the lock-in amplifier was displayed and measured by an oscilloscope, as it is depicted in Fig. 6.8. The vibration frequencies of the test object (F) are lower than oscillation frequency of the grating (F << f). If the object under test is vibrating with an amplitude A at the frequency F, the photo-emf signal measured in the oscilloscope UF at frequency F is:  2  z  A sin  2 F    U F  U 0  U C cos  ,   ZT / 4   (6.14) where U0 is the offset signal, UC is the maximum amplitude of the sinusoidal component. F Grating B.S. Collimated beam Lock‐in amplifier Piezo electric UF t Scope S.G. Photo‐emf sensor Fig 6.8. Schematic representation of the optical setup to measure the amplitude of vibration of a test object. Filtering out the offset voltage, placing the photo-emf sensor in a position such that the whole optical arrangement is in the quadrature point, and setting the low amplitudes of vibration A << ZT/8, the linear relation between the amplitude of vibrations and amplitude of the signal at frequency F measured by the oscilloscope is hold: UF  8 U C A . ZT (6.15) 163 Advances in Optics: Reviews. Book Series, Vol. 3 The voltage UC can be considered as a “calibrating” voltage and it can be determined experimentally by moving the object under test a distance larger that ZT/8. As for the previous application the value of the Talbot distance can be also experimentally determined or calculated from the specifications of the grating. So it is evident that by measuring the signal UF the amplitude of vibration of the object under test can be estimated: A ZT UF . 8 U C (6.16) The following experiment presents the determination of the amplitude of vibration of a test object vibrating at a frequency of F = 3 Hz using the proposed method. The experimental conditions were similar to that presented in the previous sub-section. Determination of the “calibrating” voltage and the position where the output signal is linearly proportional to the total distance from the grating to the photo-emf sensor (quadrature position) are obtained by axially translating the test object. Fig. 6.9 shows the oscilloscope trace obtained when the object under test is mechanically translated and it is not vibrating. From this trace and from that (similar to shown in Fig. 6.6) produced by moving the test object in steps of 100 m, the value of the Talbot distance, the “calibrating” voltage and the quadrature positions were obtained. Oscilloscope trace, V 6 5 UC 4 3 2 ZT/4 1 0 ‐6 ‐4 ‐2 0 2 4 6 Time, sec Fig. 6.9. Oscilloscope trace obtained when the test object is mechanically translated along the optical axis. Dashed line is the theoretical fit to Eq. (6.14). Left arrow indicate the quadrature position. The object under test is not vibrating. The measurements of amplitudes of vibration of the test object are performed in the quadrature point (marked with an arrow in the Fig. 6.9). Fig. 6.10 shows the oscilloscope trace when the loudspeaker is excited with a voltage of 1.45 V. 164 Chapter 6. Photo-Emf Sensors and Talbot Effect: Measurement of Displacement and Vibration Osciloscope trace, V 5,5 5,0 UF 4,5 4,0 3,5 ‐1,0 ‐0,5 0,0 0,5 1,0 Time, sec Fig. 6.10. Oscilloscope trace obtained when the test object is vibrating. The loudspeaker is excited at 1.45 V at a frequency of 3 Hz. From this trace the value of the voltage UF is readily obtained. In Fig. 6.11 the dependence of the amplitude of vibration of the test object as a function of the voltage applied to the loudspeaker is plotted. For comparison purposes in the same graph the values of the vibrations amplitudes measured by a Michelson interferometer are plotted. As clearly shown, there is an excellent agreement between the two methods. Am plitude of vibration A,  m 300 200 100 0 0,0 0,5 1,0 1,5 2,0 2,5 Voltage applied to loudspeaker, V Fig. 6.11. Amplitude of vibrations of the test object as a function of the voltage applied to the loudspeaker. Values obtained with (●) the proposed method and with (□) a Michelson interferometer. 6.4. Conclusions The self-imaging phenomenon of a binary grating (or Talbot effect) has been useful for setting up low cost and versatile devices for optical metrology. In some cases, the use of 165 Advances in Optics: Reviews. Book Series, Vol. 3 these devices is limited by the experimental technique employed for determining the visibility of the light diffracted from the grating. As an alternative technique, in this report we have presented the use of photo-emf sensors for measuring the visibility. Sensors based on photo-emf have been mainly used to detect phase-modulated optical signals. They have proved to be cheap, reliable and very robust devices for sensor applications. Here we have used the property that the electrical current induced in photoemf sensor by an oscillating light pattern, in particular by the light diffracted on oscillating diffraction grating, is proportional to the square of the visibility. We have presented two applications using the Talbot effect and the photo-emf sensors, one for measuring the displacements and other for determining the amplitude of vibrations of mirror-like objects using a GaAs: Cr photo-emf sensor. Because of the inherent adaptive properties of the photo-emf sensors, the proposed techniques are very robust to environmental perturbations. In contrast to similar applications, our technique does not require image processing. Because the Talbot distance depends on the period of the grating and the wavelength, the range of measurement can be modified by changing any of these parameters. For measurements of displacements we estimated a resolution better than 10 m in a dynamic range of 1.5 mm, and for the determination of amplitude of vibration, an excellent agreement between our technique and an interferometric one was demonstrated. References [1]. H. F. Talbot Facts related to optical science, Philos. Mag., Vol. 9, Issue 56, 1836, pp. 401-407. [2]. L. Rayleigh, On copying diffraction-gratings, and on some phenomena connected therewith, Philos. Mag., Vol. 11, 1881, pp. 196-205. [3]. K. Patorski, The self-imaging phenomenon and its applications, Prog. Opt., Vol. 27, 1989, pp. 1-108. [4]. S. P. Trivedi, S. Prakash, S. Rana, O. Sasaki, Real-time slope mapping and defect detection in bent plates using Talbot interferometry, Appl. Opt., Vol. 49, Issue 5, 2010, pp. 897-903. [5]. X. Wan, Q. Wang, H. Tao, Nanolithography in the quasi-far field based on the destructive interference effect of surface plasmon polaritons, J. Opt. Soc. Am. A. Opt. Image Sci. Vis., Vol. 27, Issue 5, 2010, pp. 973-976. [6]. A. Isoyan, et al., Talbot lithography: Self-imaging of complex structures, J. Vac. Sci. Technol. B Microelectron., Nanom. Struct., Vol. 27, Issue 6, 2009, pp. 2931-2937. [7]. H. L. Kung, A. Bhatnagar, D. a Miller, Transform spectrometer based on measuring the periodicity of Talbot self-images, Opt. Lett., Vol. 26, Issue 21, 2001, pp. 1645-1647. [8]. E. Ye, A. H. Atabaki, N. Han, R. J. Ram, Miniature, sub-nanometer resolution Talbot spectrometer, Opt. Lett., Vol. 41, Issue 11, 2016, pp. 2434-2437. [9]. J. C. Bhattacharya, Measurement of the refractive index using the Talbot effect and a moire technique, Appl. Opt., Vol. 28, Issue 13, 1989, pp. 2600-2604. [10]. C. Shakher, A. J. P. Daniel, Talbot interferometer with circular gratings for the measurement of temperature in axisymmetrical gaseous flames, Appl. Opt., Vol. 33, Issue 25, 1994, pp. 6068-6072. 166 Chapter 6. Photo-Emf Sensors and Talbot Effect: Measurement of Displacement and Vibration [11]. D. S. Mehta, P. Singh, M. S. Faridi, S. Mirza, C. Shakher, Two-wavelength lateral shearing interferometry, Opt. Eng., Vol. 44, Issue 8, 2005, 085603. [12]. P. Singh, M. S. Faridi, C. Shakher, R. S. Sirohi, Measurement of focal length with phaseshifting Talbot interferometry, Appl. Opt., Vol. 44, Issue 9, 2005, pp. 1572-1576. [13]. J. Dhanotia, S. Prakash, Automated collimation testing by incorporating the Fourier transform method in Talbot interferometry, Appl. Opt., Vol. 50, Issue 10, 2011, pp. 1446-1452. [14]. D. Podanchuk, V. Kurashov, A. Goloborodko, V. Dan’ko, M. Kotov, N. Goloborodko, Wavefront sensor based on the Talbot effect with the precorrected holographic grating, Appl. Opt., Vol. 51, Issue 10, 2012, pp. C125-C132. [15]. D. V Podanchuk, A. A. Goloborodko, M. M. Kotov, A. V Kovalenko, V. N. Kurashov, V. P. Dan’ko, Adaptive wavefront sensor based on the Talbot phenomenon, Appl. Opt., Vol. 55, Issue 12, 2016, pp. B150-B157. [16]. P. Chavel, T. C. Strand, Range measurement using Talbot diffraction imaging of gratings, Appl. Opt., Vol. 23, Issue 6, 1984, pp. 862-871. [17]. G. Schirripa Spagnolo, D. Ambrosini, Talbot effect application: measurement of distance with a Fourier-transform method, Meas. Sci. Technol., Vol. 11, 2000, pp. 77-82. [18]. G. S. Spagnolo, D. Ambrosini, D. Paoletti, Displacement measurement using the Talbot effect with a Ronchi grating, J. Opt. A Pure Appl. Opt., Vol. 4, Issue 6, 2002, p. S376. [19]. D. S. Mehta, S. K. Dubey, C. Shakher, M. Takeda, Two-wavelength Talbot effect and its application for three-dimensional step-height measurement, Appl. Opt., Vol. 45, Issue 29, 2006, pp. 7602-7609. [20]. S. K. Dubey, D. S. Mehta, A. Roy, C. Shakher, Wavelength-scanning Talbot effect and its application for arbitrary three-dimensional step-height measurement, Opt. Commun., Vol. 279, Issue 1, 2007, pp. 13-19. [21]. F. J. Torcal-Milla, L. M. Sanchez-Brea, E. Bernabeu, Talbot effect with rough reflection gratings, Appl. Opt., Vol. 46, Issue 18, 2007, pp. 3668-3673. [22]. L. M. Sanchez-Brea, J. A. Quiroga, A. Garcia-Botella, E. Bernabeu, Histogram-based method for contrast measurement, Appl. Opt., Vol. 39, Issue 23, 2000, pp. 4098-4106. [23]. E. Peli, Contrast in complex images, J. Opt. Soc. Am. A., Vol. 7, Issue 10, 1990, pp. 2032-2040. [24]. G. von B. Songcan Lai, Fringe contrast evaluation by means of histograms, Proceedings of SPIE, Vol. 3573, 1998, pp. 3573 (1-4). [25]. L. M. Sanchez-Brea, E. Bernabeu, On the standard deviation in charge-coupled device cameras: a variogram-based technique for nonuniform images, J. of Electron. Imaging, Vol. 11, Issue 2, 2002, 11. [26]. P. Rodriguez-Montero, C. M. Gomez-Sarabia, J. Ojeda-Castañeda, Adaptive photodetector for assisted Talbot effect, Appl. Opt., Vol. 47, Issue 21, 2008, pp. 3778-3783. [27]. S. Stepanov, Photo-electromotive-force effect in semiconductors, Chapter 6, in Handbook of Advanced Electronic and Photonic Materials and Devices, 2001, pp. 205-272. [28]. N. Korneev, S. Stepanov, Measurement of different components of vibrations in speckle referenceless configuration using adaptive photodetectors, Opt. Commun., Vol. 115, Issue 1-22, 1995, pp. 35-39. [29]. S. Stepanov, P. Rodriguez, S. Mansurova, M. L. Arroyo, S. Trivedi, C. C. Wang, Wavelength dependence of the photo-electromotive-force effect in CdTe: V crystal with bipolar photoconductivity, Opt. Mater., Amst., Vol. 29, Issue 6, 2007, pp. 623-630. [30]. S. Stepanov, P. Rodríguez, S. Trivedi, C. C. Wang, Effective broadband detection of nanometer laser-induced ultrasonic surface displacements by CdTe:V adaptive photoelectromotive force detector, Appl. Phys. Lett., Vol. 84, Issue 3, 2004, pp. 446-448. [31]. F. Davidson, C.-C. Wang, S. Trivedi, Optical phase-lock loops with photoconductive semiconductor phase detectors, Opt. Lett., Vol. 19, Issue 11, 1994, pp. 774-776. 167 Advances in Optics: Reviews. Book Series, Vol. 3 [32]. Y. Ding, I. Lahiri, D. D. Nolte, G. J. Dunning, D. M. Pepper, Electric-field correlation of femtosecond pulses by use of a photoelectromotive-force detector, J. Opt. Soc. Am. B Opt. Phys., Vol. 15, Issue 7, 1998, pp. 2013-2017. [33]. S. Mansurova, P. M. Zarate, P. Rodriguez, S. Stepanov, S. Köber, K. Meerholz, Non-steadystate photoelectromotive force effect under linear and periodical phase modulation: application to detection of Doppler frequency shift, Opt. Lett., Vol. 37, Issue 3, 2012, pp. 383-385. [34]. M. P. Petrov, I. A. Sokolov, S. I. Stepanov, G. S. Trofimov, Non-steady-state photoelectromotive-force induced by dynamic gratings in partially compensated photoconductors, J. Appl. Phys., Vol. 68, Issue 5, 1990, pp. 2216-2225. [35]. P. Moreno Zarate, P. Rodriguez, S. Koeber, K. Meerholz, S. Mansurova, Adaptive detection of Doppler frequency shift using ac non-steady-state photo-EMF effect, in Proceedings of the Quantum Electronics and Laser Science Conference (QELS’08), San Jose, CA, USA, 2008, p. JThA50. [36]. P. Rodriguez-Montero, D. Sánchez-de-la-Llave, S. Mansurova, Electro-optical processor for measuring displacement employing the Talbot and the nonsteady-state photo-electromotive force effects, Opt. Lett., Vol. 39, Issue 1, 2014, pp. 104-107. [37]. J. W. Goodman, Introduction to Fourier optics, Quantum and Semiclassical Opt.: J. of the Eur. Opt. Soc. Part B, Vol. 8, Issue 5, 1996, p. 491. [38]. B. C. Scruby, E. L. Drain, Laser Ultrasonics Techniques and Applications, CRC Press, 1990. [39]. C. Shakher, S. Prakash, Monitoring/measurement of out-of-plane vibrations using shearing interferometry and interferometric grating, Opt. Lasers Eng., Vol. 38, Issue 5, 2002, pp. 269-277. [40]. L. Scalise, Y. Yu, G. Giuliani, G. Plantier, T. Bosch, Self-mixing laser diode velocimetry: Application to vibration and velocity measurement, IEEE Trans. Instrum. Meas., Vol. 53, Issue 1, 2004, pp. 223-232. [41]. S. Prakash, S. Upadhyay, C. Shakher, Real time out-of-plane vibration measurement/monitoring using Talbot interferometry, Opt. Lasers Eng., Vol. 33, Issue 4, 2000, pp. 251-259. 168 Chapter 7. Advances in Label-Free Sensing of Bacteria by Light Diffraction Phenomenon Chapter 7 Advances in Label-Free Sensing of Bacteria by Light Diffraction Phenomenon Igor Buzalewicz, Katarzyna Kowal, Mariusz Linard, Agnieszka Suchwałko and Halina Podbielska1 7.1. Introduction Bacteria are found widely throughout nature and the environment in soils, water or the intestinal tract of animals. An average person carries more than 150 kinds of bacteria which exist inside and outside the body [1]. Although the majority of microorganisms can coexist with humans, plants and animals with beneficial relations, some of them are pathogenic and can be the cause of infectious diseases. Bacteria are omnipresent, however, until the establishment of optical microscopy foundations they could not have been observed visually. The optical microscopy and phase-contrast microscopy have created the perspectives for observation of the individual bacteria cells, what led to the recognition of bacteria. Since then, the use of light and optical techniques play a crucial role in the bacteria detection and identification being the basis of the microbiological diagnosis. Nowadays, the continuous increase of the bacteria resistant to commonly used antibacterial chemicals (antibiotics, sterilisation agents etc.) has been observed. National Institute of Allergy and Infectious Diseases [2] reports that the rapid and sensitive detection and accurate identification of bacteria are crucial to facilitate the narrowspectrum therapies by enabling the use of treatments targeted to a specific pathogen. Optical biosensors offer the non-invasive and nondestructive detection since they enable analysis of the amplitude and phase of light modulated by pathogens, instead of pathogens themselves. There is a variety of detection methods which base on principles. Optical techniques used in microbiology include infrared and fluorescence spectroscopy, flow cytometry, chromatography, chemiluminescence analysis, surface plasmon resonance (SPR) phenomena [3-7]. Over the past few years, it was demonstrated that the analysis of light diffraction on bacterial colonies could be used for identification of different bacteria Halina Podbielska Bio-Optics Group, Department of Biomedical Engineering, Faculty of Fundamental Problems of Technology, Wroclaw University of Science and Technology, Wroclaw, Poland 169 Advances in Optics: Reviews. Book Series, Vol. 3 species [8-18]. Diffraction signatures of bacterial colonies exhibit species-/strainsdependent features, which are suitable for bacteria differentiation and characterization. In this chapter, several aspects of the use of light diffraction on bacterial colonies in microbiological diagnosis are discussed. In Section 7.2 the biophysical model of the bacterial colony is described. The relevant properties of the bacterial colonies and their mathematical description are outlined. Section 7.3 discusses a proposed optical method for bacteria identification, followed by the detailed description of the detection system based on Fresnel diffraction patterns. In the last part of the chapter, the efficiency of the proposed optical system is presented as well as the potential alterative applications in microbiological diagnosis. 7.2. The Biophysical Model of the Bacterial Colony The transformation of the light on the bacterial colony, as on all physical objects, is influenced by its physical properties. Therefore to understand the specific light diffraction on the object, it is necessary to identify and characterise these relevant features. In case of bacterial colonies being the biological objects, the most important factors affecting its interaction with the optical field are species-/strains- dependent optical and morphological properties. A single bacterial colony is a monoculture of bacterial cells with the same genotype properties. A bacterial colony is formed by a single bacterial cell, which by metabolising the nutrients of the culture medium and by multiplication is creating a macroscopic structure consisting of the same bacterial cells and intercellular material excreted during the growth process. Therefore, the bacterial colony is a macroscopic biological object consisting of millions of individual bacterial cells, interacting with each other. The morphology of individual bacterial cells (see Fig. 7.1) forming the colony is species-dependent, and it includes various cells’ shapes (rod-shaped, spheroid-shaped, spirals-shaped), their spatial arrangement, appendages as flagella. The bacterial cells of different species and strains metabolise different nutrients, affecting the internal structure of the bacterial colony. The extracellular material fills the space between the individual cells. The oldest bacterial cells are located in the centre of the colony around the primary location where the first cell was initially situated. With the growth of the colony, that central region becomes the thickest area of the sample, with the highest concentration of intracellular material. These factors in conjunction with the dynamics of colony development, bacterial cells’ shape, its motility and spatial arrangement are influencing the colony profile and shape. The high concentration of bacterial cells and the extracellular material can be observed at a central region of the lowest intensity of phase-contrast microscopic images of bacterial colonies (see Fig. 7.2). In case of some bacterial cells containing flagella - structures protruding from the bacterial cell wall and are responsible for bacterial motility, the edges of the bacterial colonies are irregular and crown-like, or annular envelope-like structures are observed around the central region of the colony. These structures are characteristic for 170 Chapter 7. Advances in Label-Free Sensing of Bacteria by Light Diffraction Phenomenon the bacteria colonies formed by the cells exhibiting high mobility on the surface of the solid nutrient medium, which moved outside the central region of the colony. An example of such colony can be colonies formed by the Proteus mirabilis bacteria (see Fig. 7.3). Therefore, it can be seen, that colony morphology can be highly influenced by the shape of individual cells, cell wall components, the extracellular components and cellular response to nutrient availability, oxygen and other gases, salt, acidity, alkalinity, temperature [19]. The differences in bacterial cells structure, metabolism and external factors are also affecting the morphology and optical properties of bacterial colonies and can be used as foundations for their differentiation and characterization. Fig. 7.1. The examples of different morphologies of bacterial cells. Fig. 7.2. The exemplary morphology of bacterial colonies. 171 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 7.3. The exemplary experimental images of Proteus mirabilis colony recorded by the: (A) scanning confocal microscopy, and (B) phase-contrast microscopy presenting the characteristic structure of the peripheral region of the colony caused by the high motility of the bacterial cells. Depending on the optical properties of the physical object a light wave may be transmitted absorbed, re-emitted with a different wavelength, deflected, diffracted or scattered, what causes a change in the spatial distribution and the amplitude of the incoming optical wave field. Thus, the modulated light wave carries coded amplitude-phase information about the structure of the object. Therefore, the properties of the optical objects may be expressed by the amplitude transmittance function, taking into account the twodimensional distribution of the light transmission coefficient by the object under investigation as well as the two-dimensional distribution of phase changes in the plane of the examined object. This general consideration is also valid in the case of the bacterial colony. The transformation of a light wave on bacterial colonies depends on the properties of the object, which based on the wave optics [20] can be described by the amplitude transmittance function tB(x1, y1, t) which carries information about amplitude and phase modulating properties of the incoming optical wave field. This function can be expressed in the following manner: , , , , Φ , , , (7.1) where TB(x1, y1, t) represents the two-dimensional (2D) light transmission coefficient of bacteria colony and ΦB(x1, y1, t) the total phase delay of the wave field passing through the bacteria colony. This general expression enables the analysis of the complex objects as a light amplitude and phase spatial modulators, which is the most appropriate way to examine the light modulating properties of the biological object and the bacterial colonies particularly. Moreover, it should be pointed out that the bacteria colony is evolving in time, what leads to the temporal changes in the optical and morphological properties. Based on the above-described bacteria species-/strains-dependent object function it is hard to escape from the obvious conclusion that the light transmission properties of the bacterial colony in combination with their phase properties related with colony refractive index distribution and the spatial geometry of the colony will be mostly responsible for the specific amplitude-phase modulation of the illuminating light beam. The light transmission/absorption of bacterial colonies depends not only on the transmission/absorption properties of individual bacterial cells but also on the spatial variation of their density and of the intercellular material excreted during the growth 172 Chapter 7. Advances in Label-Free Sensing of Bacteria by Light Diffraction Phenomenon process, as well as on the changes in the colony thickness. They can be determined using transmission microscopy [21]. Based on wave optics, the complex amplitude U x , y′ of the optical field in the image plane in the transmission mode can be described by the following expression [20]: , ′ | | , , (7.2) where U , is the rescaled complex amplitude of the optical field in the object plane and M determines the transverse magnification of the microscope optical system, x’ = Mx and y’ = My are relationships connecting the coordinates x, y in the object plane with the coordinates x’, y’ in the image plane. Because only the intensity of the optical field is recorded in the conventional transmission microscope, the information about the phase modulation is lost. The intensity of the optical field in the image plane can be described as follows: | , , | ≅ | , | , (7.3) The microscopic image enables obtaining the rescaled two-dimensional transmission , . Therefore, it is possible to determine the transmission coefficient coefficient T T(x’, y’) of the bacterial colony, according to the following expression: , , ̅ , , (7.4) where T(i, j) is the discrete transmission coefficient for the particular pixel (i, j) of the recorded transmission image of the colony, Iob(i, j) is the discrete intensity value of the colony image pixel, and ̅ is the average discrete intensity of the pixels of the nutrient medium image outside the region occupied by the colony. T(i, j) is the relative transmission coefficient of the bacterial colony, normalized to the transmission properties of the nutrient medium on which the colonies are grown. The T(i, j) ∈ < 0; 1 > values are dimensionless units. The value 0 describes the situation of total light attenuation for an opaque object, and value 1 describes total transmission for a transparent object. The representative 2D transmission coefficient of different bacterial colonies are presented in Fig. 7.4. The reduction of the light transmission in the central region of the bacterial colony is caused by the spatial variation of the thickness of bacterial colony, bacteria cells and intercellular material density. The increase of the thickness of the colony leads to the greater light scattering and absorption. This effect is observed for the colonies of spheroid-shaped bacterial cells as Staphylococcus intermedius. These cells are forming the regular convex-shaped profiles colonies. The central region of the colony, near the nutrient medium surface, has the highest mass density as the oldest cells are located there, and the concentration of the extracellular material is the highest. On the other hand, near the colony edges where the 173 Advances in Optics: Reviews. Book Series, Vol. 3 thickness is the lowest, and as a consequence, the light transmission is higher. These features are common among different bacteria species, but they are more evident in the case of the spheroid-shaped bacteria cells (cocci). The symmetry of bacteria cells contributes to the homogeneous internal and external structure of the colony because in this case, the spatial arrangement of bacteria cells does not matter. On the other hand, the colonies of bacteria cells with different morphology (rods-shaped, spirals-shaped) have more inhomogeneous morphology what can be observed in the 2D transmission coefficients maps. The most inhomogeneous morphology and optical properties are common for bacterial cells having flagella structures responsible for high bacterial cells motility on the surface of the nutrient medium. The variations of the different bacteria cells, colony thickness and intracellular material density, can also be observed by phase contrast microscopy technique as shown in Fig. 7.5. Fig. 7.4. The exemplary experimental 2D transmission coefficients maps of the bacterial colony: (A) Bacillus subtilis; (B) Pseudomonas aeruginosa; (C) Staphylococcus intermedius. Fig. 7.5. The exemplary experimental phase-contrast microscopic (objective: 10×) images of bacterial colonies: (A) Proteus mirabilis; (B) Citrobacteri freundii; (C) Staphylococcus aureus. Phase-contrast microscopy can be used to visualize the effect of the spatial distribution of the bacterial cells and the intracellular material concentration on the spatial distribution of the colony thickness, surface roughness and colony profile but also the refractive index 174 Chapter 7. Advances in Label-Free Sensing of Bacteria by Light Diffraction Phenomenon distribution. An increase of the colony thickness and the mass density are indicated by the decrease in the intensity of light. In case of some bacteria species/strains colonies (as Bacillus subtilis and Pseudomonas aeruginosa), the high spatial variations of the colony thickness and the mass density distribution are influencing the heterogeneous spatial distribution of the 2D transmission coefficient. In this case, the strong light scattering and diffraction effects can be observed. Moreover, in some cases when the high local variation of the colony thickness and mass density impacts the roughness of the colony surface, the additional speckle effect contributes to the final light modulation on the bacterial colony. This species-/strains-dependent differences in the colony 2D transmission coefficients’ distribution will affect the optical wave field transformation on the colony. In another word, the light diffraction takes place at boundaries of each zone with different transmission properties in a similar way as in the case of the knife-edge diffraction. As it was mentioned above, the bacterial colonies can be treated as an amplitude and phase spatial modulators of the illuminating optical field. Therefore, it is necessary to describe the phase properties of bacterial colonies which contribute to the phase delays of the incident optical field. The phase delay of the wave field propagating in the medium is related to the optical path length being the product of the geometric length d of the path light follows through the medium and the index of refraction n of the medium through which it propagates: . (7.5) Therefore, the phase delay Φ (x, y) of the wave field can be expressed in the following manner: Φ , , where k is the wavenumber. , (7.6) Spatial geometry of the colony is the main factor influencing the phase delays of the optical field. In the literature, various approaches to the bacteria colonies’ profile shape are presented: a convex shape [13, 22, 23], a thin film with decreasing tailing edge [24] or a Gaussian profile [25, 26]. At the beginning for the simplicity of the general analysis, let’s consider the case when the bacterial colony has a spheroid shape as in case of the Staphylococcus aureus colonies and further on the other approaches of bacterial colony profile (Gaussian profile, profile with two different radii of curvature) will be described. Let’s consider the case of a convex shape of bacterial colony located at the object plane ∞ and rB are defined by (x1, y1) coordinates, where the radiuses of curvatures describing the colony flat input surface (on a nutrient medium) and spherical output surface respectively (see Fig. 7.6 (A)). The total phase delay of the wave field passing through such bacteria colony may be expressed, as: Φ , ∆ , ∆ , , (7.7) 175 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 7.6. Models of different profiles of bacterial colonies: (A) convex; (B) Gaussian; and (C) convex with two radii of curvature. where nB is the refractive index of the bacteria colony, H0 is the central thickness along the optical axis and ∆ , is the bacteria colony thickness function in the off-axis region. Referring to the assumed geometry of the bacteria colony, the thickness ∆ , can be described using following equation: ∆ , 1 1 . (7.8) If we expand the square root term in power series and simplify it, the thickness function can be described as   x1 , y 1   H 0  x12  y12  1 1   ,   2  r rB  (7.9) and the phase delay as  B  x1 , y1   kn B H 0  k ( n B  1) 1  x12  y12  1    . 2 r r B    (7.10) The convex shape of a bacterial colony and the similarity of the above expression to the phase delay introduced by the conventional optical lens indicate that such colony can exhibit the light focusing properties. Therefore, the total phase delay of the convex-shaped colony can be expressed in the following manner:  B  x1 , y1   kn B H 0  176     kF k x12  y12  kn B H 0  B x12  y12 , 2 fB 2 (7.11) Chapter 7. Advances in Label-Free Sensing of Bacteria by Light Diffraction Phenomenon where 1⁄ 1 and fb is the focal distance of bacterial colony. The convex profile of bacterial colony contributes to the light focusing properties of the colony, as in case of the conventional optical lenses. Finally, in the considered case the amplitude transmittance function of the convex shaped colony can be express as follows: , . , (7.12) Now, let’s consider the case of bacterial colony modeled by Gaussian profile (see Fig. 7.5 (B)) as a bell curve shape with tailing edge [26]. As in the previous case, the central thickness of the colony is expressed by H0 and radius of the colony cross-section is expressed by rB, respectively. The Gaussian function describing the colony profile can be derived from the following equation: , (7.13) . The total phase delay of the wave field passing through such bacteria colony is expressed, as: Φ ∆ , , , , Finally, above equation can be written as: Φ , and the amplitude transmittance function as , , , . , 1 1 (7.14) (7.15) . (7.16) In the case of convex shaped bacterial colony with two different radii of curvature (see Fig. 7.5 (C)) rB1 in the central region of colony (Σ1) and rB2 near the colony edges (Σ2), the total phase delay can be described in a similar way as in case of the convex colony with single radius of curvature, but for each region of colony separately: and Φ , , (7.17) Φ , , (7.18) 1 and 1⁄ 1 and where 1⁄ H1+H2 = H0. Above expression have shown that the convex-shaped bacterial colony 177 Advances in Optics: Reviews. Book Series, Vol. 3 exhibits the light focusing properties, with two different focal lengths in each colony region. The amplitude transmittance function in this case can be expressed as: , , , , , ∈Σ , . (7.19) ∈Σ , The examined the Escherichia coli colonies [22] has shown that analyzed biological object exhibits the light focusing properties similar to the conventional microlenses. It should be pointed out that the conventional optical lens does not change the intensity but induces a nonuniform phase shift of the wave, which transforms into visible intensity variations during the propagation through a suitable distance. However, contrary to the classical lens, the bacterial colony is a semi-transparent object, and the light transmission through the colony will be limited, therefore beside the introduced phase shift, the additional amplitude modulation of incoming wave is observed, and the bacterial colony can be treated as amplitude and phase light modulator. Moreover, it was shown, that the intensity distribution in the focal point is more spread or extended in axial and lateral directions than in case of conventional optical lenses. The bacterial colony can also be considered as an optical diffuser, which is affecting the spread of focal point and spatial light intensity variations in the focal plane. Therefore, for the validation of the most general model of the bacterial colonies the phase delay caused by the colony surface roughness: Φ , , where , , (7.20) should be added to the previously described total phase delays expressed by Eq. (7.11), Eq. (7.15), Eq. (7.17) and Eq. (7.18), or be introduced to the amplitude transmittance function of bacteria colony: , , Φ , Φ , . (7.21) This function describes the amplitude and phase properties of a bacterial colony, which are responsible for specific spatial modulation of the illuminating optical wave field. Moreover, the 2D transmission coefficient and the phase delay associated with colony geometry are describing the complex illuminating optical wave field amplitude and phase modulation. 7.3. The Optical System for Analysis of Light Diffraction on Bacterial Colonies The differences in morphology and optical properties of the bacterial colony can be used as foundations for their differentiation and characterization. Moreover, they are also influencing the illuminating optical wave field, what means that the presence of the species-/strains-dependent light transformation on bacterial colony can provide optical signatures, which can be used for bacteria differentiation and characterization. Nowadays, 178 Chapter 7. Advances in Label-Free Sensing of Bacteria by Light Diffraction Phenomenon two bacteria identification systems based on forward-light scattering or diffraction approaches are developed: BARDOT (Bacterial Rapid Detection using Optical scattering Technology) at Purdue University [8-10] and BISLID (Bacteria Identification System by Light Diffraction) at Wroclaw University of Science and Technology [13-18]. There are also some intermediate approaches based on the both BARDOT and BISLID systems [11, 12]. In BARDOT system, no assumption about the type of the illuminating beam is made. However, in each of the colony illumination using either plane wave (collimated beam), spherical divergent or convergent wave illumination, the angular divergence of diffracted optical wave field is different and in different locations from the colony various diffraction patterns are registered. Based on the configuration of the optical correlators carrying out the optical Fourier transform, the type of the illuminating beam can offer significant advantages, which can improve the conditions of the diffraction patterns registration [27]. The BISLID system is based on the optical system with converging spherical wave illumination, which is generated by the transforming lens located before the object plane, where the bacterial colonies are placed. In the proposed system, it is possible to record both the Fresnel and Fraunhofer diffraction patterns, because the converging spherical wave illumination eliminates the need of large observation distances for recording the Fraunhofer pattern. The extended analysis of the converging spherical wave illumination system properties [28] showed that this system allows to compress and distort the observation space along an optical axis into the finite region of the space between the diffracting object and the Fourier transform plane. Moreover, in this optical system, it is possible to control scaling of the registered diffraction patterns, what enables the fitting the lateral size of the diffraction patterns to the size of the matrix of the camera. It should be pointed out, that the optical system with converging spherical wave illumination possesses more advantages comparing to the other Fourier transform systems. The transforming lens, which is placed in front of the object, must be corrected only for a pair of on-axis points to produce the spherical wave and not for all aberrations in the infinity – focal plane points pairs as in configuration with the plane wave illumination. Therefore, the setup with converging spherical wave illumination is simpler, so the level of coherence noises on optical elements is reduced. Moreover, choosing the same size of the lens as the size of the object allows avoiding the bandwidth limitation of the lens. These properties additionally lower the costs of the system construction. 7.3.1. The Optical Wave Field Transformation in Proposed Optical System To understand the main features of the proposed optical system with the converging spherical wave illumination, it is necessary to present the theoretical model of light transformation [13]. For simplicity, let assume that a coherent plane wave Uin.(x0, y0) = A with the amplitude A, propagating along optical axis z perpendicularly to the (x0, y0) plane, falls on the transforming lens L0. It means that the point light source is located at an infinite distance from the transforming lens (see Fig. 7.7). The lens L0 with the focal distance f is transforming the incident plane wave into the converging spherical wave, which can be described as: 179 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 7.7. Proposed optical system configuration for characterization of bacteria colonies diffraction patterns: L0 transforming lens in (x0, y0) plane, bacteria colonies on Petri dish in (x1, y1) plane, observation plane (x2, y2). , , , Ψ , , ; (7.22) where λ is the wavelength of the incident wave, F = 1/f, P(x0, y0) is a pupil function of the transforming lens, and the function  x , y , p  represents a Gaussian function Ψ , ; (7.23) . Between the lens L0 and the object plane (x1, y1) located in the distance z1 from the lens, the free propagation takes place. Therefore the optical field can be described using the Fresnel diffraction approximation, as follows: U ( x m 1 , y m 1 )  exp ikz m 1   i z m 1     U x m      i x m 1  x m 2   y m 1  y m 2 , y m exp  z   m 1    C  , Z m 1   U ( x m dx  m dy m , , y m ) x m 1  x m , y m 1  y m , Z m 1 dx m dy m   (7.24) where Z m1  1 zm1 . Finally, the wave field illuminating the object plane can be expressed as:   U in ( x1 , y1 )  C  , Z 1   U out . x 0 , y 0  x1  x 0 , y1  y 0 , Z 1 dx 0 dy 0 ,    180      (7.25) Chapter 7. Advances in Label-Free Sensing of Bacteria by Light Diffraction Phenomenon where C(λ, Z1) is a constant characteristic for Fresnel approximation depending on λ and Z1. Additionally, it was assumed that the colony is fully illuminated by the converging spherical wave, therefore its pupil function can be ignored. After simple transformations and substitutions, the optical wave field Uin.(x1, y1) illuminating the bacterial colony can be expressed, as follows: U in ( x1, y1)  i A Z1  F      C , Z1  x1, y1, Z1F  , Z1  F  (7.26) or after simple transformation, as:  f   Z F  A  exp ikz1  x1, y1, 1 ,  f z   Z1  F    1    Uin ( x1, y1)   (7.27) The above expression represents a spherical wave converging towards the plane z = f and the amplitude changes proportionally to the ratio f/(f-z1), what is in agreement with the geometrical optics predictions. When this wave illuminates the single bacteria colony on Petri dish placed in the object plane (x1, y1), its amplitude and phase are modulated by the analyzed object as it was discussed in the previous sections. The contribution of the nutrient medium and Petri dish is limited to the presence of exponential phase shift along the optical axis, as well as to the attenuation of a primary amplitude of the incident wave. The amplitude and phase transformations on bacteria colony of the wave field Uin.(x0, y0), can be simply presented by , , , . (7.28) Similarly, as in the case of the free propagation of the optical field from the lens L0 to the object plane, we are using the Fresnel approximation to obtain scattered wave field in the observation plane:   U in ( x 2 , y 2 )  C 2  , Z 2   U x1 , y1  x 2  x1 , y 2  y1 , Z 2 dx 1 dy 2 .         (7.29) After rearrangement of Eq. (7.29) and appropriate substitutions, the optical wave field in the observation plane (x2, y2) takes a form:      x , y , Z 2 2 2  1  f fAz U in ( x 2 , y 2 )  C  , Z1 , Z 2       ~ x Z y Z   t b ( x1 , y1 )  x1 , y1 , Z f x 2 2; f y  2 2   (7.30) , where ZF ~ Z  Z2  1 . Z1  F (7.31) 181 Advances in Optics: Reviews. Book Series, Vol. 3 The operator … represents the two-dimensional Fourier transform. It can be seen that 0 Eq. (7.30) is describing the Fresnel transform of the bacteria colony amplitude for transmittance. However, there are some important differences between this expression and the conventional Fresnel diffraction formula known from scalar theory of diffraction. Presented above expression should be considered as a Fresnel diffraction formula for tB(x1, y1) alone, and not for entire scattered wave field U.(x1, y1), as it is commonly ~ regarded. Moreover, the parameter Z is not describing the distance to the observation plane, but rather the nature of diffraction pattern (Fresnel or Fraunhofer), which is observed. If the observation plane is shifted to the back focal plane of the transforming lens, then z1 + z2 = f and the parameter Z~  0 . Moreover, the exponential quadratic phase term inside the two-dimensional Fourier transform is eliminated and Eq. (7.30) takes a form of the Fourier transform of the bacteria colony amplitude transmittance alone representing the Fraunhofer diffraction formula:  fA   x , y , Z  t ( x , y ) x Zˆ y Zˆ U ( x f , y f )  U ( x2 , y2 )  C  , Z1, Z 2   f  z  2 2 2 B 1 1 f x  2 ; f y  2 , 1       (7.32) where Z1F 1 1 Zˆ    . zˆ f  z1 Z 1  F (7.33) If the location of the observation plane ranges from the object plane z = z1 to the Fourier transform plane z = f, then, it is possible to observe the Fresnel diffraction pattern of the ~ object. The scale of the patterns depends directly on the value of the parameter Z and indirectly on the relation between the distance z1 and f (see Fig. 7.8). If the observation plane is near the Fourier transform plane, then Z~  0 and the Fraunhofer diffraction pattern of the object can be observed with the scaling factor of Zˆ depending on the distance zˆ  f  z1 . It means that by increasing the distance ẑ , the size of the diffraction pattern is greater until the object is directly behind the lens. If the distance ẑ decreases, the size of the pattern is reducing. Moreover, when the point light source is moved closer to the front focal plane of the transforming lens, then the illuminating beam converges less rapidly, and the Fourier transform plane moves away from the lens. Therefore, the scale changes of the observed diffraction patterns will be larger. In such optical system, the matrix size of the detectors may be smaller due to the possibility of adjusting an appropriate scale of the observed diffraction pattern. However, it should be pointed out, that results reported in [22] have shown that the bacterial colony is evolving over the time, and it can be considered as an optical element with the adaptive light focusing properties as the focusing properties are changing in time. This feature affects the use of Fraunhofer diffraction patterns for bacteria species identification, because the location of the Fourier plane due to the changing of the focal length of the colony, is continuously shifting along the optical axis over the time. Therefore, in the proposed optical method for bacteria species identification based on 182 Chapter 7. Advances in Label-Free Sensing of Bacteria by Light Diffraction Phenomenon diffraction patterns of bacterial colonies, the additional system of bacterial profile determination should be included. Moreover, also the use of Fresnel diffraction patterns of bacterial colonies for their species identification recorded in fixed collection distance is affected by the light focusing properties of bacterial colonies as well because the Fresnel patterns will be shifted along the optical axis for bacterial colonies incubated in different times. To omit this problem, it is necessary to additionally characterize the locations of Fresnel diffraction patterns observation plane for different times of bacterial colonies incubation or the time of incubation should be fixed [18]. This factor should be considered as standardized parameter limiting the efficiency of bacteria species identification by the methods based on analysis of bacterial colony diffraction patterns. Fig. 7.8. The experimental results of the change of the lateral size of Fresnel diffraction patterns in the case of Salmonella Enteritidis colony with decreasing the distance z1:(a) 28 cm; (b) 26.5 cm; (c) 25.3 cm, and (d) 24.5 cm (bacteria colony diameter: approx. 0.8 mm, beam diameter: approx. 1 mm) [13]. 7.3.2. The Configuration of the Experimental Optical System for Bacteria Identification The optical system for recording the Fresnel diffraction patterns of bacterial colonies presented in Fig. 7.9 includes: (1) the laser diode module (635 nm, 1 mW, collimated Thorlabs), (2) beam expander BE (1.5X, Thorlabs), (3) amplitude filters (OD: 0-4.0, Thorlabs) (4) iris diaphragm (diameter: 0-2, 5 cm, Thorlabs) (5) transforming lens L0 (f = 45 cm, Edmund Optics), (6) beam splitter (T:R = 50:50, Thorlabs), (7) holder with the sample of bacterial colonies in Petri Dish integrated with an automatic X-Y-Z translation stage, (8) beamsplitter (T:R = 50:50, Thorlabs), (9) CCD camera (EO-1312, Edmund Optics) with imaging objective (f = 3.5 cm, Edmund Optics) for diffraction patterns recording and (12) computer unit. 183 Advances in Optics: Reviews. Book Series, Vol. 3 The system also contains an additional channel for registration of image of all bacterial colonies on Petri dish and the automatic localization of bacterial colonies grown on the medium. It combines of (6) beam splitter, which enables recording of bacterial colonies on Petri dish in reflection mode, (10) the camera (CMOS, Basler ace) with imaging objective (f = 12 mm, Edmund Optics) and (11) the ring illuminator for uniform illumination of the Petri dish. Fig. 7.9. The experimental BISLD system configuration (description in text). 7.4. Bacteria Identification Based on Fresnel Diffraction Patterns of Bacterial Colonies 7.4.1. The Bacteria Sample Preparation The bacterial samples were prepared in the laboratory of the Department of Epizootiology and Veterinary Administration with Clinic of Infectious Diseases of the Wroclaw University of Environmental and Life Science. Bacteria suspensions of different bacteria species/ strains were first incubated for 18 hours at the temperature of 37 °C. The suspensions in respectively 10-5 and 10-6 dilutions were seeded on the surface of the solid nutrient medium in Petri dishes with Columbia agar (Oxoid), so as to obtain 12-20 colonies per plate, and were again incubated at 26 °C for the next 18 hours. The same solid nutrient medium was used for all bacterial species/strains to omit the initial differentiation and a priori information about their species/strains introduced by their metabolic properties. Fixing the parameters of the incubation process (nutrient medium, temperature and time of incubation etc.) and defined diameter of the colony, guarantees the proper comparison of the bacteria colonies diffraction patterns. Details of the sample preparation and incubation standardization can be found in [13, 17, 18]. 184 Chapter 7. Advances in Label-Free Sensing of Bacteria by Light Diffraction Phenomenon 7.4.2. The Experimental Fresnel Diffraction Patterns of Bacterial Colonies As it was mentioned above, the bacterial colonies exhibit the variety of the morphologies and optical properties among different bacteria species, but also among different strains of the same species. These factors are also influencing the optical wave field transformation on bacterial colonies, manifested by the spatial intensity distribution of the registered Fresnel diffraction patterns of these colonies (see Fig. 7.10). Fig. 7.10. The exemplary experimental Fresnel diffraction patterns of bacterial colonies. Under the visual inspection, it is possible to distinguish the unique features of the diffraction patterns of bacterial colonies among different species and strains, as a different number of the round-shaped intensity maxima or radial-spokes intensity maxima, as well as the presence of the round spot or disc and more complex spatial intensity modulations. The variety of colonies of different bacteria species/ strains exhibit unique properties and diffraction signature. Therefore it is hard to describe the light diffraction on bacterial colonies in details in each case. Therefore, this consideration will be limited to the case of the Staphylococcus aureus (ATCC43300) and Escherichia coli (ATCC35401) colonies (see Fig. 7.11). In case of the S.aureus colonies, the 2D transmission coefficient shows that the light transmission is limited in all regions of the colony with a slightly higher transmission near the colony edges, suggesting that it can be treated as the nearly opaque disc. 185 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 7.11. The comparison of (A) 2D transmission coefficients; (B) phase-contrast images, and (C) Fresnel diffraction of (I) Staphylococcus aureus and (II) Escherichia coli colonies. Moreover, the phase contrast image demonstrates regular and continuous changes of the intensity contrast, what suggests that it has a convex-shaped profile. In the region near the colony edges, the thickness of the colony is minimal, what is responsible for higher transmission coefficient in this region. In consequence, the diffraction pattern of this colony contains the set of the round-shaped intensity maxima with two strong maxima in the central region of the patterns and weaker maxima in the peripheral regions. This Fresnel pattern is similar to the diffraction patterns characteristic for opaque discs. However, the additional light focusing properties of the convex-shaped colony are affecting the distance between these intensity maxima. On the other hand, in case of the Escherichia coli colony, the two zones with different light transmission can be distinguished in the central and in the peripheral region. However, the central zone has irregular boundaries and contains the concentric spokes-like features, which are observed pointing outwards from the centre of the colony. Moreover, the phase-contrast image of the colony shows that around the central region of the colony with the lowest contrast the crescent-like structures are observed, what can be associated with the process of the colony forming and local inhomogeneity of the internal structure of the colony. The Fresnel pattern of this colony contains two round intensity maxima, but their shape is not as regular as in case of the S.aureus diffraction pattern. It can be associated with the irregular boundaries of the central zone with lower light transmission and irregular shape of the bacterial colony edges. Moreover, the Fresnel patterns include the radial spokeslike features, which presence can also be caused by shape irregularities of the central colony zone with the lowest light transmission or the crescent-like features observed on the phase-contrast image of the colony. It should be pointed out that the bacterial colonies are evolving in time, changing the morphology and the optical properties, and in consequence is also affecting the spatial distribution of the Fresnel diffraction patterns intensity (see Fig. 7.12). 186 Chapter 7. Advances in Label-Free Sensing of Bacteria by Light Diffraction Phenomenon Fig. 7.12. The representation of time-dependent changes in the Fresnel diffraction patterns of the Escherichia coli colony. For longer (36 and 40 hours) incubation times of the Escherichia coli colony, the second circular maximum occurs inside the region of the colony shadow and the diameter of the diffraction rings decreases. This effect can be caused by the change of the size of the bacterial colony, as well as the size of the zones with different 2D transmission coefficients inside the colony. According to the Eq. (7.30), the bacterial colony diffraction pattern can be treated as a Fourier transform of the colony amplitude transmittance function. According to the similarity theorem of the Fourier transform, when the diameter of bacterial colony and diameter of internal zones of the colony with different transmission properties increases with the incubation time, the diameter of the diffraction ring of Fresnel patterns decreases. However, it should be mentioned that some significant influence of the central thickness of bacterial colony on maximal diffraction angle and number of diffraction rings, is observed. Also, the different chemical compositions of the nutrient medium and temperature of incubation cause the changes of bacterial colony morphology, size and its transmission properties. However, after standardization of the bacterial colony incubation conditions: type of nutrient medium, the temperature of incubation and the time of incubation, it is possible to obtain high repeatability of the registered Fresnel diffraction patterns for different colonies of the same bacteria species/strains (see Fig. 7.13). Therefore, preservation of the above-indicated incubation conditions enables the use these optical signatures for bacteria species/strains differentiation. Fig. 7.13. The repeatability of the experimental Fresnel diffraction of Citrobacter freundii colonies. 187 Advances in Optics: Reviews. Book Series, Vol. 3 7.4.3. The Analysis of the Diffraction Patterns It was shown that the visual inspection of the Fresnel diffraction patterns of bacterial colonies enables the differentiation of different bacteria species or strains. However, from the clinical point of view in microbiological diagnosis, it is necessary to obtain quantitative indicators characterizing the accuracy of the bacteria species/strains identification based on the registered diffraction patterns. Evaluation of the system can be carried out using sectional analysis of the diffraction patterns was proposed. In this approach, the Fresnel diffraction patterns of bacterial colonies are partitioning by annularshaped limitation zones (partitioning ring) from which the quantitative and statistical features are extracted. The extraction of the new and interpretable features from the diffraction patterns was preceded using a dedicated macro written in the ImageJ free software (http://rsb.info.nih.gov/ij/) with human interaction for distinguishing the center and edges of the diffraction patterns [29]. Marking of the edges and the center was followed by partitioning each of Fresnel patterns into 10 disjoint rings of the equal thickness. Partitioning into 10 rings was shown to be the best of fixed splits [17, 20]. Then, the normalization process of the Fresnel patterns was performed with use of the standard algorithm for histogram stretching. For each partitioning ring of each pattern the values of the following features were calculated: mean and standard deviation denoting brightness and roughness of the regions of interest, respectively; skewness and kurtosis, relative smoothness, uniformity and entropy [17]. Additionally, the Fresnel diffraction pattern radius was calculated, as it is an important predictor. The selection of features, which are the best for building the classification models, was performed by the use of ANOVA (analysis of variance) [31] and Fisher divergence measure [32] to estimate the features separation possibilities measure. Fisher divergence is also called signal to noise ratio and thus further will be called SNR. The proposed measure denotes possibilities of the given feature to separate various classes (bacteria strains). In the study, several classification models were investigated: LDA (Linear Discriminant Analysis), QDA (Quadratic Discriminant Analysis) and SVM (Support Vector Machine). However, the best results of differentiating the bacteria on the strain level were obtained for QDA and SVM. The cross-validation (CV) was chosen as a classifier performance assessment method. The task of the performance assessment method is to estimate the unknown classification error that will occur after using the classification model on other, independent data sets. CV accomplishes the task by splitting the data set into two disjoint subsets (learning and test sets). The model is built with the application of given feature set (predictors) on the learning set, while its performance is tested on the test set. The procedure is repeated given a number of times, and upon the results, the classification error is estimated. The schema of the proposed Fresnel patterns of bacterial colonies was shown below (Fig. 7.14). The exemplary results of the classification assessment were performed on bacterial colonies of Escherichia coli (PCM O119), Staphylococcus aureus (PCM 2267), Proteus mirabilis (PCM 547), Salmonella Enteritidis (ATCC 13076) and Salmonella typhimurium (ATCC 14028) bacteria. The 250 Fresnel patterns of bacterial colonies were recorded at 188 Chapter 7. Advances in Label-Free Sensing of Bacteria by Light Diffraction Phenomenon the fixed distance from the colony for which the highest accuracy of identification on the bacteria strain level was achieved. The performed analysis (see Table 7.1) gives better results for SVM classification than for QDA, while there is the statistically non-significant difference between SNR and ANOVA feature selection methods on the same data. Fig. 7.14. The schematic of the proposed method of the bacteria identification based on Fresnel diffraction patterns. Table 7.1. Identification results of the optimized analysis. Two feature selection methods (ANOVA and SNR) were used for the features predictive properties ranking. The predictors used in the model building, identification error, multi-class sensitivity and specificity are depicted in the table. Most significant results are marked in bold. Number of features Error [%] Sensitivity Specificity SNR 18 3.7 0.9708 0.9626 QDA ANOVA 18 3.92 0.9791 0.9576 SNR 27 1.34 0.9759 0.9903 SVM ANOVA 27 1.38 0.9751 0.9900 The proposed method enables identification of the bacteria with an error as small as 1.34 %, sensitivity equal 0.9759 and specificity 0.9903. Obtained results indicate that the bacteria identification based on the Fresnel diffraction patterns of bacterial colonies can 189 Advances in Optics: Reviews. Book Series, Vol. 3 provide significant advances in the microbiological diagnosis, because its accuracy is comparable with conventional used biomolecular techniques, but can offer the costs limitation and facilitation of the measurements procedure. 7.5. The Use of the Fresnel Diffraction Patterns of Bacterial Colonies for Evaluation of the Efficiency of Antibacterial Factors The problem of bacteria identification and limitation of the bacteria contamination risk are the most important issues of the contemporary science in the context of increasing bacterial drug resistance. Therefore, there is a need to develop new, faster and more common methods of bacteria detection and identification, antibacterial agents and factors, as well as techniques to characterize their effectiveness against bacteria. Performed experiments have shown that it is possible to use also the Fresnel diffraction patterns of bacterial colonies for examination of the selected physicochemical factors with antimicrobial properties. The investigation was performed on the samples of Yersinia entercolitica (ATCC 23715) colonies, prepared as reported in the previous sections. Three physicochemical factors were used in the study: low temperature (4° C), UV radiation (340-390 nm) and chemical bactericide Skinsept® Pur (Ecolab). Skinsept® Pur in 100 g of active ingredient contains: 46.0 g ethanol (96 % denatured), 27.0 g of isopropyl alcohol; 1.0 g benzyl alcohol and excipients: hydrogen peroxide, purified water. Alcohols cause denaturation of proteins and dissolution of lipids of bacterial membrane, whose continuity can be interrupted, leading to cytoplasm leakage into the external environment and cell death. Skinsept® Pur containing a mixture of alcohols was used to investigate the effect of this agent. Approximately 0.09 g of the product was applied to the colonies. At the low temperature, there is a disturbance in the synthesis of essential compounds needed to carry out the life processes of the bacterial cell, which results in a slower metabolism. For this purpose, colonies of bacteria were placed at 4 ° C ± 0.5 °C for 55 minutes and then further incubated until the measurement. The ultraviolet radiation directly affects the structure of the deoxyribonucleic acid of the bacterial cell. It should be noted, that if UV irradiation does not destroy the bacteria in a way that will prevent them from continuing development, then the genes responsible for the reconstruction of bacterial cells will be activated and further microbial development will be possible. In the experiment, a UVA lamp emitting radiation at the wavelength of 340-390 nm was used. In the sample plane, the radiated power was 923.6 ± 1.99 μW, and the irradiance was 1175.96 ± 2.53 μW/cm2. Colonies of bacteria were illuminated for 40 minutes and then further incubated until the measurement was taken. Each of the applied antibacterial factors affects the morphology and size of the bacterial colonies, however with different efficiency (see Fig. 7.15). Compared to the colonies of the control sample, bacterial colonies treated with selected physicochemical agents have considerably smaller spatial sizes, indicating a slowdown or inhibition of colony development. To quantify the change in the diameter of the colonies treated with these factors to the control sample, the size of these colonies was determined using microscopic images. The obtained results show that, among all studied factors, the greatest reduction in colony diameter occurred with UV radiation. Colonies of bacteria 190 Chapter 7. Advances in Label-Free Sensing of Bacteria by Light Diffraction Phenomenon treated with UV radiation were 64 % smaller than the control colonies. For Skinsept® Pur and low temperature, the colony diameter decreased by 29 % and 9 %, respectively. The greatest inhibition of bacterial colonies has been obtained in samples irradiated with UV radiation. Slowing down the development of bacterial colonies leads not only to limiting their spatial sizes but also to diversify their internal structure. Changing the size of the colonies and their structure affects the spatial distribution of the Fresnel diffraction patterns generated by them, which are closely correlated with the morphological characteristics of the colonies. Based on the scalar diffraction theory, the Fresnel diffraction patterns can be expressed as a Fourier transform function of the amplitude transmittance function of the colony and an additional quadratic phase component strictly dependent on the distance of the observation plane from the objective plane. Therefore, given the properties of the Fourier transform, it is expected that the spatial dimensions of the diffraction spectra of bacterial colonies subjected to the physicochemical properties tested will also change as they change the size of the colonies. According to Fourier's similarity or scaling of the Fourier transform, the change in transverse dimensions of the object will lead to the inverse proportional change of Fourier spectra, of which the special case is the Fresnel spectra. In the case of bacterial colonies, this will be affected by changes in the diffraction patterns diameter, the distance between the outer and inner round-shaped intensity maxima and the width of these maxima. Fig. 7.15. The exemplary of microscopic images (objective: 4×) of Yersinia entercolitica bacterial colonies (incubation time 23h): (A) control sample and samples treated with: (B) low temperature - 4 °C, (C) Skinsept® Pur, (D) UV radiation. In case of the Fresnel pattern of the control colony shown in Fig. 7.16, two main diffraction rings are visible in the center and at the periphery of the Fresnel patterns. The interior of the diffraction patterns is characterized by a specific "reticular" modulation of intensity associated with the internal structure of the colony. On the other hand, in case of colony treated with physicochemical factors, the increasing effectiveness of their action 191 Advances in Optics: Reviews. Book Series, Vol. 3 leads to the inhibition of colonization and formation of colonies with significantly smaller diameter. In consequence, in their Fresnel patterns, the distance between the two major diffraction rings increases. It can be noted that the diameter of the central diffraction ring increases with the increase of the diameter of the colony, what means that its size depends on the stage of development of the colony. Comparing the Fresnel patterns of treated colonies with the pattern of the control colony, there is a gradual decline in the characteristic “reticular” modulation of the intensity of the diffraction patterns. It is worth emphasizing that this effect strengthens with increasing inhibitory effect of the factor under consideration. Fig. 7.16. The exemplary Fresnel diffraction patterns of Yersinia entercolitica bacterial colonies (incubation time 23 h): (A) control sample and samples treated: (B) at low temperature 4 °C, (C) Skinsept® Pur, (D) UV radiation (red dotted lines marked major intensity maxima). Obtained experimental results have shown that the Fresnel diffraction patterns, which spatial distributions are related to the morphology and optical properties of bacterial colonies, are highly sensitive to the induced changes in the colony development. The quantitative analysis of the mean and standard deviation of the pixels intensity in each diffraction pattern partitioning zone extracted in a similar way as in the case of features extraction for bacteria identification have shown that they can be used to characterize differences between diffraction patterns of colonies treated by different antimicrobial factors. These create perspectives on the future use the analysis of the Fresnel diffraction patterns of bacterial colonies for characterization of the minimal inhibitory concentration (MIC), which is the lowest concentration of a drug which prevents visible growth and development of the bacterial colonies, at which this drug has bacteriostatic activity. This examination is commonly used in microbiology for examination of the antimicrobial properties of different drugs or chemical agents and the light diffraction on the bacterial colonies can offer significant improvements and advantages in future microbiological diagnosis. 192 Chapter 7. Advances in Label-Free Sensing of Bacteria by Light Diffraction Phenomenon 7.6. The Perspectives for Exploiting of Light Diffraction on Bacterial Colonies Using Digital Holography Previously described results of the research have shown that diffraction patterns analysis enables the bacteria identification, however, are based on Fresnel patterns recorded in the specific observation plane. In analyzed case, the Fresnel diffraction takes place, the spatial distribution of diffraction patterns is significantly affected by the distance between the sample and detector. Described above technique utilizes light scattering/diffraction in one selected direction and at a fixed distance from the colony, although the spatial distribution of scattering/diffraction patterns is significantly affected by the observation distance. This dependence is characteristic for the Fresnel diffraction patterns, and it is caused by the presence of the quadratic phase factor in the Eq. (7.29). During the single measurement, only one diffraction/scattering pattern can be recorded. However, the diffraction patterns recorded at different distances from the sample can exhibit different unique features depending on this distance. Therefore, it is necessary to perform series of measurements of diffraction patterns in different distances from the colony, what is a time-consuming process. Moreover, the bacterial colony profile can be asymmetric and tilted respectively to the optical axis, and in consequence, the light diffracted on colony will not propagate in-line along the optical axis. Some of these disadvantages can be eliminated using digital holography, which enables reconstruction of the amplitude and phase properties of examined objects, as well as the amplitude and phase patterns of the optical field scattered/diffracted by the object in a chosen observation plane behind the object from one single digital hologram. The preliminary performed experiments by the use of point-source digital in-line holography have shown that digital holography can be used for the characterization of bacterial colonies and can find the potential use for microbiological investigation [21]. To the best of our knowledge, it was the first attempt at characterizing the species-dependent properties of bacterial colonies and their diffraction patterns by digital in-line holography. The performed analysis of the reconstructed amplitude and phase patterns of examined bacterial colonies (Escherichia coli and Staphylococcus intermedius) revealed unique species-dependent optical properties. Moreover, the single measurement digital hologram recording and its numerical reconstruction enabled obtaining a reference database of additional bacterial diffraction signatures from all observation space. In consequence, it allows for the extraction of additional differentiating features, in contrast to the already proposed methods based on single scattering/diffraction patterns recorded in a fixed observation plane. This method can provide new optical discriminators for bacterial species, which can extend the classification vector and improve the bacterial identification accuracy. The comparison of the reconstructed Fresnel patterns obtained with pointsource digital in-line holography microscopy with previously recorded Fresnel diffraction patterns of the same colonies demonstrated that both patterns are highly correlated. The potential of bacteria species differentiation by digital holographic signatures was also demonstrated by means of the Principle Components Analysis of examined optical signatures of bacterial colonies and it was demonstrated that these optical signatures obtained by digital holography exhibit unique species-dependent features. Future research 193 Advances in Optics: Reviews. Book Series, Vol. 3 will be focused on more extended investigation of potential application of digital holographic sensors for different bacteria species/strains characterization. 7.7. Conclusions The light diffraction on bacterial colonies being the macroscopic objects enables noncontact and nondestructive examination and offer significant facilities as no need for advanced and time-consuming sample preparation or the use of additional chemical reagents /fluorescence/immunological markers. Moreover, the bacteria samples can be used for further verification or investigation, what distinguishes this method from others conventionally used in microbiological diagnosis: biochemical or biomolecular. The phenomenon of light diffraction on bacterial colonies grew on solid nutrient media can be used for accurate bacteria identification using the examination of Fresnel diffraction patterns supported by statistical analysis as well as for characterization of the antimicrobial agents. The digital holography technique based on light diffraction phenomenon and numerical reconstruction algorithms also offers significant advantages which can enable the more complex and extended analysis of the optical signatures of bacterial colonies. Acknowledgements This work was partially supported by statutory funds of Wroclaw University of Science and Technology. I.B. is funded by a scholarship (No. 489/2014, Contract No. 0159/E366/STYP/9/2014) of the Polish Ministry of Science and Higher Education. References [1]. M. T. Madigan, J. M. Martinko, J. Parker, Biology of Microorganisms, Prentice Hall, 1997. [2]. National Instititute of Allergy and Infectious Diseases Report, https://www.niaid.nih.gov/sites/default/files/arstrategicplan2014.pdf [3]. S. J. Mechery, X. J. Zhao, L. Wang, L. R. Hilliard, A. Munteanu, W. Tan, Using bioconjugated nanoparticles to monitor E. coli in flow channel, Chemistry - An Asian Journal, Vol. 1, 2006, pp. 384-390. [4]. P. Leonard, S. Hearty, J. Quinn, R. O'Kennedy, A generic approach for the detection of whole Listeria monocytogenes cells in contaminated samples using surface plasmon resonance, Biosensors and Bioelectronics, Vol. 19, Issue 10, 2004, pp. 1331-1335. [5]. D. Ivnitski, I. Abdel-Hamid, P. Atanasov, E. Wilkins, Biosensors for detection of pathogenic bacteria, Biosensors & Bioelectronics, Vol. 14, 1999, pp. 599-624. [6]. A. C. Samuels, A. P. Snyder, D. K. Emge, D. Amant, J. Minter, M. Campbell, A. Tripathi, Classification of select category A and B bacteria by Fourier transform infrared spectroscopy, Applied Spectroscopy, Vol. 63, Issue 1, 2009, pp. 14-24. [7]. A. Maninen, M. Putkiranta, A. Rostedt, J. Saarela, T. Laurila, M. Marjamӓki, J. Keskinen, R. Hernberg, Instrumentation for measuring fluorescence cross-sections from airborne microsized particle, Applied Optics, Vol. 47, Issue 7, 2008, pp. 110-115. 194 Chapter 7. Advances in Label-Free Sensing of Bacteria by Light Diffraction Phenomenon [8]. P. P. Banada, S. Guo, B. Bayraktar, E. Bae, B. Rajwa, J. P. Robinson, E. D. Hirleman, A. K. Bhunia, Optical forward-scattering for detection of Listeria monocytogenes and other Listeria species, Biosensors and Bioelectronics, Vol. 22, 2007, pp. 1664-1671. [9]. P. P. Banada, K. Huff, A. Aroonnual, E. Bae, B. Rajwa, B. Bayraktar, A. Adil, J. P. Robinson, E. D. Hirleman, A. K. Bhunia, Label-free detection of multiple bacterial pathogens using light scattering sensor, Biosensors and Bioelectronics, Vol. 24, 2009, pp. 1685-1692. [10]. E. Bae, D. Ying, D. Kramer, V. Patsekin, B. Rajwa, C. Holdman, J. Sturgis, V. J. Davisson, J. P. Robinson, Portable bacterial identification system based on elastic light scatter patterns, Journal of Biological Engineering, Vol. 6, Issue 12, 2012. [11]. U. Minoni, A. Signoroni, G. Nassini, On the application of optical forward-scattering to bacterial identification in an automated clinical analysis perspective, Biosensor Bioelectronics, Vol. 68, 2015, pp. 536-543. [12]. P. R. Marcoux, M. Dupoy, A. Cuer, J. L. Kodja, A. Lefebvre, F. Licari, et al., Optical forwardscattering for identification of bacteria within microcolonies, Applied Microbiology and Biotechnology, Vol. 98, 2014, pp. 2243-2254. [13]. Buzalewicz, A. Wieliczko, H. Podbielska, Influence of various growth conditions on Fresnel diffraction patterns of bacteria colonies examined in the optical system with converging spherical wave illumination, Optics Express, Vol. 19, 2011, pp. 21768-21785. [14]. I. Buzalewicz, A. Wieliczko, K. J. Bednarek, H. Podbielska, Diffraction signature of bacteria colonies and the influence of different incubation conditions, in Proceedings of the Conference Frontiers in Optics 2011/Laser Science XXVII, Washington, DC, USA, 2011, p. JWA6. [15]. A. Suchwalko, I. Buzalewicz, H. Podbielska, Statistical identification of bacteria species, in Microbial Pathogens and Strategies for Combating Them: Science, Technology and Education (A. Méndez-Vilas, Ed.), Formatex Research Center, 2013, pp. 711-721. [16]. A. Suchwalko, I. Buzalewicz, H. Podbielska, Identification of bacteria species by using morphological and textural properties of bacterial colonies diffraction patterns, Proceedings of SPIE, Vol. 8791, 2013, 87911M. [17]. A. Suchwałko, I. Buzalewicz, A. Wieliczko, H. Podbielska, Bacteria species identification by the statistical analysis of bacterial colonies Fresnel patterns, Optics Express, Vol. 21, Issue 9, 2013, pp. 11322-11337. [18]. A. Suchwałko, I. Buzalewicz, H. Podbielska, Bacteria identification in an optical system with optimized diffraction pattern registration condition supported by enhanced statistical analysis, Optics Express, Vol. 22, Issue 21, 2014, pp. 26312-26327. [19]. S. Nagai, Y. Nishizawa, M. Onodera, S. Aiba, Effect of dissolved oxygen on growth yield and aldolase activity in chemostat culture of azotobacter vinelandii, Journal of General Microbiology, Vol. 66, 1971, pp. 197-203. [20]. J. W. Goodman, Introduction to Fourier Optics, 3rd edition, Robert & Company Publishers, 2005. [21]. I. Buzalewicz, M. Kujawińska, W. Krauze, H. Podbielska, Novel perspectives on the characterization of species-dependent optical signatures of bacterial colonies by digital holography, PLoS ONE, Vol. 11, Issue 3, 2016, 0150449. [22]. I. Buzalewicz, K. Liżewski, M. Kujawińska, H. Podbielska, Degeneration of Fraunhofer diffraction on bacterial colonies due to their light focusing properties examined in digital holographic microscope system, Optics Express, Vol. 21, 2013, pp. 26493-26505. [23]. P. E. Bae, P. P. Banada, K. Huff, A. K. Bhunia, J. P. Robinson, E. D. Hirleman, Biophysical modelling of forward scattering from bacterial colonies using scalar diffraction theory, Applied Optics, Vol. 46, Issue 7, 2007, pp. 3639-3648. [24]. M. A. Bees, P. Andresen, E. Mosekilde, M. Givskov, The interaction of thin-film flow, bacterial swarming and cell differentiation in colonies of Serratia liquefaciens, Journal of Mathematical Biology, Vol. 40, Issue 1, 2000, pp. 27-63. 195 Advances in Optics: Reviews. Book Series, Vol. 3 [25]. H. Kim, A. K. Singh, A. K. Bhunia, E. Bae, Laser-induced speckle scatter patterns in Bacillus colonies, Frontiers in Microbiology, Vol. 5, 2014, 537. [26]. E. Bae, N. Bai, A. Aroonnual, J. P. Robinson, A. K. Bhunia, E. D. Hirleman, Modeling light propagation through bacterial colonies and its correlation with forward scattering, Journal of Biomedical Optics, Vol. 15, Issue 4, 2010, pp. 045001 (1-10). [27]. D. Joyeux, S. Lowenthal, Optical Fourier transform: What is the optimal setup?, Applied Optics, Vol. 21, 1982, pp. 4368-4372. [28]. J. D. Gaskill, Linear Systems, Fourier Transform and Optics, John Wiley & Sons, New York, 1978. [29]. M. D. Abràmoff, P. J. Magalhães, S. J. Ram, Image processing with ImageJ, Biophotonics International, Vol. 11, Issue 7, 2004, pp. 36-42. [30]. A. Suchwalko, I. Buzalewicz, H. Podbielska, Computer-based classification of bacteria species by analysis of their colonies Fresnel diffraction patterns, Proceedings of SPIE, Vol. 8212, 2012, 82120R. [31]. P. Dalgaard, Introductory Statistics with R, Springer New York, New York, 2008. [32]. Feature Extraction: Foundations and Applications (Studies in Fuzziness and Soft Computing) (I. Guyon, S. Gunn, M. Nikravesh, L. Zadeh, Eds.), Springer, 2006. 196 Chapter 8. Integrated Terahertz Planar Waveguides for Molecular Sensing Chapter 8 Integrated Terahertz Planar Waveguides for Molecular Sensing Borwen You and Ja-Yu Lu1 8.1. Introduction Terahertz (THz) waveguides have been developed and basically categorized as the dielectric and metal waveguides with simple media to guide THz waves [1]. For the further waveguide development of molecular sensing applications, the simple waveguide medium is not so efficient to detect the target molecules because of two sensing issues, the poor lateral-field overlap and the power-limited propagation length. Therefore, the waveguide structures composed of multiple layers of dielectrics and metals are critical to achieve the engineering purpose for sensing analytes in a waveguide medium. Eventually, the target analyte is sensitively detected by THz waves based on the efficient interaction in the optimal waveguide length and modal field. The porous dielectrics are commonly the low-loss waveguide media to deliver THz waves for the long-distance transmission because the high attenuation media reduces inside the modal field instead of the air space [2]. The solid-core dielectrics are operated as subwavelength-scaled waveguides to weakly guide THz waves inside the solid media, spreading most of the modal field outside the core, i.e., toward the surrounding air space [3]. Such dielectric waveguides based on large percentages of air space successfully realize the low-loss transmission in THz frequency region and can be applied as the important optical components for THz fiber communication. However, THz fiber mode intensity is so weak without strong interaction ability to detect analytes around the fiber cores. To achieve the sensing purpose, the evanescent field of a THz fiber sensor should be powerful or the sample amount should be large to identify analytes in power absorption [4]. The metal medium in THz waveguides mainly works as the reflective surface to confine THz waves and propagate along the hollow core space. The parallel-plate waveguide (PPWG) is the typical structure and simply assembled by two metal plates to efficiently Ja-Yu Lu Department of Photonics, National Cheng Kung University, Tainan, Taiwan 197 Advances in Optics: Reviews. Book Series, Vol. 3 deliver THz-broadband waves [5]. Although THz-PPWG is an enclosed waveguide structures, the open frame metal surface is also able to guide THz waves as SommerfeldZenneck surface waves [6]. Because THz wave frequencies are considerably lower than the metal plasmon frequencies, THz waves cannot penetrate metals to directly excite surface plasmon polaritons (SPPs). Hence, THz Sommerfeld-Zenneck surface waves propagate in the delocalized manner on metal surfaces, which are difficult to detect those analytes on the metal surfaces with very small across sections. Tightly confined and powerful surface waves in THz frequency are the solution, i.e., in the high field intensity, to achieve the sensitive detection purpose for analytes on a metal surface. However, surface-confined waves are commonly suffered from very strong attenuation at the waveguide medium without powerful interaction on the analytes. The metamaterial concept currently shows it is possible to flexibly tune THz wave confinement and the longest propagation length based on certain periodic metal structures, and the molecular sensing ability can be enhanced via the periodic metal structures due to the distinctly strong THz resonance or interference. In this chapter, integrated planar THz waveguides can be used for bio-chemical sensing applications when the composed metal and dielectric layers are engineered to control THz waves with the sufficient propagation lengths and the optimal lateral fields. The presented THz waveguide sensor basically contains three main parts, including the waveguide, superstrate, and analyte. In the chapter, the waveguide sensing mechanism is investigated form the waveguide transmission spectrum of the evanescent field. The detection sensitivity is significantly affected by the decay length of the evanescent waveguide mode, which is determined by the refractive indices or thicknesses of the waveguide and superstrate layers. Two kinds of THz waveguides are presented in this chapter, including the metal-grating and -rod-array structures, respectively, for THz-frequency- and -phasesensitive waveguide sensors. The detection sensitivity of a THz planar integrated waveguide sensor based on the periodical metal structures can be optimized by adjusting their geometrical parameters of metal, instead of modifying the refractive index of the waveguide/superstrate layers with various dielectric materials. 8.2. THz Frequency Sensitive Detection 8.2.1. Waveguide Configuration and Terahertz Spectral Characterization The waveguide configuration with THz-frequency sensitive response in molecular sensing is illustrated in Fig. 8.1 and demonstrated as a THz hybrid plasmonic waveguide [7, 8]. There are two planar constructions, containing a 220 mm-long dielectric ribbon and a 50 mm-long metal grating. The dielectric ribbon is made of the polyethylene (PE) material with a 15 mm width, a 220 mm length and a 20 m thickness. Because the ribbon thickness is smaller than the diffraction-limited-beam size of the guided THz waves (0.1-1 THz), the modal field range along the Y axis is not confined to propagate, resembling a THz plastic-wire waveguide with strong evanescent powers outside the core [3]. Such the weak-confined THz wave propagation is performed in the 17 cm-long PE ribbon after the input end, and then interacts the attached metal grating, performing as the 198 Chapter 8. Integrated Terahertz Planar Waveguides for Molecular Sensing hybrid THz plasmonic waveguide (Fig. 8.1). The metal grating is considered as a metallic diffraction grating because the structural sizes and an operated field wavelength are equivalent. The metal grating is prepared from periodically perforated slits on one brass sheet, where the structural period and slit width are, respectively, 1.5 and 1 mm. The brass gratings with three different thicknesses (100, 200 and 400 m), and two different slit widths (1 and 0.5 mm) are experimentally measured in THz spectroscopy to optimize the power transmission and field confinement (Y-axis) at the integrated waveguide section, i.e., the metal grating waveguide in Fig. 8.1. Contrarily, THz field along the waveguide width (X-axis) is ignored in the study because the width is sufficiently larger than that of input THz wave beam size, corresponding to a beam divergence along an unlimited axis without boundary. Fig. 8.1. A THz hybrid plasmonic waveguide (reprinted from [7]). THz wave polarization along the ribbon waveguide is consistent with that along the metal grating to smoothly deliver THz waves to perform a hybrid plasmonic waveguide at the overlapping section. Basically, a dielectric ribbon waveguide can support transverse magnetic (TM)-polarized THz waves while the electric field oscillation of the coupled waves is perpendicular to the ribbon surface [Fig. 8.1]. The proposed hybrid plasmonic waveguide consequently enables the THz wave transmission that are straightforwardly coupled from the PE ribbon. The TM-polarized modal field to enter the metal grating section are stabilized via the 170 mm-long propagation on a PE ribbon, where a large portion of waveguide mode is in the air cladding. When THz waves are illuminated on the metal grating from the air cladding of a ribbon waveguide, partial power transmits and reflects due to the mismatch in modal size or waveguide refractive index. The reflected THz waves from the meatal grating follows the momentum conservation relation,    K in  K   K R , (8.1) where the vectors of Kin, KR and Kare the propagation constants of the input, reflected THz waves and the wave vector of a metal grating respectively. The grating wave vector, K, equals 2m/ where m and  are, respectively, the Bragg diffraction order and a structural period. The directions of propagation constants, Kin and KR, are opposite but 199 Advances in Optics: Reviews. Book Series, Vol. 3 have the same value, 2neff/C where , C and neff are, respectively, THz wave frequency, the light speed in a vacuum and an effective waveguide refractive index. Therefore, the diffraction grating can reflect THz waves exactly at Bragg frequencies in different orders, derived as mC/2neff based on the momentum conservation relation. Fig. 8.2 (a) illustrates THz power transmittance of the 50 mm-long metal grating at different metal thicknesses. The transmittance is measured from the transmission power behind of the 220 mm-long ribbon waveguides with and without attaching the metal grating. The 100 m-thick grating waveguide has the largest transmission spectrum and highest transmittance, comparing to those of the 200 and 400 m-thick gratings. It means the wave transmittance along the 200 and 400 m-thick gratings are obviously restricted, only delivering low frequency waves, respectively, less than 0.285 and 0.250 THz. There are two transmission dips around 0.3 and 0.4 THz for all the three gratings. The low transmittance feature is caused from Bragg reflection among the periodical slits of the metal grating under the phase-matching condition, which are the 3rd- and 4th-order Bragg frequencies for the 1.5 mm structural period. For increasing the metal thickness, the power distinction at the spectral dip increases and the related dip frequency is also red shifted (Fig. 8.2 (a)). The measured results show the 3rd- and 4th-order transmission dips for the 100 m thick grating waveguide are, respectively, at 0.3 and 0.4 THz, exhibiting shallow spectral depths. For the 200 m-thick grating, the 3rd- and 4th-order transmission dips are shifted to 0.296 THz and 0.396 THz with lower transmittance, about 0.005 and 0.002, respectively. When the metal thickness is increased to 400 m, the 3rd- and 4th-order transmission dips are further red shifted to 0.280 THz and 0.380 THz, with the lowest transmittances about 0.003 and 5×10-5, respectively. The power-decrement and red-shift effects of the Bragg-reflection waves for the large-thickness gratings are both resulted from the raised scattering cross section [9] because the large slit depth (or the increased metal thickness) causes strong scattering and deflects the waves with THz frequency lower than the Bragg frequency. Fig. 8.2. Transmittance of different 50 mm-long metal gratings with a 1.5 mm structural period, having (a) the same slit width of 1 mm in different metal thicknesses; (b) the same metal thickness of 0.2 mm with different slit-widths (reprinted from [7]). 200 Chapter 8. Integrated Terahertz Planar Waveguides for Molecular Sensing Besides of the grating thickness to affect the waveguide spectrum, two different slit widths of a 200 m-thick grating are observed in the study. Fig. 8.2 (b) shows the grating waveguide transmittance for the slit widths of 1.0 and 0.5 mm based on the same structural period of 1.5 mm. The transmittances at 0.296 and 0.396 THz are clearly raised about one order of magnitude when the 1.0 mm-slit width reduces to 0.5 mm. This indicates that the wide slit width causes the large scattering cross section, resulting in the reduced transmittance. Such low transmittance also represents the large slit width of the grating waveguide makes high evanescent power transferred from the ribbon waveguide mode into the THz surface plasmon polariton (SPP) modes that are confined on the periodical metal structure. Therefore, suitably tailoring the geometrical parameters of the metal grating at the thickness and slit width enables the optimal coupling efficiency of the SP modes along this THz hybrid plasmonic waveguide. 8.2.2. Integration of a Superstrate and the Sensing Method THz waves generally cannot be confined to a general metal surface with the surface plasmons because THz frequency is much lower than that of the intrinsic plasmon frequency of a metal. Periodical structures on a metal surface, such as 2D hole arrays [10], periodic slits [11], and various patterns of metamaterials [12, 13], have been demonstrated as THz spoof surface plasmons (THz-SSPPs) and thus enhance lateral field confinement. The THz-SSPPs can be generated at the grating structure of the THz hybrid plasmonic waveguide and have the features of the tightly lateral confinement, long transmission lengths, and distinct spectral dips related to the refractive index of the analyte, which are the critical specifications of a sensitive waveguide sensor. Those THz propagation waves on the hybrid waveguide are generally called as THz surface plasmonic waves (SPW) because of the optical performances in wave guidance, local field resonance and surface field confinement. THz-SPWs are generated by transferring the plastic ribbon waveguide modes to THz-SSPPs by means of the integrated diffraction metal grating structure [14, 15]. The generated THz-SPWs can be subwavelength confined to propagate on a 50 mm-long grating metal surface, and resonantly reflected under the phase matching condition of Eq. (8.1). Fig. 8.3 schematically illustrates the side view of a hybrid plasmonic waveguide for THz wave sensing. The diffraction grating is made of a 200 m thick brass with a 1.5 mm structural period, including air and brass sections with lengths of 1.0 and 0.5 mm, respectively. The evanescent field of a THz-SPW can be sensitively modified by the attached analytes on a superstrate, covering the metal grating. Such sensing scheme enables THz-SPW interacting the analytes on the grating for a sufficiently long distance to generate obvious spectral response for sensing applications. To test the sensing capability of the hybrid plasmonic waveguide based on the THz-SPW resonance, PE-film superstrates with thicknesses of 20, 50, and 90 m are attached to the grating surface. Fig. 8.4(a) shows the transmitted spectrum of the hybrid plasmonic waveguide, attached with various superstrate thickness. The original spectral dip at 0.38 THz for the blank hybrid plasmonic waveguide obviously shifts to the low-frequency range while the superstrate is attached on the grating. However, the spectral shift of the 3rd-order SPW resonance at 0.296 THz does not occur because the 0.296 THz modal field is not efficiently coupled to the metal grating to form the THz-SSPPs. In the low Bragg 201 Advances in Optics: Reviews. Book Series, Vol. 3 resonance mode shown in Fig. 8.2(a) (i.e., 1st- and 2nd- Bragg order), the modal field is dominated by the TM modes of the ribbon waveguide to directly pass through the outer region of the metal grooves without being modulated by the metal grating. Thus, the surface wave of 0.296 THz directly propagates along the attached PE film with weak resonance reflection by the metal grating. Contrarily, a large portion of the modal field for the 4th-order Bragg resonance at 0.380 THz is confined as THz-SSPPs inside the metal grating, and a small portion of the modal field evanesces in the air cladding with a subwavelength modal tail strongly interacting with the covered superstrate. The Bragg resonance reflections of THz-SSPPs are greatly influenced by the effective refractive index of the covered thin film because of the waveguide lateral field matching to the superstrate thickness. Therefore, the frequency of the 4th-order SPW-Bragg resonance, 0.380 +/– 0.020 THz, can be completely modified when the effective refractive index of the waveguide is changed by the attached superstrate. Based on the phase matching condition, the resonant frequency is defined in Eq. (8.2) and inversely proportional to the effective refractive index of the waveguide. ν mC . 2  n eff  Λ (8.2) Fig. 8.3. Molecular sensing scheme of a THz hybrid planar plasmonic waveguide (reprinted from [8]). Fig. 8.4 (b) summarizes the spectral positions of THz-SPW resonance, responding from the 20, 50 and 90 m thick superstrates on the grating, respectively, at 0.318, 0.281, and 0.263 THz. These spectral positions show the effective refractive indices of waveguide increase due to the additional superstrate coverage. As shown in Fig. 8.4 (c), the effective refractive index of a blank grating is 1.05 and raises to 1.25, 1.42, and 1.51, respectively, for 20, 50 and 90 m thick superstrates. The effective waveguide index increment with the superstrate thickness is determined from the power-occupied ratio of the modal field inside each superstrate along the Y-axis. The effective index can thus be estimated by the formula, neff = (1 – ).nair + nPE, where nairandnPE represent the volume ratio of the waveguide field occupied in a superstrate, refractive indices of air and PE-film superstrate, respectively. The occupied ratios of the resonant THz-SPWs in the 2050and 90 m thick superstrates are estimated as 0.5, 0.8, and 1.0, respectively, as shown in Fig. 8.4(d). The overall power distribution along the Y-axis outside the grating can be 202 Chapter 8. Integrated Terahertz Planar Waveguides for Molecular Sensing clearly covered by the 90 m thick superstrate, which leads the effective waveguide index to approach the refractive index of a PE material in THz frequency range [16]. Furthermore, the knife-edge measurement of the waveguide modal field expresses the power distribution along the X-axis of the hybrid waveguide is about 2 mm whatever the PE film is attached or not, representing the superstrate attachment is critical to THz-TM modal field confinement along the lateral dimension of Y-axis, not the waveguide width dimension in X-axis. Fig. 8.4. (a) Power transmission spectra for various thicknesses of PE films integrated to the metal grating, showing the performed (b) resonant frequencies, (c) relating effective refractive indices of the waveguide, and (d) the occupied power ratios inside each PE thin film (reprinted from [8]). The evanescent field of the propagated THz-SPWs is significantly confined to interact the analytes on the superstrate based on the Bragg resonant dip in a transmission spectrum. The decay lengths of the 4th-order Bragg resonance for various PE-film superstrates on the metal grating can be estimated from the standard PE film thickness (d) and the occupied field ratio () inside the PE films as shown in Fig. 8.4 (d). The lateral decay length (Y) of waveguide field is defined in the inset of Fig. 8.5 (a) and can be calculated from the relation, Y = d/. Fig. 8.5 (a) shows the calculated decay lengths of the resonant THz-SPWs at 0.318, 0.281, and 0.263 THz are approximately 40 m (~/21), 60 m (~/16), and 90 m (~/12) apart from the metal-grating surface in the Y direction, respectively. The evanescent decay lengths of THz-SPWs can also be obtained from the waveguide transmittance (T) based on the relation, T = (ꞏ.e-L)+(1-), where L denote the absorption coefficient (~1 cm-1) [16] and the PE-film superstrate length of 50 mm, respectively. The occupied field ratio ()of the guided THz-SPWs is thus expressed as  = (1- T)/(1-e-L). The measured transmittance T of a PE film in 0.160 ~ 0.280 THz is obtained from the transmission power in Fig. 8.4(a) and the decay lengths are then estimated, as shown in Fig. 8.5 (b). The estimated minimal decay lengths of the delivered THz-SPWs for PE thicknesses of 20, 50, and 90 m are approximately 40, 60, and 90 m (Fig. 8.5 (b)), respectively, which are approximately consistent with the measured results in Fig. 8.5 (a). The spectral ranges of the best confined THz-SPWs with the minimal decay 203 Advances in Optics: Reviews. Book Series, Vol. 3 length at the 4th-order Bragg resonance are around 70 GHz within 0.250~0.318 THz, 0.210~0.281 THz and 0.190~0.263 THz for 2050and 90 m thick PE films, respectively. It means THz-SSPPs are excited within these spectrum ranges, performing the best field confinement and the sensitive response of spectral shift for the ambient index variation. Fig. 8.5. (a) The decay lengths of the 4th-order resonant THz-SPW for different thicknesses of PEfilm superstrate on the grating. (inset) The definition of the decay lengths of THz-SPWs; (b) Frequency-dependent decay lengths of guided THz-SPWs for different thicknesses of PE-film superstrate (reprinted from [8]). 8.3. Phase Sensitive Detection 8.3.1. Waveguide Configuration and Terahertz Spectral Characterization The waveguide configuration with THz-phase sensitive response in molecular sensing is illustrated in Fig. 8.6 and demonstrated as a metal-rod-array (MRA)-based THz plasmonic waveguide [17]. The MRA-structural planar waveguide medium is composed of the uniform metal rods with 2D periodic arrangement. The MRA structure period () is determined from a rod diameter (D) and an air gap size (G). The rod diameter and height are, individually, 160 µm and 1 mm, considered as the structural unit of a MRA. The 1 mm-thick MRAs have sufficiently large cross section for the input THz beam, and applicable in the edge-coupled configuration as one slab waveguide. The propagation length of the MRA-based waveguide is 30 rows of MRA along the Y-axis, and the polarization of the input THz waves is in the X-direction, as shown in Fig. 8.6 (a) and perpendicular to the rod axis [18]. The MRA width is approximately 9 mm along the X-axis and considerably larger than the 1 mm-long rod. For the MRA slab-waveguide, THz wave divergence in the Z-axis would be confined, but a metal blade should be placed ahead of the input end of the MRA to prevent the detection of leaky and scattered THz waves before the stabilized modal field (Fig. 8.6 (a)). 204 Chapter 8. Integrated Terahertz Planar Waveguides for Molecular Sensing Fig. 8.6. (a) MRA waveguide configuration. Microscopic photos in the top view for the (b) 420 µm-and (c) 620 µm- MRAs. Microscopic photos in the side view for the (d) 420 µm-and (e) 620 µm- MRAs (reprinted from [17]). The MRA is fabricated from the cylinder polymer rods with a periodical arrangement in a square array, which is constructed through bottom-up 3D micro-stereolithography using a UV curable photopolymer [19, 20]. The high aspect ratio of a MRA cannot be prepared from the traditional mechanical or 3D printing methods. After the micro-stereolithography process, a 100 nm-thick aluminum metal film is deposited on the surface of polymer rods based on the sputter coating method. The coated metal thickness is larger than that of the skin depth of the invasive THz waves in 0.1-1 THz [21]. Figs. 8.6 (b) and 8.6 (c) show the fabrication results of MRAs for the top-view photographs of the 420 and 620 µm- MRA, respectively. Obviously, this micro-stereolithography fabrication method makes the periods along the X- and Y-dimensions approximately equal. Figs. 8.6 (d) and 8.6 (e) illustrate the side-view photographs of the MRAs, showing the straightness and uniformity of each rod. Figs. 8.7 (a) and 8.7 (b), respectively, show the normalized transmission spectra of the 420 and 620 µm- MRA waveguides. There are two transmission bands for these two MRA structures in 0.1-0.6 THz. The transmittance of 620 µm- MRA is obviously higher than that of the 420 µm- MRA because of the higher air-filling ratio among the rods. The rejection bands in Fig. 8.7 are resulted from the destructive interference of multiple reflections among the metal rods while THz waves enter the 30 rows of MRA. The central frequency of the rejection band is consistent to the Bragg reflection principle based on the equation of c/2n,where c, n, and are the light speed in vacuum, an effective refractive index, and a MRA period, respectively. The rejection band of the 420 µm- MRA ranges from 0.284 THz to 0.425 THz with a 141 GHz bandwidth (Fig. 8.7 (a)). For the 620 µm- MRA, the rejection band shifts to the lower frequency range, comparing to that of 420 µm- MRA, and ranges from 0.24 THz to 0.27 THz (Fig. 8.7 (b)). The spectral performance of two different MRA periods is consistent to those calculated results, using the finite-difference time-domain (FDTD) method. 205 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 8.7. Measured transmittance of (a) 420 µm-, and (b) 620 µm- MRAs (reprinted from [17]). 8.3.2. Integration of a Superstrate and the Sensing Method To apply the THz plasmonic MRA waveguide for the molecular sensing applications, one superstrate is required to be attached on the top of the rods with the solid analytes. The PP superstrates are full covered on the top surfaces of the 420 and 620 µm-MRA structures. The superstrates have the same width of 9 mm but different lengths of 13 and 18 mm, respectively, for 420 and 620 µm-MRAs. PP superstrates integrated to 420 and 620 µm- MRAs are discussed in this section for different thicknesses, including 30, 50, 70, and 90 µm. Fig. 8.8 (a) illustrates the transmitted THz electric field oscillations (E) of the 420 µm-MRA waveguide integrated with various PP superstrate thicknesses. The waveform of the blank device is obviously changed when the PP-superstrate is top integrated on the waveguide. The waveform main peaks shift as the superstrate thickness increases from 30 to 90 µm, where the main peaks for the 30, 50, 70, and 90 µm-thick superstrates are located at 19.8, 24.3, 24.8, and 29.3 ps, respectively. Such apparent time-domain shift of the waveform originates from the phase retardation along the PP film superstrate. Fig. 8.8. THz waves propagate through a 420 µm- MRA waveguide with and without attaching PP superstrates with various thicknesses: (a) Electric field oscillations, and (b) Power spectra (reprinted from [17]). 206 Chapter 8. Integrated Terahertz Planar Waveguides for Molecular Sensing The corresponding transmission spectra (|E|2) of the electric field oscillation of the PP superstrate integrated 420 µm-MRA are illustrated in Fig. 8.8(b). The rejection band for each PP superstrate thickness is approximately in 0.284-0.425 THz without obviously change, but the transmission power of the rejection and transmission bands increases as the thickness of the PP superstrate increases. Fig. 8.9(a) summarizes the peak power (|Epeak|2) at the low- and high-frequency transmission bands, which are denoted as 1st and 2nd peaks, respectively, for different superstrate thicknesses. The peak power in the transmission band gradually increases with a slight sinusoidal fluctuation as the superstrate thickness increases. The proportionally increasing effect between the transmission peak power and the superstrate thickness implies that the partial THz power of the confined modal field at the MRA-air cladding interface is coupled toward the aircladding region via the PP superstrate and then the interference interaction inside the MRA structure decreases. It also means the partial transmitted power that originates from the constructive and deconstructive interference of MRA is guided in the air-cladding space when a PP superstrate is attached to the top of the MRA structure. Fig. 8.9(b) schematically illustrates the power distribution of THz plasmonic wave guided on the 420 µm-MRA waveguide with a PP superstrate thickness larger than 90 µm. A large fractional power is distributed in the air-cladding region with a decay length more extended than that of the blank device, thereby increasing the THz power transmission. Fig. 8.9. (a) The power transmission of the peak power at the low- and high-frequency bands, respectively, denoted as the 1st and 2nd peaks for different PP superstrate thicknesses on a 420 µm- MRA; (b) Schematic description about the Z-axial power distribution of a 420 µm- MRA, integrated by a PP superstrate thickness larger than 90 µm (reprinted from [17]). Figs. 8.10 (a) and 8.10 (b) respectively show the measured THz time-domain waveforms and the corresponding transmitted power spectra for the 620 µm- MRA waveguide integrated with different PP superstrate thicknesses. The measured waveforms in Fig. 8.10 (a) are quite similar, except of the broadening oscillations. A continuous timedelay shift can thus be observed, where the second electric field peak are taken as the example in observation and displayed in the dashed lines. The correlating power spectra of the electric field oscillations show the MRA rejection and transmission bands still exist 207 Advances in Optics: Reviews. Book Series, Vol. 3 at similar spectral positions for attaching all thickness conditions. Fig. 8.11 (a) summarizes the peak powers of the 1st and 2nd transmission bands for different thicknesses of superstrates to integrate the 620 µm- MRA waveguide. The peak power obviously decreases even with a slight sinusoidal fluctuation as the superstrate thickness increases, contrary to the result of the 420 µm- MRA waveguide. It represents the extended decay length of the blank 620 µm- MRA waveguide becomes more concentrated inside the MRA after integrating a thick superstrate. In other words, the superstrate top conjugated on the 620 µm- MRA waveguide can confine the extending power in the air-cladding region toward the MRA structure. This phenomenon results in a more fractional THz power both inside the PP superstrate and the MRA structure to attenuate the transmission power. Fig. 8.11(b) schematically shows the modal power distribution of the superstrate integrated the 620 µm- MRA waveguide. Most of the modal power immerses inside the 620 µm- MRA structure and only a small amount of the modal power evanesces to the air-cladding region. The strong interference interaction is eventually reserved inside the MRA structure, like the blank waveguide condition. The transmission peak power variations represent a certain thickness of a PP superstrate integrated on a MRA waveguide certainly changes Z-axial modal power distributions (Figs. 8.9 (a) and 8.11 (a)). Because the lateral power distributions are different between the two MRAs, their detection sensitivities to detect analytes on the superstrate are certainly different. Fig. 8.10. THz waves propagate through a 620 µm- MRA waveguide with and without attaching PP superstrates with various thicknesses: (a) Electric field oscillations; (b) Power spectra (reprinted from [17]). The waveform variations in Figs. 8.8 (a) and 8.10 (a) express that the phase retardations of THz electric field oscillations interacting the superstrates are approximately proportional to the thickness variation. Figs. 8.12 (a)-8.12 (d) summarize the phase retardations induced by the 30, 50, 70, and 90 µm-thick PP superstrates attached on the 420 and 620 µm- MRA waveguides at frequencies of 0.520, 0.424, 0.322, and 0.226 THz, respectively. The normalized phase retardation () is obtained by comparing the phases of the transmitted THz wave with and without attaching superstrates and divided by the waveguide lengths. All the phase retardations at the four THz frequencies and along the two different MRAs are approximately proportional to the PP film 208 Chapter 8. Integrated Terahertz Planar Waveguides for Molecular Sensing thicknesses. The slopes of linearly fitting curves represent the phase detection sensitivities of the two MRA waveguides at different frequencies. The phase retardation contributed by a PP-film superstrate per micrometer thickness can be estimated from the linearly fit slopes. Fig. 8.11. (a) The power transmission of the peak power at the low- and high-frequency bands, respectively, denoted as the 1st and 2nd peaks for different PP superstrate thicknesses on a 620 µm- MRA; (b) Schematic description about the Z-axial power distribution of a 620 µm- MRA, integrated by a PP superstrate thickness larger than 90 µm (reprinted from [17]). Fig. 8.12. PP superstrate induced phase retardations on the 420 and 620 µm- MRA waveguides at different THz frequencies: (a) 0.520 THz; (b) 0.424 THz; (c) 0.322 THz, and (d) 0.226 THz; (e) Sensitivities of phase change detection (reprinted from [17]). Fig. 8.12 (e) summarizes the phase detection sensitivities at different THz wave frequencies for the two MRA waveguides. The sensitivity is approximately proportional to THz frequency in 0.20-0.55 THz. Obviously, the phase detection sensitivity of the 620 µm-MRA waveguide for different superstrate thicknesses is superior to that of the 420 µm-waveguide above 0.226 THz. The difference in phase detection sensitivity between the 420 and 620 µm-MRAs are resulted from different lateral power 209 Advances in Optics: Reviews. Book Series, Vol. 3 distributions of the two superstrate-integrated MRA waveguides (Figs. 8.9 (b) and 8.11 (b)). For the superstrate-integrated 620 µm-MRA waveguide, the lateral modal field has a considerably longer optical path in the 30 rows of MRA because most of the waveguide power is guided inside the MRA structure. Thus, the increased optical-pathdifference (OPD) induced by the superstrate-thickness increment per micrometer (Fig. 8.12 (b)) on the 620 µm-MRA waveguide promotes the phase retardation. Contrarily, most of the modal power guided on the superstrate-integrated 420 µm-MRA waveguide extends toward the air-cladding region, thereby reducing the optical path in the 30 rows of MRA. Consequently, the superstrate-induced OPD per micrometer in the multiple reflections of the 420 µm-MRA is smaller than that of the 620 µm- MRA. 8.4. Conclusions THz planar waveguides are experimentally studied based on the metal grating and rodarray structures, and further integrated with dielectric superstrates for the molecular sensing purpose. The metal-grating-based THz planar waveguide performs the frequencysensitive response to detect analytes on the superstrate because the generated THz-SSPPs in THz-SPWs are not only subwavelength confined to the metal surface but also delivered over a long distance with resonant reflection by the periodic metal structure. THz-SPW resonance follows the Bragg principle, and the transmission spectrum depends on the ambient refractive index of the grating. For the MRA-based THz planar waveguide, analytes on the superstrate can be sensitively detected and monitored from wave phase change of a THz electric field oscillation, categorized as a phase-sensitive THz planar waveguide sensor. An MRA-based THz plasmonic waveguide can be engineered via the interspace for the suitable modal field to integrate a superstrate, which has a significantly large optical path to interact analytes and is highly sensitive to the phase variation of the surrounding analytes. The superstrate integrated in the MRA or metal grating THz waveguides can be applied for loading various molecules in the thin-film or particle states. Such analyte-loaded superstrates can also be considered as the biochips or lab-on-a-chip to be detected by THz waves. Such detectable OPDs are considerably smaller than THz coherent length and valuable to investigate intermolecular attraction, perturbed by THz electromagnetic waves. Acknowledgements This work was supported by the grants in Ministry of Science and Technology of Taiwan (MOST 104-2221-E-006-163-MY3) and Japan Society for the Promotion of Science (JSPS), Grants-in-aid for scientific research (KAKENHI, JP17K45678). References [1]. S. Atakaramians, S. Afshar V., T. M. Monro, D. Abbott, Terahertz dielectric waveguides, Adv. Opt. Photonics, Vol. 5, Issue 2, 2013, pp. 169-215. 210 Chapter 8. Integrated Terahertz Planar Waveguides for Molecular Sensing [2]. S. Atakaramians, S. Afshar V., H. Ebendorff-Heidepriem, M. Nagel, B. M. Fischer, D. Abbott, T. M. Monro, THz porous fibers: design, fabrication and experimental characterization, Opt. Express, Vol. 17, Issue 16, 2009, pp. 14053-14062. [3]. L.-J. Chen, H.-W. Chen, T.-F. Kao, J.-Y. Lu, C.-K. Sun, Low-loss subwavelength plastic fiber for terahertz waveguiding, Opt. Lett., Vol. 31, Issue 3, 2006, pp. 308-310. [4]. M. Walther, M. R. Freeman, F. A. Hegmann, Metal-wire terahertz time-domain spectroscopy, Appl. Phys. Lett., Vol. 87, Issue 26, 2005, 261107. [5]. R. Mendis, D. Grischkowsky, Undistorted guided-wave propagation of subpicosecond terahertz pulses, Opt. Lett., Vol. 26, Issue 11, 2001, pp. 846-848. [6]. T.-I. Jeona, D. Grischkowsky, THz Zenneck surface wave (THz surface plasmon) propagation on a metal sheet, Appl. Phys. Lett., Vol. 88, Issue 6, 2006, 061113. [7]. B. You, J.-Y. Lu, W.-L. Chang, C.-P. Yu, T.-A. Liu, J.-L. Peng, Subwavelength confined terahertz waves on planar waveguides using metallic gratings, Opt. Express, Vol. 21, Issue 5, 2013, pp. 6009-6019. [8]. B. You, J.-Y. Lu, T.-A. Liu, J.-L. Peng, Hybrid terahertz plasmonic waveguide for sensing applications, Opt. Express, Vol. 21, Issue 18, 2013, pp. 21087-21096. [9]. J. G. Rivas, M. Kuttge, P. H. Bolivar, H. Kurz, J. A. Sánchez-Gil, Propagation of surface plasmon polaritons on semiconductor gratings, Phys. Rev. Lett., Vol. 93, Issue 25, 2004, 256804. [10]. C. R. Williams, S. R. Andrews, S. A. Maier, A. I. Fernandez-Dominguez, L. Martin-Moreno, F. J. Garcia-Vidal, Highly confined guiding of terahertz surface plasmon polaritons on structured metal surfaces, Nat. Photonics, Vol. 2, Issue 3, 2008, pp. 175-179. [11]. W. Zhu, A. Agrawal, A. Nahata, Planar plasmonic terahertz guided-wave devices, Opt. Express, Vol. 16, Issue 9, 2008, pp. 6216–6226. [12]. C. R. Williams, M. Misra, S. R. Andrews, S. A. Maier, S. Carretero-Palacios, S. G. Rodrigo, F. J. Garcia-Vidal, L. Martin-Moreno, Dual band terahertz waveguiding on a planar metal surface patterned with annular holes, Appl. Phys. Lett., Vol. 96, Issue 1, 2010, 011101. [13]. A. I. Fernández-Domínguez, E. Moreno, L. Martín-Moreno, F. J. García-Vidal, Terahertz wedge plasmon polaritons, Opt. Lett., Vol. 34, Issue 13, 2009, pp. 2063-2065. [14]. G. Nemova, R. Kashyap, Fiber-Bragg-grating-assisted surface plasmon-polariton sensor, Opt. Lett., Vol. 31, Issue 14, 2006, pp. 2118-2120. [15]. R. Kashyap, G. Nemova, Surface plasmon resonance-based fiber and planar waveguide sensors, J. Sens., Vol. 2009, 2009, 645162. [16]. J. W. Lamb, Miscellancous data on materials for millimetre and submillimetre optics, Int. J. Infrared Millim. Waves, Vol. 17, Issue 12, 1996, pp. 1997-2034. [17]. B. You, C.-C. Peng, J.-S. Jhang, H.-H. Chen, C.-P. Yu, W.-C. Lai, T.-A. Liu, J.-L. Peng, J.-Y. Lu, Terahertz plasmonic waveguide based on metal rod arrays for nanofilm sensing, Opt. Express, Vol. 22, Issue 9, 2014, pp. 11340-11350. [18]. A. Mazhorova, J. F. Gu, A. Dupuis, M. Peccianti, O. Tsuneyuki, R. Morandotti, H. Minamide, M. Tang, Y. Wang, H. Ito, M. Skorobogatiy, Composite THz materials using aligned metallic and semiconductor microwires, experiments and interpretation, Opt. Express, Vol. 18, Issue 24, 2010, pp. 24632-24647. [19]. J.-W. Choi, R. Wicker, S.-H. Lee, K.-H. Choi, C.-S. Ha, I. Chung, Fabrication of 3D biocompatible/biodegradable micro-scaffolds using dynamic mask projection microstereolithography, J. Mater. Process. Technol., Vol. 209, Issue 15-16, 2009, pp. 5494-5503. [20]. M. Farsari, F. Claret-Tournier, S. Huang, C. R. Chatwin, D. M. Budgett, P. M. Birch, R. C. D. Young, J. D. Richardson, A novel high-accuracy microstereolithography method employing an adaptive electro-optic mask, J. Mater. Process. Technol., Vol. 107, Issue 1-3, 2000, pp. 167-172. 211 Advances in Optics: Reviews. Book Series, Vol. 3 [21]. M. A. Ordal, L. L. Long, R. J. Bell, S. E. Bell, R. R. Bell, R. W. Alexander, C. A. Ward, Optical properties of the metals Al, Co, Cu, Au, Fe, Pb, Ni, Pd, Pt, Ag, Ti, and W in the infrared and far infrared, Appl. Opt., Vol. 22, Issue 7, 1983, pp. 1099-1120. 212 Chapter 8. Integrated Terahertz Planar Waveguides for Molecular Sensing Chapter 9 Integrated-Optics Solutions for Biomedical Optical Imaging B. Imran Akca1 9.1. Introduction Several novel optical imaging techniques have been developed over the past years to be used in basic biological research and clinical applications. Flourescence, confocal, Raman, and Brillioun microscopy are just a few of these optical techniques that are actively investigated by many research groups. Optical coherence tomography (OCT) is also one of these hot research areas which is a non-invasive, three-dimensional imaging technique that offers close-to-histology-level image quality [1]. Based on broadband spectral interferometry, OCT has enabled clinical applications ranging from ophthalmology to cardiology that revolutionized in vivo medical diagnostics. There are currently two distinct OCT technologies commercially available: time domain (TD) and Fourier domain (FD) OCT technology. Integrated optics offers unique solutions for OCT systems. Integrating several complex optical devices as miniaturized components on a single microchip improves mechanical stability for maintenance-free operation and accesses lithographic high-volume fabrication for dramatic cost reduction and improved repeatability. Miniaturized OCT systems have recently garnered attention for mainly their high potential in overcoming the size and cost problems of the bulky OCT systems [2, 3]. One major advantage of integrated optics is that the operation of the existing optical components can be reconfigured by controlling the material properties using temperature, voltage, or pressure. Despite this unique feature, OCT based upon integrated optical components has as yet not utilized it properly. In this chapter, by exploiting the unique features of integrated optics, the design of a novel multiple-reference TD-OCT system, an akinetic beam scanner, and high-speed spectrometers with ultra-high resolution or broad bandwidth are presented. The TD-OCT system and the akinematic beam scanner are designed to work at 1300 nm wavelength range whereas spectrometers are designed for 800 nm range. In conventional TD-OCT B. Imran Akca Institute for Lasers, Life and Biophotonics Amsterdam, Department of Physics and Astronomy, VU University Amsterdam, Amsterdam, The Netherlands 213 Advances in Optics: Reviews. Book Series, Vol. 3 systems depth scanning is achieved by modifying the relative optical path length difference of the reference and the sample arms in a sequential way using mechanical scanners. Due to the speed limitations and accompanying significant sensitivity decrease, and motion artifacts conventional TD-OCT systems fall behind FD-OCT systems in many applications. Formation of 2D images in many imaging techniques requires precise lateral scanning of the incident light beam using mechanical scanners such as galvanometer actuated mirrors. These mechanical scanners cause some major problems such as image distortion, phase errors, beam jitter, and inaccuracies due to non-uniform scan pattern, etc. Spectrometers are the core of many optical imaging modalities, such as Brillioun microscopy, SD-OCT, Raman microscopy etc. Having high speed, broad bandwidth, high resolution, and small footprint are the main requirements of a spectrometer. The designs that are discussed in this chapter are all comprised of electro-optic switches and very compact delay lines. Firstly, the electro-optic switch design for the devices working at 1300 nm wavelength range will be discussed and later on the details of each device design will be given. Spectrometers that are designed at 800 nm wavelength range will be discussed separately in the following section. 9.2. Designs at 1300 nm 9.2.1. Material System The proposed sample arm configuration was simulated for the lithium niobate (LN)-onsilicon waveguide platform as it is being one of the most versatile and well-developed active optical materials [4]. The material system is 300-nm-thick ion-sliced lithium niobate film on oxidized silicon wafer. The oxide thickness is 3 μm. The refractive index of the LN layer is 2.22 at 1300 nm, and its electro-optic (EO) coefficient is (r33 ~ 30 pm/V) [4]. Single mode rib waveguides with 0.2 µm of slab height and 1 µm of waveguide width were designed. The effective refractive index of the rib waveguide was calculated to be 1.85 by using beam propagation method (BPM) simulations. The three-dimensional illustration of the waveguide structure as well as the optical mode profile are given in Fig. 9.1. The minimum bending radius of the curved waveguides was calculated to be R = 100 µm with a bending loss of 0.01 dB/cm. The propagation loss of the LN waveguides defined by the ion-implantation-assisted wet etching is around 0.23 dB/cm. Metallic electrodes can be defined using gold or chromium. A 500-nm-thick silicon dioxide (SiO2) top cladding will be used to prevent propagation losses induced by the electrodes. The fiber-to-chip coupling losses (~6 dB) can be reduced to < 0.5 dB by using a high numerical aperture fiber [5]. 9.2.2. Working Principle of the Electro-Optic Switch One of the main components of the proposed designs is an electro-optic switch which is a wavelength-insensitive Mach-Zehnder-type interferometric coupler as illustrated in Fig. 9.2 [6, 7]. 214 Chapter 8. Integrated Terahertz Planar Waveguides for Molecular Sensing Fig. 9.1. (a) Three-dimensional view of the waveguide stack with relevant design parameters. (b) Beam propagation method simulation of the optical mode. The blue outline shows the crosssectional profile of the waveguide geometry. Fig. 9.2. (a) The schematic of the wavelength-insensitive electro-optic switch. An electrode is placed on top of the right arm of the coupler which is drawn in gray. (b) Beam propagation method simulation of the switch; (Left) No phase difference between coupler arms, the input light will stay in the same arm (bar state). (Right) For a π phase difference between coupler arms, the input light will cross-couple to the other arm (cross state). L, D, Δx are the lengths of the straight sections of the directional couplers, and the delay section, and the separation between coupler arms, respectively. It is comprised of a pair of directional couplers connected by a delay section in which a phase shift is introduced. The second directional coupler cancels deviations introduced by the former, if these deviations are similar in both couplers. An electrode is placed on the right arm of the electro-optic switch as shown in gray in Fig. 9.2(a). When the voltage is off, the lights on both arms will be in phase and the input light will stay in the same arm, i.e. bar state as illustrated in the left part of Fig. 9.2(b). With the applied voltage, the effective refractive index of that arm is locally increased due to the electro-optic effect 215 Advances in Optics: Reviews. Book Series, Vol. 3 which induces a phase difference between two arms. At a certain voltage value corresponding to a π phase difference between arms, the sample beam cross-couples to the other arm, i.e. cross state (Fig. 9.2(c) right part). It is also possible to achieve switching operation using pressure-induced refractive index change by choosing the material technology accordingly. 9.2.3. Akinetic Beam Scanner Layout and Its Working Principle The akinetic scanner design is comprised of two main components; namely electro-optic switches and two-mode interference based beam splitters/combiners. Fig. 9.3(a) shows the akinetic scanner layout implemented in an integrated-OCT system centered at 1300 nm. For simplicity, only 4 arms of the scanner are shown in the figure. Input light is divided into two arms with an integrated 3 dB beam splitter; half of it towards the reference arm which is integrated on the same chip for further size reduction, the other half towards the sample arms. Each sample arm consists of an electro-optic switch (Fig. 9.3(b)) for beam steering and a beam splitter/combiner (Fig. 9.3(c)). The electrooptic switch changes the propagation direction of the sample beam from bar state to cross state by an applied voltage corresponding to π phase difference between coupler arms. By activating each switch sequentially, the sample beam can be steered from one imaging location to the next until whole imaging range is scanned. Fig. 9.3. (a) Akinetic beam scanner implemented in an integrated-OCT system. Light coming from the input waveguide is divided into two; half towards on-chip reference arm, half towards the sample arm where several electro-optic switches are placed. The end of each sample arm is divided into two branches with a constant length difference, i.e. ΔL, between each for simultaneous imaging. (b) Schematic of the electro-optic switch. (c) Schematic of the two-mode interference based beam splitter/combiner. The splitting ratio is 50/50. (d) Due to ΔL, signals from two different physical locations on the sample will be detected at two different depth locations which are separated by 2ΔL. 216 Chapter 8. Integrated Terahertz Planar Waveguides for Molecular Sensing In order to double the imaging speed, each sample arm is divided into two branches with a certain length difference, i.e. ΔL, between them. In this way, two physical locations on the sample can be simultaneously illuminated. When images are formed, signals from two different locations will be detected at two different depths separated by 2ΔL as depicted in Fig. 9.3(d). The number of branches can be increased further in accordance with the desired speed improvement. The separation between each branch is chosen to be d = 20 µm. For a scanning range of 1 mm, 24 electro-optic switches are used which results in a scanner size of around 1 mm × 9 mm (9 mm2). Using the central part of a focusing lens, light can be successfully delivered into the tissue and collected back through same path. Returned signal from different sample locations are combined at the 3 dB coupler and interfered with light from the reference arm. A single detector and a high-speed data acquisition card is utilized to record interference signal from all beams simultaneously. The switching time of an electro-optic coupler is only few nanoseconds, (<10 ns), therefore scanning of a 1 mm wide area on the sample would take approximately 200 nanoseconds. However, in order to avoid data acquisition related problems and increase the integration time for higher signal to noise ratio, it is necessary to apply some time delay between each imaging point. Even for a long time delay, e.g. 1 millisecond, a reasonably high scanning speed, i.e. ~1 kHz, can be achieved. 9.2.4. Multiple-Reference TD-OCT Layout and Its Working Principle Fig. 9.4 is the schematic of the integrated-optics-based multiple-reference TD-OCT system in which the micro-chip is outlined by the red dashed-rectangle. For ease of understanding the first two levels of the light tapping mechanism are demonstrated. Here, a central wavelength of 1300 nm, an axial resolution of 20 µm, and a depth range of 1 mm are aimed at. Light coming from a broadband light source will be divided into two arms with a 3 dB coupler; half towards the sample, half towards the reference arm. There will be several electro-optically-controlled directional couplers placed on both sample and reference arms at certain distances. Imaging of different depths will be controlled by the additional length increment between consecutive reference beams (i.e. a in Fig. 9.4). According to the Nyquist sampling theorem the step size of the beam scanning should not be more than half of the axial resolution, i.e. 10 µm, which defines the length difference between consecutive reference points. Consequently, a is calculated to be 14 μm for this design by reckoning in the round trip of the light inside the tissue as well as the effective refractive indices of tissue and waveguides (ne (tissue) × 10 μm × 2 = ne (waveguide) × a, where ne (tissue) = 1.4, ne (waveguide) = 2.01)). For larger bandwidths, the group refractive indices of the waveguide and tissue have to be used for calculating a. There is no restraint on the additional length between tapping sections, i.e. d, however smaller d is favorable for compact devices. 9.2.5. Design Parameters of the Electro-Optic Switch The design of the electro-optic switch was made in two steps. Firstly, lengths of the straight coupling sections of the directional coupler and the delay part were calculated to 217 Advances in Optics: Reviews. Book Series, Vol. 3 achieve full coupling using equations given in [7] as L = 95 µm and Δx = 0.28 µm, respectively. The separation between coupler arms was chosen as D = 1 µm in order to reduce the overall device length to 0.65 mm. Secondly, the coupler designed in the first step was used to simulate the required mode effective refractive index increment for a π phase difference by scanning the refractive index difference (Δn) between coupler arms from 0 to 5 × 10-3 with 10-4 step size (Fig. 9.5(a)). It was found to be Δn = 2 × 10-3. The required voltage value to induce such index difference was calculated to be 21 Volts by using below equation [8] 1 V n V   ne3 r33 Γ, 2 t (9.1) for an overlap factor of Г = 0.3, electro-optically active layer thickness of t = 0.3 μm, effective refractive index of ne = 1.85, and electro-optic coefficient of r33 = 30 pm/V. The coupling loss was simulated to be 0.04 dB. The simulated splitting ratio between two arms stays constant over 100 nm bandwidth, as shown in Fig. 9.5(b). The coupling ratio remains the same even after a certain voltage applied on one arm of the coupler (Fig. 9.5(b), bottom). Fig. 9.4. Schematic of the proposed multiple-reference arm TD-OCT system based on integrated optics. Electro-optically-controlled directional couplers act as optical switches, which keep it on the same arm when there is no voltage, and cross-couple the light when there is a π phase difference due to electro-optic effect. The cross-coupled light from reference and sample arms will be combined at the beam combiner and sent to a photodetector (PD). Imaging of different depths will be controlled by the additional length increment between consecutive reference beams, i.e. a. The micro-chip is outlined by the red dashed-rectangle. 218 Chapter 8. Integrated Terahertz Planar Waveguides for Molecular Sensing Fig. 9.5. Simulation results of the electro-optic switch (a) Refractive index difference between coupler arms versus power on the same arm (I2). For Δn = 2 × 10-3, 99 % of the input light is crosscoupled to the other arm. (b) The coupler is wavelength independent for a wavelength range of 100 nm, and its wavelength independency does not change after the voltage is turned on. The change in coupling ratio of the electro-optic switch due to the process non-uniformity and limitations in reproducibility has been investigated. The refractive index of the cladding layer can have non-uniformities of up to ± 3 × 10-4, and the core layer can show thickness variations up to ± 1 % over the wafer. The waveguide width can vary by ± 0.1 μm. The simulation results of the effects of these process-dependent deviations are summarized in Table 9.1. The wavelength-independent couplers used in electro-switches are relatively fabrication tolerant devices as indicated in Table 9.1. Variations in the refractive index of the cladding layer has the minimum effect on coupling ratio whereas the maximum variation in coupling ratio was calculated to be 0.2 % for ±1 % change in core thickness which is still insignificant. Table 9.1. The Effect of the Technological Tolerances on Electro-optic Switch Performance. Parameters Δw = ± 0.1 µm δdcore = ± 1 % Δncladding = ± 3×10-4 w = 1.1 µm w = 0.9 µm dcore = 303 nm dcore = 297 nm ncladding = 1.4488 ncladding = 1.4482 Effective index change (×10-3) 5.5 -6.5 2.6 -2.6 0.2 -0.1 Coupling ratio change (%) 0.09 -0.1 0.2 -0.2 0.01 -0.009 9.2.6. Two-Mode Interference Beam Splitter/Combiner Design The beam splitter/combiner used in the end of each arm of the TD-OCT system as well as in the akinetic scanner is based on two-mode interference (TMI). It is wavelength independent and compared to an optical Y junction it is more fabrication tolerant, and 219 Advances in Optics: Reviews. Book Series, Vol. 3 reproducible. Fig. 9.6(b) and Fig. 9.6(c) demonstrate the beam propagation simulation results of the TMI-based beam splitter and combiner, respectively. The separation between input waveguides, the width and the length of the slab region are h = 0.8 μm, w = 3 μm, and l = 9 μm, respectively. The splitting ratio is constant over 200 nm wavelength range as shown in Fig. 9.6(d). The overall loss of the beam combiner/splitter was simulated to be 0.18 dB. Fig. 9.6. (a) Schematic of the TMI-based beam splitter (b) and combiner (c). The loss of the splitter and combiner was simulated to be 0.18 dB. (d) The splitting ratio remains constant over 200 nm bandwidth range. 9.3. High-Speed Spectrometer Designs Optical spectroscopy is an essential tool in numerous areas including biochemical sensing, material analysis, optical communication, and medical applications [9]. The development of a high-resolution on-chip spectrometer could enable compact, low-cost spectroscopy for portable sensing and increase lab-on-a-chip functionality. Motivated by this demand, several integrated microspectrometers have been realized in different configurations [10-13]. Most of these spectrometers rely on dispersive components which are inevitably bulky because their spectral resolution scales inversely with optical path length. Fourier transform spectroscopy (FTS) is a technique that uses interference of light rather than dispersion to measure the spectrum of a sample [14]. It is basically a Michelson interferometer with a movable mirror. The basis of this technique is the Fourier-pair relationship between the interferogram of a sample and its spectrum. The primary advantages of FTS compared to dispersive spectrometers are high optical throughput thereby greater signal-to-noise ratio, compact size, and relatively easily attainable high resolution which is constant over the entire spectral region as determined by the mirror displacement from the origin. 220 Chapter 8. Integrated Terahertz Planar Waveguides for Molecular Sensing Although FTS can be more compact in size, its scanning interferometric configuration makes it slow for some applications where speed is a critical constraint [15]. Spatial heterodyne spectroscopy (SHS) is an interferometric Fourier-transform (FT) technique based on a modified Michelson interferometer with no moving parts and relying on analysis of stationary interference patterns [16]. The SHS concept was successfully implemented in bulk optics so far, and recently it has been proposed for planar waveguide implementation by Cheben et al. as a Fourier-transform arrayed waveguide grating (FT-AWG) microspectrometer [12]. Florjańczyk et al., have generalized the waveguide SHS FT concept into a waveguide Mach-Zehnder interferometer (MZI) array which was based on an array of independent MZIs with different phase delays [17]. Even though it is a promising technique, it is still challenging to place long delay lines on a single wafer to achieve ultrahigh resolution. As a follow-up, they have presented a spiral-based SHS FT design with a spectral resolution of 40 pm, and footprint of 12 mm2 [18]. However, besides being quite lossy, the spectrometer size will still be significant if ultrahigh resolution is aimed at. In this section, two novel FT spectrometer layouts are introduced; one with an ultrahighresolution of 500 MHZ (~ 1 pm) in a very small footprint of 1 cm2 and other one with larger bandwidth (40 nm). The ultrahigh-resolution spectrometer design is comprised of N = 60 MZIs whereas large-bandwidth spectrometer has N = 80 MZIs that are sequentially activated by voltage-controlled directional couplers. Compared to spiral-based FT spectrometer described in [18], the ultrahigh-resolution spectrometer layout will provide much smaller size for the same resolution in addition its N times larger throughput. The long optical delay between MZI arms is introduced by sequentially tapping the propagating light out at several locations on the light path which makes the overall device size very compact. The tapping operation is provided by electro-optically-controlled directional couplers that are placed on both interferometer arms with a certain length difference between consecutive tapping locations. Lithium niobate (LN)-on-silicon material technology was chosen for these specific designs, however it can be applied to other electro-optic materials. The proposed designs can be easily adjusted to realize spectrometers with different bandwidth and resolution combinations. The electro-optic switch and the waveguide geometry were redesigned in accordance with the new wavelength range, i.e. 800 nm. Since the bandwidth range of the spectrometers is not more than 40 nm, directional couplers are preferred in electro-optic switch design in contrary to 1300-nm designs as it has a shorter length which reduces the device size significantly. 9.3.1. Material System at 800 nm The proposed spectrometer ideas are simulated for the LN-on-silicon waveguide platform. The material system is 250-nm-thick ion-sliced lithium niobate film on oxidized silicon wafer. The oxide thickness is 3 μm. The refractive index of the LN layer is 2.25 at 800 nm. Single mode rib waveguides with 0.2 µm of slab height and 0.9 µm of waveguide width were designed. Fig. 9.7(b) demonstrates the cross-sectional beam profile of the mode obtained by using beam propagation method (BPM). The minimum bending radius 221 Advances in Optics: Reviews. Book Series, Vol. 3 of the curved waveguides was calculated to be R = 150 µm with a bending loss of -0.005 dB/cm. A 500-nm-thick silicon dioxide (SiO2) top cladding will be used to prevent propagation losses induced by the electrodes. The simulated beam profile and the relevant waveguide parameters are given in Fig. 9.7(b). 9.3.2. Electro-Optic Switch Design at 800 nm Directional couplers used in both layouts were designed to act as voltage-controlled electro-optic switches with nanosecond switch time. A metallic electrode was placed on top of one of the straight waveguides of the directional coupler as shown in Fig. 9.7(a). Fig. 9.7. (a) The schematic of the electro-optically-controlled integrated- optics-based directional coupler. Here I1 is the input light, I2 is the transmitted light, I3 is the cross-coupled light, Lc is the electrode length, and d is the separation between coupler arms. (b) Beam propagation method simulation of the optical mode. The blue outline shows the cross-sectional profile of the waveguide geometry. Relevant waveguide parameters are given. (c) The amount of cross-coupling of the input light at different voltage values for different electrode lengths. The most optimum combination was obtained for an electrode length of 300 μm and an applied voltage value of V = 18 Volts. (d), right Voltage is OFF, the light will be cross-coupled to the other channel. (d), left Voltage is ON, V = 18 Volts, a π phase difference will be generated between coupler arms and the input light will stay in the same arm. When there is no voltage on the electrodes, the lights on both arms will be in phase and the incoming light will be cross-coupled to the other arm. (Fig. 9.7(d) left side). At a 222 Chapter 8. Integrated Terahertz Planar Waveguides for Molecular Sensing certain voltage value (i.e. V = 18 Volts for this design), a π phase difference is generated between coupler arms which avoids cross-coupling of the light and forwards it to the next stage where a different spectral information is obtained (Fig. 9.7(d) right side). This operation will be performed in a sequential order by switching the directional couplers on and off until all range is scanned. It is assumed that the coupling ratios at each directional coupler does not vary substantially across the entire bandwidth which is relatively small for the example considered here (i.e. 15 GHz). However, the non-uniformity of light coupling among individual directional couplers can be calibrated out as described in [17].The effective refractive index of the waveguide stack was calculated to be 2.0 for TE polarization. The expected change in the mode effective refractive index due to applied voltage (Δn (V)) was calculated to be 12 × 10-5 × V using Eq. (9.1) for an overlap factor of Г = 0.25, electro-optically active layer thickness of t = 0.25 μm, effective refractive index of ne = 2.0, and electro-optic coefficient of r33 = 30 pm/V. BPM simulations were performed for designing and optimizing the optical components. The directional coupler was designed in two steps. Firstly, in-phase case was designed and the separation between waveguides was calculated to be d = 0.9 μm for full crosscoupling (Fig. 9.7(d) left). There is a trade-off between the length of the electro-optically defined part of the coupler (i.e. electrode length, Lc) and the applied voltage. For a π phase difference this length is defined as: Lc  0 2  n (V ) . (9.2) In the second stage, the electrode length was scanned from 250 μm to 350 μm with a 50 μm step size while applied voltage value was scanned from 0 to 25 Volts in 1 Volt steps as given in see Fig. 9.7(c). The most optimum case was obtained for an electrode length of 300 μm and an applied voltage value of 18 Volts to generate a π phase difference. At this voltage level, light will stay in the same arm and be directed to the next MZI section (Fig. 9.7(d) right). 9.3.3. Ultrahigh-Resolution Spectrometer Layout and Its Working Principle Fig. 9.8 is the schematic of the ultrahigh-resolution FT spectrometer layout. For ease of understanding the first two levels of the light tapping mechanism are demonstrated. Here a central wavelength of 800 nm is aimed at. There are several MZIs that are electrooptically-controlled in a sequential order. Input light will be divided into two arms with an integrated 3 dB directional coupler. Half of the light will travel through a multi-S-shaped path that is comprised of several curved waveguides and straight waveguide sections. This arm will be used for providing additional length difference between MZI arms. The other half of the light will be sent towards a straight waveguide section that can be considered as the reference arm of the interferometer. The end of the both arms can be a waveguide termination such as a matched load that decreases reflection. 223 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 9.8. Schematic of the ultrahigh-resolution FT spectrometer design. The input light is divided equally by an on-chip 3 dB coupler and sent towards two different paths; one has several S-shaped waveguides and the other has a straight waveguide. There are several electro-opticallycontrolled directional couplers on both arms that act as optical switches. They cross-couple the light when there is no voltage, and keep it on the same arm when there is a π phase difference due to electro-optic effect. The cross-coupled light from both arms will be combined at the beam combiner and sent to a photodetector (PD). Here Ls is the length of the straight sections; R is the radius of the curved waveguides on the S-shaped path. Each MZI consists of two electro-optically-controlled directional couplers one in each arm. The first MZI will have zero path length difference between its arms, therefore the length of segments |AB| and |DE| are equal. The length of the next segment on the multiS-shaped path is |BC| = (2 × π × R) + (2 × Ls) where Ls is the length of the straight parts, and R is the radius of the curved waveguide sections. The length difference between the arms of the next MZI (i.e. |BC| – |EF|) is chosen in accordance with the resolution requirements. The resolution of the spectrometer is defined by the maximum delay ∆Lmax, which also determines the spectrometer size. The calculation of ∆Lmax refers to the Littrow condition and can be expressed as [18]: Lmax  02 ,   ng (9.3) where δλ is the resolution of the spectrometer, ng is the group refractive index of the waveguide stack, and λ0 is the center wavelength. For a spectral resolution of 500 MHz a maximum delay of ∆Lmax ≅ 30 cm is needed which makes it very challenging to accommodate such a spectrometer on a standard 10-inch wafer using existing waveguide SHS designs. The proposed design solves the size problem by using light tapping approach. As the light travels through a waveguide, the optical length gets increased, and by tapping the propagating light out at certain locations on the waveguide, the required optical delay can be obtained for each MZI. The multi-S-shaped sections of this waveguide will keep the length of the device short while straight sections in between (with 224 Chapter 8. Integrated Terahertz Planar Waveguides for Molecular Sensing length Ls) will mainly provide the optical delay. As a constraint on proper operation of the spectrometer, each MZI path length must be an integer multiple of some fixed path length, i.e. ∆Lmax/N. The light coming from both MZI arms will get interfered on an on-chip beam combiner and sent to a matched photodetector. The output power distribution digitally processed by using a Fourier transform to retrieve the input spectrum. Photodetectors can be fabricated on the same chip or a commercial photodetector array can be externally buttcoupled to the chip. Based on the Nyquist sampling theorem, in order to scan 15 GHz of bandwidth with 500 MHz resolution, N = 60 directional couplers are needed which results in an overall device size of around 2 cm × 0.5 cm (1 cm2). The time needed for scanning 15 GHz bandwidth in 60 steps will be less than a millisecond. 9.3.4. Broadband Spectrometer Layout and Its Working Principle Some applications require high speed spectrometers working over a large bandwidth. The previously explained spectrometer design provides ultrahigh resolution over a very small bandwidth. By changing the spectrometer layout slightly, one can achieve large bandwidth and high speed over a small footprint. Fig. 9.9 demonstrates the FT spectrometer working over 40 nm bandwidth with a spectral resolution of 1 nm at the central wavelength of 800 nm. Such a spectrometer can be used in Raman spectroscopy which can outperform in terms of size and speed compared to existing bulky counterparts. By using Eq. (9.3) the required maximum length difference is calculated to be 372 μm. Assuming 80 MZIs (according to the Nyquist sampling theorem), the required length increment for each step is calculated to be ΔL = 4.7 μm which can be provided by applying a curved waveguide section to one side of the MZI parts. The working principle of this design is as follows; input light will be divided into two arms with an integrated 3 dB directional coupler. Half of the light will travel through a path that is comprised of several curved waveguides and straight waveguide sections. This arm will be used for providing additional length difference between MZI arms. The other half of the light will be sent towards a straight waveguide section (with small curved sections for avoiding coupling of light between straight sections) that can be considered as the reference arm of the interferometer. The end of the both arms can be a waveguide termination such as a matched load that decreases reflection. In the case of no voltage on the first couplers and a certain voltage (i.e. V = 18 Volts) on the rest of the couplers, the light will cross couple to the other branch and stay in the same branch until it reaches the detector. At this stage, there will be no path length difference between two arms. When a certain voltage is applied to the first couplers while there is no voltage on the rest of the couplers, there will be a ΔL path length difference between two arms. By increasing the number of couplers that are under a certain voltage, the path length difference will be increased by ΔL at every step until the maximum path length difference is reached. The overall device size will be around 2.5 cm × 0.5 cm. The time needed for scanning 40 nm bandwidth in 80 steps will be less than a millisecond. 225 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 9.9. Schematic of the broadband FT spectrometer design. The input light is divided equally by an on-chip 3 dB coupler and sent towards two different paths; one has several curved waveguides and the other has a straight waveguide. There are several electro-optically-controlled directional couplers on both arms that act as optical switches. They cross-couple the light when there is no voltage, and keep it on the same arm when there is a π phase difference due to electro-optic effect. The cross-coupled light from both arms will be interfered at the photodetector (PD). Here ΔL is the extra path length difference between curved and straight waveguide sections. 9.4. Conclusions In summary, a novel OCT beam scanner, a TD-OCT system and two high-speed FT spectrometer designs were presented to be used in various biomedical optical applications such as OCT, confocal and Brillioun microscopy, and Raman spectroscopy. The proposed designs are comprised of dynamic delay lines which are based on voltage-controlled directional couplers. The proposed ideas can be applied to other imaging modalities as well as other measurement techniques. Different switching mechanisms (e.g. pressurebased) can be applied in different material platforms, depending on the power consumption, and switching speed requirements. It is expected that the layouts described in here will evoke some experimental interest and will be followed up by several research groups and companies. Acknowledgements The author thanks Dr. Bob van Someren and Prof. Ton van Leeuwen for the fruitful discussions. 226 Chapter 8. Integrated Terahertz Planar Waveguides for Molecular Sensing References [1]. D. Huang, E. A. Swanson, C. P. Lin, J. S. Schuman, W. G. Stinson, W. Chang, M. R. Hee, T. Flotte, K. Gregory, C. A. Puliafito, J. G. Fujimoto, Optical coherence tomography, Science, Vol. 254, 1991, pp. 1178-1181. [2]. B. I. Akca, B. Považay, A. Alex, K. Wörhoff, R. M. de Ridder, W. Drexler, M. Pollnau, Miniature spectrometer and beam splitter for an optical coherence tomography on a silicon chip, Opt. Express, Vol. 21, 2013, pp. 16648-16656. [3]. G. Yurtsever, B. Považay, A. Alex, B. Zabihian, W. Drexler, R. Baets, Photonic integrated Mach-Zehnder interferometer with an on-chip reference arm for optical coherence tomography, Biomed. Opt. Express, Vol. 5, 2014, pp. 1050-1061. [4]. E. L. Wooten, K. M. Kissa, A. Yi-Yan, E. J. Murphy, D. A. Lafaw, P. F. Hallemeier, D. Maack, D. V. Attanasio, D. J. Fritz, G. J. McBrien, D. E. Bossi, A review of lithium niobate modulators for fiber-optic communications systems, IEEE J. Sel. Top. Quantum Electron., Vol. 6, 2000, pp. 69-82. [5]. O. D. Herrera, K.-J. Kim, R. Voorakaranam, R. Himmelhuber, S. Wang, V. Demir, Q. Zhan, L. Li, R. A. Norwood, R. L. Nelson, J. Luo, A. K.-Y. Jen, N. Peyghambarian, Silica/electrooptic polymer optical modulator with integrated antenna for microwave receiving, J. Lightwave Technol., Vol. 32, Issue 20, 2014, pp. 3861-3867. [6]. B. I. Akca, C. R. Doerr, G. Sengo, K. Wörhoff, M. Pollnau, R. M. de Ridder, Broad-spectralrange synchronized flat-top arrayed-waveguide grating applied in a 225-channel cascaded spectrometer, Opt. Express, Vol. 20, Issue 16, 2012, pp. 18313-18318. [7]. B. E. Little, T. Murphy, Design rules for maximally flat wavelength-insensitive optical power dividers using Mach-Zehnder structures, IEEE Photon. Technol. Lett., Vol. 9, 1997, pp 1607-1609. [8]. B. Imran Akca, A. Dana, A. Aydinli, M. Rossetti, L. Li, A. Fiore, N. Dagli, Electro-optic and electro-absorption characterization of InAs quantum dot waveguides, Opt. Express, Vol. 16, 2008, pp. 3439-3444. [9]. N. V. Tkachenko, Optical Spectroscopy, Elsevier Science, 2006. [10]. E. Le Coarer, S. Blaize, P. Benech, I. Stefanon, A. Morand, G. Lerondel, G. Leblond, P. Kern, J. M. Fedeli, P. Royer, Wavelength-scale stationary-wave integrated Fourier-transform spectrometry, Nat. Photonics, Vol. 1, 2007, pp. 473-478. [11]. B. I. Akca, B. Považay, A. Alex, K. Wörhoff, R. M. de Ridder, W. Drexler, M. Pollnau, Miniature spectrometer and beam splitter for an optical coherence tomography on a silicon chip, Opt. Express, Vol. 21, Issue 14, 2013, pp. 16648-16656. [12]. P. Cheben, I. Powell, S. Janz, D.-X. Xu, Wavelength-dispersive device based on a Fouriertransform Michelson-type arrayed waveguide grating, Opt. Lett., Vol. 30, 2005, pp. 1824-1826. [13]. R. F. Wolffenbuttel, State-of-the-Art in integrated optical microspectrometers, IEEE Trans. Instrum. Meas., Vol. 53, 2004, pp. 197-202. [14]. M.-L. Junttila, J. Kauppinen, E. Ikonen, Performance limits of stationary Fourier spectrometers, J. Opt. Soc. Am. A, Vol. 8, 1991, pp. 1457-1462. [15]. G. Scarcelli, S. Hyun Yun, Confocal Brillouin microscopy for three-dimensional mechanical imaging, Nat. Photon., Vol. 2, 2008, pp. 39-43. [16]. J. Harlander, R. J. Reynolds, F. L. Roesler, Spatial heterodyne spectroscopy for the exploration of diffuse interstellar emission lines at far-ultraviolet wavelengths, Astrophys. J., Vol. 396, 1992, pp. 730-740. [17]. M. Florjańczyk, P. Cheben, S. Janz, A. Scott, B. Solheim, D.-X. Xu, Multiaperture planar waveguide spectrometer formed by arrayed Mach–Zehnder interferometers, Opt. Express, Vol. 15, Issue 26, 2007, pp. 18176-18189. 227 Advances in Optics: Reviews. Book Series, Vol. 3 [18]. V. Velasco, P. Cheben, P. J. Bock, A. Delâge, J. H. Schmid, J. Lapointe, S. Janz, M. L. Calvo, D. Xu, M. Florjańczyk, M. Vachon, High-resolution Fourier-transform spectrometer chip with microphotonic silicon spiral waveguides, Opt. Lett., Vol. 38, 2013, pp. 706-708. 228 Chapter 10. Video Based Heart Rate Estimation Using Facial Images from Video Sequences Chapter 10 Video Based Heart Rate Estimation Using Facial Images from Video Sequences Siong-Shi Ling, Yong-Poh Yu and Raveendran Paramesran1 10.1. Introduction Human heart rate is measured as the number of heart beats per minute (BPM). It is an important parameter used to reveal the health condition of an individual. The pattern of the measured heart rate can be used to indicate levels of fitness, the presence of disease, stress or fatigue and even blockages in the artery due to diabetes or high cholesterol level. Currently, the most common method used for human heart rate measurements is by using the Electrocardiography (ECG) machine. The electrodes are attached to the surface of the skin around the wrist and chest of the subject. The electrical activity of the human heart is captured through the attached electrodes. Heart rate measurements using ECG machine is a contact based method which might not be suitable for skin-burned patients and person with autistic disorder (sensitive to touch). Garbey et al. introduced a new approach for human cardiac pulse measurement based on thermal signal analysis of the major blood vessels near the skin surface (Garbey et al., 2007 [11]). The modulation of the temperature measured from these blood vessels is caused by the variations in blood flow. In the same year, Pavlidis et al. measured the human heart rate and breath rate through bio-heat modeling of facial imagery using a thermal camera (Pavlidis et al., 2007 [12]). The cardiac pulse detection at the forehead proposed by Gatto was extracted from the video infrared thermography (Gatto, 2009 [15]). This approach is based on the principle that the variations of blood flow during the cardiac cycle will cause the fluctuation of thermal energy released by the body tissue. Takano and Ohta developed a system to measure the human heart rate and respiratory rate based on the images from the Charge-Coupled Device camera (Takano & Ohta, 2007 S.-S. Ling Department of Electrical Engineering, Faculty of Engineering, University of Malaya, Kuala Lumpur, Malaysia 229 Advances in Optics: Reviews. Book Series, Vol. 3 [13]). The variations of the average brightness in the region of interest within the subject’s skin were recorded. These data were processed through a sequence of operations which involve interpolation, low pass filter and auto-regressive spectral analysis in order to obtain the heart rate and the respiratory rate. In the following year, Verkruysse et al. measured human respiration and heart rates through remote sensing of plethysmographic signals under ambient light using digital camera (Verkruysse et al., 2008 [14]). Jonathan and Leahy utilized the camera on the smartphone to capture a series of video frames of a human index finger (Jonathan & Leahy, 2010 [16]). The reflections of plethysmographic signals obtained from these video frames were used to compute the human heart rate. The engineering model created by Shi et al. was used for cardiac monitoring through reflection photoplethysmography (Shi et al., 2010 [18]). This noncontact model is made up of a light source that consists of a Vertical Cavity Surface Emitting Laser (VCSEL) and a photo-detector that consists of a high-speed silicon PiN photodiode. Photoplethysmography (PPG) is a non-invasive and inexpensive method to measure the variations of blood volume through the variations of light absorption or reflection (Kamshilin et al., 2011 [19]). The variations of blood volume in the blood vessels are due to the contraction and relaxation of heart muscles during each cardiac cycle. The relationship between the blood volume pulses and the light in reflection PPG has been investigated by some researchers (Hertzman, 1938 [1]; Weinman et al., 1977 [2]) since a few decades ago. The principle of PPG is based on the fact that body tissue is less opaque than the blood. Therefore, the increase in blood volume will reduce the intensity of the reflected light from the trans-illuminated tissue. The variations in blood volume will change the intensity of the reflectance accordingly. Therefore, the human heart rate which is the same as the frequency of cardiac cycle can be measured from the plethysmographic signals captured in the video. Heart rate measurement from video sequences is considered as low cost since the color can be captured using any available video recording device such as video camera, webcam or mobile phone. This remote and non-contact (without using any special device) heart rate measurement is very suitable for home-based health care applications and telemedicine. Poh et al. developed a non-contact technique to estimate the heart rate of a subject whose body was stationary (Poh et al., 2010 [17]; Poh et al., 2011 [20]). This contact-free approach is based on automatic face tracking and the use of blind source separation on color channels within the facial region. Besides that, their proposed method is robust to motion artifacts and able to extract the heart rate of multiple people at the same time. They showed that human heart rate can be measured from video recorder, such as webcam, under ambient light. However, the whole frontal face is used as the Region of Interest (ROI) which includes the regions with less or without blood vessels such as the eyes, hair and nostrils. Their model used a video with duration of 60 seconds that includes the entire facial region of a subject to compute the subject’s average heart rate variability for that duration. The Red, Green and Blue (RGB) pixel values of each video frame were used as the raw input signals. Blind source separation (BSS) method was utilized to extract the 230 Chapter 10. Video Based Heart Rate Estimation Using Facial Images from Video Sequences source signals (that contain the heart rate PPG signals) from the RGB input signals. The heart rate was calculated by using peak detection algorithm. The results obtained from their proposed method were compared to the ECG raw signals. Their results showed that BSS is able to extract the heart rate source signals from the facial images under stationary condition. Pursche et al. modified this technique by transforming the BSS source signals (the heart rate signals) into frequency domain (Pursche et al., 2012 [22]). They divided the facial region into three parts, and concluded that the area around eye and nose (center of the face region) provides better information compared to the other two parts. The time series signals were transformed into frequency domain using Fourier transform. They concluded that this method has higher correlation compared to the peak detection algorithm. On the other hand, Xu et al. showed a simplified mathematical model to obtain the BVP signals from images of human skin (Xu et al., 2014 [24]). They developed a model for pigment concentration in human skin, and used it to estimate the heart rate. They computed the heart rate readings from video recordings lasting from 45 s to 90 s. The subjects are required to keep still during the recording. Their heart rates did not vary much. Kumar et al. proposed a model, known as DistancePPG, to improve the signal-to-noise ratio of the camera-based PPG signal by combining the color change signals obtained from different regions of the face using a weighted average (Kumar et al., 2015 [25]). Additionally, they introduced a method to track different regions of the face separately to extract the PPG signals under motion. The method was evaluated on people having diverse skin tones, under various lighting conditions and natural motion scenarios. Kumar et al. concluded that the accuracy of heart rate estimation was significantly improved using the proposed method. Section 10.2 introduces the component analysis (both ICA and PCA). The way how component analysis (both ICA and PCA) can be used to estimate dynamic heart rate variation is discussed in Section 10.3. In Section 10.3, the experimental study consists of two experiments. The first experiment is on the dynamic heart rate estimation using facial images from video sequences. Although ICA can be used to separate the PPG signal from color components of a video clip, the amount of independence of the ICA sources may be decreased due to the short video duration. A method using ICA combined with mutual information is introduced to identify the minimum video duration needed to estimate the heart rate without compromising the accuracy of heart rate readings (Yu et al., 2015 [26]). In the second experiment we show that the PCA can be used to de-correlate the color components of the PPG signal to estimate the dynamic heart rates (Yu et al., 2015 [27]). This is possible because the color components in log-space are correlated to each other. Section 10.4 examines the impact of varying the distance between the subject and the video camera and also fixing the distance but varying the video duration. Section 10.5 concludes the study. 231 Advances in Optics: Reviews. Book Series, Vol. 3 10.2. Introduction to Component Analysis This section introduces both ICA and PCA. ICA is one of the blind source separation (BSS) methods. Generally, BSS is used to uncover the independent signals from a set of sensor observations that are linear mixtures of statistical independent sources. On the other hand, PCA is a way of identifying the patterns in a group of high dimensional data and expressing or analyzing the data by highlighting their similarities and differences (Lindsay, 2002 [9]). In other words, PCA is used to decorrelate the linearly dependent signals into uncorrelated signals. 10.2.1. Independent Component Analysis ICA is used to decorrelate the signals and reduce higher-order statistical dependencies (Lee et al., 2000 [6]). Assume that there are n linear mixtures (sensors) y1, …,yn of n independent components yj  mj1c1  mj 2c2 ... mjncn , for all j, (10.1) and each mixture yj as well as the independent component ck is the random variable, instead of a proper time signal. Let y denotes the mixture y, …,yn, c denotes c1, …,cn, and M denotes the mij, then (10.1) can be written as y = Mc (10.2) The statistical model in (10.2) is known as independent component analysis. It describes how the observed sensors yi are generated by a process of mixing the components si (Hyvärinen & Oja, 2000 [5]). The mixing matrix M is unknown but can be estimated. The independent components can be obtained by computing inverse of mixing matrix M, denoted by W. Hence, c = Wy (10.3) 10.2.2. Principal Component Analysis To utilize PCA, an important assumption has to be made, i.e. linearity (John, 2002 [8]). In other words, a new set of data can be formed as a linear combination of its basis vectors. Let A be the original data set, B be the representation of A and T be the linear transformation matrix that transforms A into B, then B = TA (10.4) Geometrically, T is a rotation and a translation matrix which transforms A into B. Considering both A and B are a m × n matrix, then the covariance matrix of A, CA can be defined as: 232 Chapter 10. Video Based Heart Rate Estimation Using Facial Images from Video Sequences CA = 1 AAT, n 1 (10.5) where AT is the transpose matrix of A. The respective eigenvectors and eigenvalues can be obtained from the covariance matrix. By sorting the eigenvalues in descending orders, the k eigenvectors that correspond to the k largest eigenvalues can be selected, where k is the number of dimensions of the new subspace. A new projection matrix T can be constructed from the selected k eigenvectors. Hence, the original dataset A can be transformed via T to obtain a k-dimensional feature subspace B, where the components of B are uncorrelated to each other. 10.3. Dynamic Heart Rate Estimation Using Component Analysis This section shows how dynamic heart rate measurements that are typically obtained from sensors mounted near to the heart can also be obtained from video sequences. A short video duration is needed for dynamic heart rate estimation. However, short video duration may decrease the amount of independence of ICA sources and may render to inaccurate readings. Hence, ICA is combined with mutual information to ensure accuracy is not compromised in the use of short video duration. Besides ICA, PCA may also be used to estimate the dynamic heart rates. It is found that the color components in log-space are correlated to each other. The color components in log-space can be de-correlated using PCA to recover the PPG signal. An important consideration for accuracy of the dynamic heart rate estimation is to determine the shortest video duration that realizes it. This video duration is chosen when the six principal components (PC) are least correlated amongst them. When this is achieved, the first PC is used to obtain the heart rate. 10.3.1. Experimental Setup All experiments were set up under constant office fluorescent light. A Sony camcorder (HDR-PJ260VE) was used for the video recording purposes. All videos were recorded and sampled at 50 frames per second. The camcorder was fixed at a position with a distance of about 0.60 m from the subject’s face. In both experiments, four subjects were selected and requested to carry out a cycling activity. All subjects were asked to cycle at different speeds for about two minutes. Then they were asked to stop for twenty seconds. The camcorder was used to capture their facial images during that time. Throughout the video recordings, all subjects were asked to remain stationary. For the first experiment, twenty heart rate readings (sampled at each second) were computed using ICA approach (Yu et al., 2015 [26]) for every subject. For the second experiment, another twenty heart rate readings (sampled at each second) were computed using PCA approach (Yu et al., 2015 [27]) for every subject. 233 Advances in Optics: Reviews. Book Series, Vol. 3 As reference, the instantaneous heart rates of each subject that obtained from the ICA and PCA methods were compared to the actual heart rate readings measured from Polar Heart Rate Monitor – Polar Team2 Pro (Schönfelder et al., 2011 [21]; Wallén et al., 2012 [23]). 10.3.2. Experimental Results Using ICA Method A total of 80 instantaneous heart rate readings were obtained for this experiment. In the experiment, the subjects’ heart rates were varying between 127 BPM and 153 BPM. Fig. 10.1 shows the detail workflow of the ICA method. The block diagram of the proposed model for dynamic heart rate measurement is illustrated in Fig. 10.1. Fig. 10.1. Workflow of ICA approach (Yu et al, 2015 [26]). After the ROI of each frame was identified, the mean of pixel values for red (R), green (G) and blue (B) components were computed separately, where  μR: the mean of all pixel values for R component;  μG: the mean of all pixel values for G component;  μB: the mean of all pixel values for B component. 234 Chapter 10. Video Based Heart Rate Estimation Using Facial Images from Video Sequences The respective μR, μG, and μB of all these 50 continuous frames were calculated. Therefore, at that instant, a set of three raw sensors R(n), G(n), B(n) were formed. Each raw sensor consists of 50 elements. The set of raw sensors were then detrended using algorithm developed by Tarvainen et al. (Tarvainen et al., 2002 [10]). ICA model developed by Cardoso and Souloumiac (Cardoso, 1999 [4]; Cardoso and Souloumiac, 1993 [3]) was then used to separate one set of 3 independent sources from the set of sensors. The set of ICA sources were bandpass filtered (128-point Hamming window, 0.6-4 Hz), and the mutual information was applied to obtain the independence of the ICA sources. The entire process was repeated by increasing the number of previous frames, one-by one until it fulfilled the criterion. The criterion was based on the convergence of the curve fitting coefficients. Table 10.1 summarizes the details of the computed heart rate readings of all subjects. The highest mean absolute error and the highest standard deviation of absolute errors are 1.48 and 1.19 BPM. Fig. 10.2 shows the scattered plot of all computed and actual heart rate readings. It shows that the computed heart rate readings are closely correlated to the actual heart rate readings. The correlation coefficient between the computed and actual heart rate readings is 0.97. The Bland Altman plot is shown in Fig. 10.3. It shows that only a small number of computed heart rate readings are located outside the 95 % limit of agreement interval. Table 10.1. Summary of heart rate readings results obtained from ICA approach. Subject 1 2 3 4 Heart Rate Readings (BPM) Highest Lowest 150 127 153 133 141 134 153 142 Mean absolute error (BPM) 1.29 1.37 1.38 1.48 Standard deviation of absolute errors (BPM) 1.06 1.09 1.19 0.81 Fig. 10.2. Comparison of all actual and estimated heart rate readings. 235 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 10.3. Bland-Altman plot for all estimated heart rate readings. 10.3.3. Experimental Results Using PCA A total of 80 instantaneous heart rate readings were obtained from this experiment. In the experiment, the subjects’ heart rates were varying between 129 BPM and 153 BPM. Fig. 10.4 shows the detail workflow of the PCA method. The face region was identified by using the model described in (Viola and Jones, 2001 [7]) and the region of interest (ROI) was fixed at the area below eyes and above the upper lip of mouth. For each frame, the spatially average of the RGB and YCbCr components, i.e.: μR, μG, μB, μY, μCb, and μCr are computed respectively. All six color components were projected into log-space. Therefore, at any time instant, a set of six input features log PR, log PG, log PB, log PY, log PCb and log PCr were formed. The set of input features were then detrended using the model developed by Tarvainen et al. (Tarvainen et al., 2002 [10]). PCA was then used to recover six PCs from these six input features. The set of PCs was bandpass filtered (128-point Hamming window, 0.8-4 Hz). The entire process was repeated by increasing the number of previous video frames, until the stopping criterion was met. At this point, the corresponding number of frames was chosen as the video duration needed to compute the instantaneous heart rate reading. The first PC was then chosen as the PPG signal. The corresponding frequency of this PPG signal was considered as the instantaneous heart rate reading for that particular instant. Table 10.2 summarizes the details of the computed heart rate readings of all subjects. The highest mean absolute error and the highest standard deviation of absolute errors are 1.94 and 1.21 BPM. Fig. 10.5 shows the scattered plot of all computed and actual heart rate readings. It shows that the computed heart rate readings are closely correlated to the actual heart rate readings. The correlation coefficient between the computed and actual heart rate readings is 0.93. The Bland Altman plot is shown in Fig. 10.6. It shows that only a small number of computed heart rate readings are located outside the 95 % limit of agreement interval. 236 Chapter 10. Video Based Heart Rate Estimation Using Facial Images from Video Sequences Fig. 10.4. Workflow of PCA approach (Yu et al, 2015 [27]). Table 10.2. Summary of heart rate readings results obtained from PCA approach. Subject Heart Rate Readings (BPM) Highest Lowest Mean absolute error (BPM) Standard deviation of absolute errors (BPM) 1 141 134 1.94 1.21 2 153 133 1.18 1.09 3 153 135 1.93 1.00 4 135 129 1.18 1.09 237 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 10.5. Comparison of all actual and estimated heart rate readings. Fig. 10.6. Bland-Altman plot for all estimated heart rate readings. 10.4. Distance between the Subject and Video Camera By increasing the distance between the subject and video camera, the high quality resolution is lost thereby affecting the accuracy of the heart rate estimation. In this study, two experiments were carried out to examine the impact of varying the distance between the subject and the video camera and to its heart rate estimation. In the first experiment, the distance between the subject and video camera was varied with fixed duration. The distance was varied between 30 cm and 200 cm with fixed durations of 5 and 8 seconds. In the second experiment, the duration was varied between 3 and 9 seconds with fixed distance of 50 cm and 100 cm. Four subjects took part in the experiments. The obtained data from both experiments were analyzed using an integrated method consisting of detrend, ICA, bandpass filters, and fast Fourier transform as shown in Fig. 10.7. 238 Chapter 10. Video Based Heart Rate Estimation Using Facial Images from Video Sequences Fig. 10.7. Flow chart of video-based heart rate estimation using ICA. 10.4.1. Varying Distance with Fixed Video Duration All experiments followed the setup described in Section 10.3.1. The subject was seated on a chair facing the video camera with an initial distance of 30 cm and was recorded for 60 seconds while his/her instantaneous heart rate reading was recorded simultaneously. The subject was requested to stay still with no motion. Viola et al model used to detect the face region (Viola et. al, 2001 [7]). The obtained data were processed and analyzed offline using MATLAB R2013a. The experiments were repeated by increasing the distance 10 cm for each case till the final distance reaches 200 cm. A total of 18 sets of videos and heart rate readings were captured for each subject. Four subjects’ heart rates were measured and they varied from 62 BPM to 90 BPM. The method as shown in Fig. 10.7 was used to estimate the heart rate readings from videos. A 5-second video clip was used in each case. The obtained results were compared to actual heart rate readings for each subject. This study also looks into the accuracy of the heart rate estimation when the duration is increased from 5-seconds to 8-seconds. Tables 10.3 and 10.4 show the Root Mean Square Error (RMSE) for all four subjects with distance varying from 30 cm to 200 cm for both fixed durations of 5-seconds and 8-seconds respectively. The error rates are much lower for 8-seconds duration when compared to 5-seconds duration. The results for both 5-seconds and 8-seconds duration gave the lowest scores when the distance was 50 cm. More details of the average error rates obtained from the differences between the estimated heart rates and actual heart rates for four subjects are shown using Box-Plot Graph in Fig. 10.8 and Fig. 10.9. Each box in the Figs. 10.8 and 10.9 indicate the average of the errors of the four subjects combined. More error rates are observed when the duration is for 5-seconds as shown in Fig. 10.8. 239 Advances in Optics: Reviews. Book Series, Vol. 3 Table 10.3. RMSE values for distance varying from 30 cm to 200 cm with 5-seconds duration. Distance (cm) 30 40 50 60 70 80 90 100 110 120 130 140 150 160 170 180 190 200 RMSE with 5-seconds duration S1 S2 S3 S4 1.056 9.331 3.090 0.841 2.282 2.388 7.002 5.187 1.238 2.075 2.757 0.850 1.609 4.047 4.610 1.785 2.183 12.597 2.680 1.585 1.724 2.750 2.493 2.219 2.677 5.006 7.723 1.887 2.473 5.473 7.231 2.664 2.408 4.929 31.580 1.648 1.762 3.897 8.979 5.590 4.378 4.714 8.968 1.492 4.594 4.270 25.483 2.493 18.463 9.818 9.672 3.109 16.995 4.787 34.688 1.536 4.431 3.453 32.981 2.081 2.003 39.025 30.779 22.533 2.906 7.923 7.501 5.807 36.442 30.142 61.079 7.240 Table 10.4. RMSE values for distance varying from 30 cm to 200 cm with 8-seconds duration. Distance (cm) 30 40 50 60 70 80 90 100 110 120 130 140 150 160 170 180 190 200 240 RMSE with 8-seconds duration S1 S2 S3 S4 1.048 3.202 2.859 0.896 1.304 1.792 6.252 0.768 1.368 1.300 2.675 0.646 0.984 2.448 2.917 0.836 0.746 3.635 2.644 0.566 1.381 2.560 1.341 1.304 1.476 1.230 5.505 1.285 1.843 4.942 1.913 1.808 1.650 3.448 5.958 0.722 1.105 2.450 0.920 4.462 1.663 6.173 3.078 1.860 7.831 3.299 25.008 1.400 21.131 1.475 8.664 1.565 1.985 2.684 15.970 0.779 0.969 2.514 4.136 1.391 1.485 0.858 10.929 1.923 1.739 9.385 8.417 6.115 22.674 29.222 44.849 11.181 Chapter 10. Video Based Heart Rate Estimation Using Facial Images from Video Sequences Fig. 10.8. Box-Plot Graph of error values for all subjects with duration time of 5 seconds. Fig. 10.9. Box-Plot Graph of error values for all subjects with duration time of 8 seconds. 10.4.2. Fixed Distance with Varying Video Duration The same four subjects from previous experiments were involved in this experiment. The same set-up as in the previous experiments was used except in this case the distance is fixed at 50 cm and 100 cm. The video frames used to analyze the data were varied from 3-seconds to 9-seconds. The same integrated ICA method shown in Fig. 10.7 was used to obtain the estimated heart rate readings. Tables 10.5 and 10.6 show the RMSE values for the four subjects with varying video duration for a fixed distance at 50 cm and 100 cm 241 Advances in Optics: Reviews. Book Series, Vol. 3 respectively. The results in Table 10.5 shows that beyond 5-second duration gave acceptable errors between the estimated and actual readings. Similar observations can be seen in Table 10.6 where in this case, the video duration beyond 6-second produces acceptable error rates. More details of the errors are shown using Box-Plot graph as shown in Fig. 10.10 and 10.11 for the distance 50 cm and 100 cm respectively. Table 10.5. RMSE values for varying duration time at distance of 50 cm. Time Duration (s) 3 4 5 6 7 8 9 RMSE at distance of 50 cm S1 S2 S3 S4 2.448 3.570 4.683 8.424 1.642 2.081 4.724 8.198 1.238 2.075 2.757 0.850 1.406 1.297 2.720 0.828 1.473 1.231 2.183 0.653 1.368 1.300 2.675 0.646 1.473 1.300 2.543 0.588 Table 10.6. RMSE values for varying duration time at distance of 100 cm. Time Duration (s) 3 4 5 6 7 8 9 RMSE at distance of 100 cm S1 S2 S3 S4 5.168 14.738 7.830 3.907 3.220 5.919 13.076 3.250 2.473 5.473 7.231 2.664 2.166 5.376 2.465 2.418 1.297 4.585 3.447 1.151 1.843 4.942 1.913 1.808 0.936 4.395 2.522 0.720 10.5. Conclusion This chapter presents the methods to estimate dynamic heart rates using both ICA and PCA approaches. A method using ICA combined with mutual information is introduced to identify the minimum video duration needed to estimate the heart rate without compromising the accuracy of heart rate readings. Four subjects took part in an experiment involving cycling activity. The obtained results were compared to the actual heart rate readings from the Polar Team2 Pro and acceptable errors were observed. In the next experiment, PCA is used to de-correlate the color components of the PPG signals to estimate the dynamic heart rates. Similar error rates were observed between the actual and the estimated heart rate. In addition to that, we also examine the impact of varying the distance between the subject and the video camera and also fixing the distance but varying the video duration. In the experiment where varying the distance up to 130 cm with fixed 242 Chapter 10. Video Based Heart Rate Estimation Using Facial Images from Video Sequences duration showed that acceptable error rates between the actual and computed method are observed. In the last experiment where the video duration is varied with fixed distance of 50 cm and 100 cm showed that beyond 6-second video duration gave acceptable error rates for both fixed distances. Fig. 10.10. Box-Plot Graph of error values for all subjects with distance of 50 cm. Fig. 10.11. Box-Plot Graph of error values for all subjects with distance of 100 cm. 243 Advances in Optics: Reviews. Book Series, Vol. 3 References [1]. A. B. Hertzman, The blood supply of various skin areas as estimated by the photoelectric plethysmograph, American Journal of Physiology, Vol. 124, Issue 2, 1938, pp. 328-340. [2]. J. Weinman, A. Hayat, G. Raviv, Reflection photoplethysmography of arterial-blood-volume pulses, Medical and Biological Engineering and Computing, Vol. 15, Issue 1, 1977, pp. 22-31. [3]. J. F. Cardoso, A. Souloumiac, Blind beamforming for non-Gaussian signals, IEE Proceedings F - Radar and Signal Processing, Vol. 140, No. 6, 1993, pp. 362-370. [4]. J. F. Cardoso, High-order contrasts for independent component analysis, Neural Computation, Vol. 11, Issue 1, 1999, pp. 157-192. [5]. A. Hyvärinen, E. Oja, Independent component analysis: algorithms and applications, Neural Networks, Vol. 13, Issue 4, 2000, pp. 411-430. [6]. T. W. Lee, M. S. Lewicki, T. J. Sejnowski, ICA mixture models for unsupervised classification of non-Gaussian classes and automatic context switching in blind signal separation, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 22, Issue 10, 2000, pp. 1078-1089. [7]. P. Viola, M. Jones, Rapid object detection using a boosted cascade of simple features, in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’01), Kauai, Hawaii, USA, 8-14 Dec. 2001, I511. [8]. S. John, A Tutorial on Principal Components Analysis, Princeton University, 2002, pp. 1-16. [9]. I. S. Lindsay, A Tutorial on Principal Components Analysis, University of Otago, 2002, pp. 1-27. [10]. M. P. Tarvainen, P. O. Ranta-Aho, P. A. Karjalainen, An advanced detrending method with application to HRV analysis, IEEE Transactions on Biomedical Engineering, Vol. 49, Issue 2, 2002, pp. 172-175. [11]. M. Garbey, N. Sun, A. Merla, I. Pavlidis, Contact-free measurement of cardiac pulse based on the analysis of thermal imagery, IEEE Transactions on Biomedical Engineering, Vol. 54, Issue 8, 2007, pp. 1418-1426. [12]. I. Pavlidis, J. Dowdall, N. Sun, C. Puri, J. Fei, M. Garbey, Interacting with human physiology, Computer Vision and Image Understanding, Vol. 108, Issue 1-2, 2007, pp. 150-170. [13]. C. Takano, Y. Ohta, Heart rate measurement based on a time-lapse image, Medical Engineering & Physics, Vol. 29, Issue 8, 2007, pp. 853-857. [14]. W. Verkruysse, L. O. Svaasand, J. S. Nelson, Remote plethysmographic imaging using ambient light, Optics Express, Vol. 16, Issue 26, 2008, pp. 21434-21445. [15]. R. G. Gatto, Estimation of instantaneous heart rate using video infrared thermography and ARMA models, PhD Thesis, University of Illinois, Chicago, 2009. [16]. E. Jonathan, M. Leahy, Investigating a smartphone imaging unit for photoplethysmography, Physiological Measurement, Vol. 31, Issue 11, 2010, pp. N79-N83. [17]. M. Z. Poh, D. J. McDuff, R. W. Picard, Non-contact, automated cardiac pulse measurements using video imaging and blind source separation, Optics Express, Vol. 18, Issue 10, 2010, pp. 10762-10774. [18]. P. Shi, V. A. Peris, A. Echiadis, J. Zheng, Y. S. Zhu, P. Y. S. Cheang, S. J. Hu, Non-contact reflection photoplethysmography towards effective human physiological monitoring, Journal of Medical and Biological Engineering, Vol. 30, Issue 3, 2010, pp. 161-167. [19]. A. A. Kamshilin, S. Miridonov, V. Teplov, R. Saarenheimo, E. Nippolainen, Photoplethysmographic imaging of high spatial resolution, Biomedical Optics Express, Vol. 2, Issue 4, 2011, pp. 996-1006. [20]. M. Z. Poh, D. J. McDuff, R. W. Picard, Advancements in noncontact, multiparameter physiological measurements using a webcam, IEEE Transactions on Biomedical Engineering, Vol. 58, Issue 1, 2011, pp. 7-11. 244 Chapter 10. Video Based Heart Rate Estimation Using Facial Images from Video Sequences [21]. M. Schönfelder, G. Hinterseher, P. Peter, P. Spitzenpfeil, Scientific comparison of different online heart rate monitoring systems, International Journal of Telemedicine and Applications, Vol. 2011, 2011, 631848. [22]. T. Pursche, J. Krajewski, R. Moeller, Video-based heart rate measurement from human faces, in Proceedings of the IEEE International Conference on Consumer Electronics (ICCE’12), Las Vegas, NV, USA, 2012, pp. 544-545. [23]. M. B. Wallen, D. Hasson, T. Theorell, B. Canlon, W. Osika, Possibilities and limitations of the polar RS800 in measuring heart rate variability at rest, European Journal of Applied Physiology, Vol. 112, Issue 3, 2012, pp. 1153-1165. [24]. S. C. Xu, L. Y. Sun, G. K. Rohde, Robust efficient estimation of heart rate pulse from video, Biomedical Optics Express, Vol. 5, Issue 4, 2014, pp. 1124-1135. [25]. M. Kumar, A. Veeraraghavan, A. Sabharwal, DistancePPG: Robust non-contact vital signs monitoring using a camera, Biomedical Optics Express, Vol. 6, Issue 5, 2015, pp. 1565-1588. [26]. Y. P. Yu, P. Raveendran, C. L. Lim, Dynamic heart rate measurements from video sequences, Biomedical Optics Express, Vol. 6, Issue 7, 2015, pp. 2466-2480. [27]. Y. P. Yu, P. Raveendran, C. L. Lim, B. H. Kwan, Dynamic heart rate estimation using principal component analysis, Biomedical Optics Express, Vol. 6, Issue 11, 2015, pp. 4610-4618. 245 Chapter 11. Implementing Differential Signalling in Free Space Optical Communication Link Chapter 11 Implementing Differential Signalling in Free Space Optical Communication Link Mojtaba Mansour Abadi1 11.1. Introduction to Free Space Optics One of the current challenges in wireless communications is to be able to provide a cost effective high speed data link in applications, where the radio frequency (RF) based technology cannot be used or is not suitable. For example, in highly populated indoor environments (train station, airports, etc.), and ‘the last mile access’ network, where the end users, using the RF based wireless technologies, do experience lower data rates and low quality services due to the spectrum congestion (i.e., bandwidth bottleneck). The high speed optical wireless connection is defined as a data link with a minimum speed of few Gbps, which is also known as gigabit data connection [1], in emergency situations such as flooding, earthquake, etc., and massive public events including concerts, festivals, as well as optical fibre networks maintenance and repair. Nowadays, using the internet and, in general, having access to the data network have become a typical daily task for everyone. With the rapid growth of smart devices, the RF spectrum, which is already being stretched too thinly, is experiencing congestion at a global level, which needs addressing. Nowadays, there are a growing number of applications as shown in Fig. 11.1, which require access quality to the data services anywhere, anytime and under all conditions. In a perfect scenario, all end users should have access to the optical fibre based backbone network with an ultra-high capacity, to benefit from truly high-speed data communications with a very low end-to-end transmission latency. Of course, for environment where deployment of optical fibre is not economical a combination of satellite communications and optical fibre communications technologies would be the most suitable option. However, this could also be quite costly and therefore may not be feasible in the long run. Mojtaba Mansour Abadi School of Engineering, University of Glasgow, Glasgow, Scotland 247 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 11.1. Variety of demands for high bit rate access to the data. Therefore, because of the limited bandwidth, and high cost of the RF technology [2], there is the need to consider alternative technologies. The cost and challenges associated with installation of optical fibre particularly in rural areas as well as maintenance of such a network is rather high, therefore is not considered for the last mile access network. Whereas bandwidth limited RF technologies are also not suitable, thus the need for the most attractive and relatively cost effective solution still exists. Free space optical (FSO) communications also known as optical wireless communications have been used to provide high speed data service for aforementioned applications [3]. The FSO technology is license free, easily deployable, secure and capable of offering low bit error rate (BER) as well as high speed link over a range of link span up to 10 km for civilian applications [4], which has been adopted in a number of applications including:  Broadband internet to rural areas [5] – FSO based link could replace optical fibre access technologies such as fibre to the home (FTTH) in order to provide connectivity between in-building networks and to broadband and backbone data networks;  Inter-building connectivity [4] and electronic commerce [6] – FSO provides highspeed, flexibility and high security;  Audio and video streaming [7] – FSO is an attractive solution for video surveillance and monitoring, as well as live broadcasting of sporting events, in emergency situation (e.g., the tsunami in Japan in 2011 that almost wiped out all the telecommunications infrastructure [11]) etc. Unmanned aerial vehicle (UAV) and high attitude platforms [4, 8] – UAVs and high attitude platforms have been used for military surveillance, monitoring traffic and disaster 248 Chapter 11. Implementing Differential Signalling in Free Space Optical Communication Link areas, or broadcasting vital data. UAVs generate a large volume of data, which needs downloading as quickly as possible. This can be achieved by employing the on-board FSO technology. 11.2. FSO Communications 11.2.1. Background Considering any information transmission via light as a form of optical communication; then the ancient Greeks and Romans around 800 B.C. who used fire beacons for signalling over a medium range distance were the first users of FSO links [9]. Since then, a growing interest in research, development and deployment in FSO is observed. The first modern system was developed by Alexander Graham Bell back in 1880 by inventing the “Photophone” that used sun rays to transfer voice over a distance of 200 m [10]. However, not much happened until the discovery of the laser in 1960s. In 1962 researchers from MIT Lincolns Laboratory demonstrated a television signal transmission over 48 km using a light emitting GaAs diode based FSO link [9]. For years FSO was used in military and deep space applications with very little commercial use. The reasons were that i) existing communication technologies were more than adequate to meet the demand; ii) lack of cost-efficient reliable optical components; and iii) impact of the atmospheric conditions on the performance of FSO links [10]. As mentioned before, the demand for higher data rate was the main motivation for researchers to reconsider FSO as a promising alternative and complementary technology to the RF. The growing number of research and development activities in FSO both within the academy and industry supported by a large number of scientific articles clearly demonstrate the potential of this emerging technology with optical fibre like capabilities. Currently, there are commercial FSO products available in the market offering data rates beyond gigabits per second [11, 12]. Research and development is going on to push the data rate to higher limits (e.g., 10 Gbps commercial FSO transceiver [13]) and improve the link quality (e.g., it is desirable to achieve ideal 100 % link availability in all weather conditions. In practice, using a hybrid link the availability of five nines (99.999 %) is reported [5]) as well as to reduce the cost of the complete system. The cost effectiveness of FSO system compared to the RF system is more obvious, when the RF system is supposed to deliver the same high data rate connection service [14-16]. To illustrate the comparison between existing technologies and FSO and to show the advantage of FSO one can refer to Fig. 11.2. [17, 18] In summary the key features of FSO systems for long range line-of-sight (LOS) high speed data connections are outlined as follow:  High data rate: At the moment RF provides 1 to 2 Mbps for unregulated 2.4 GHz ISM bands [19], 20 Mbps 875 Mbps at 5.7 GHz 4G mobile and 60 GHz millimetre wave technologies, respectively [20]. Potentially FSO can provide bandwidth as large 249 Advances in Optics: Reviews. Book Series, Vol. 3 as 2000 THz, which is far beyond the maximum data rate of RF technologies [21, 22], see Fig. 11.2 (a). (a) 180,000 160,000 Fibre RF FSO 140,000 Cost [$] 120,000 100,000 80,000 60,000 40,000 20,000 0 20 40 60 80 100 Data Rate [Mbps] 120 140 160 (b) Fig. 11.2. Comparison of communication technologies: (a) in terms of data rate and link coverage, figure taken from [17] with permission, and (b) in terms of cost and data rate (the figure is plotted using the data from [18]).  License free spectrum: FSO spectrum is not regulated therefore there are no license fees.  Power consumption: The global information and communications technology is responsible for 2 to 10 % of the global energy consumption according to the report 250 Chapter 11. Implementing Differential Signalling in Free Space Optical Communication Link smart 2020 [23]. The global warming and the existing concern to reduce the power consumption is a critical motivation to replace RF with FSO in LOS applications since FSO is potentially green in terms of energy consumption compared to RF.  Low cost: Covering installation, maintenance and license fees, see Fig. 11.2 (b).  High security: FSO links are inherently secure due to highly narrow, well confined and directional beam profile.  Back-bone network compatibility: FSO operating at all three optical transmission windows of 850, 1300 and 1550 nm are compatible with optical fibre based back bone networks. 11.2.2. FSO Structure Fig. 11.3 illustrates the general schematic system block diagram of a typical FSO communication link (Note that in the following chapters, the detailed block diagram will be presented for each case). Fig. 11.3. The fundamental system diagram of an FSO link. LD and PD are laser diode and photodetector, respectively. The modulated or unmodulated version of the transmit information, (i.e., a digital bit steam) is used for intensity modulation of the optical source. Note that, for much higher data rates (i.e., in excess of 10 Gbps) external modulation schemes should be used). As shown in Fig. 11.3, the modulation procedures are assumed to be performed in the transmitter (Tx) module. The optical source adopted could be a light emitting diode (LED) or a laser diode (LD). The latter is more widely used because of the LOS transmission requirement, longer transmission span and higher data rates in outdoor environments. Note that additional optics such as lens, beam splitters, beam polarizer, optical filter, optical fibre, etc. are also used as part of the Tx. Particularly, locating the LD at the focal length of a lens is a common practice used in beam forming and collimation of the laser output. The generated optical beam is launched into the free space channel and is captured at the receiver (Rx) using a combination of optics and an optical photodetector (PD). As in the Tx, typically a lens is used at the Rx to focus the received beam into the PD. The generated electrical signal at the output of the 251 Advances in Optics: Reviews. Book Series, Vol. 3 PD is then amplified, processed and converted back into the digital bit stream at the Rx, see Fig. 11.3. Depending on the application, the free space channel condition and the data rate, the aforementioned elements (i.e., LD, PD, and interface block) can be different. For instance, in a short range clear channel, a single Tx and Rx as well as a simple on-off-keying nonreturn-to-zero (NRZ-OOK) modulation scheme would be sufficient to meet the link requirements [17]. For most cost-effective typical systems, intensity-modulation/direct-detection (IM/DD) based FSO links are more adequate [7, 24-27]. In IM/DD technique, based on the information data, only the intensity of the light is modulated. Another technique to modulate and detect the light is the external modulation and the coherent detection, where in contrary to IM/DD, both intensity and phase/frequency of the light can be modulated [28]. Coherent systems require an external modulator such as Mach-Zehnder modulator to perform the modulation operation. Besides at the Rx, a light source synchronized with the one at the Tx is needed to down-convert the received optical beam [29]. Although coherent systems are shown to have impressive performance in terms of background noise rejection, atmospheric-induced fading mitigation and higher sensitivity of the Rx; the cost and complexity of practical implementation, in particular stability and synchronization of laser sources at the Tx and the Rx, make them less popular than IM/DD based systems [7]. Although the FSO technology has many benefits, it cannot provide 100 % link availability under all weather conditions as outlined below [8]: 1. Turbulence induced fading - This is due to the temperature gradient along the optical propagation path and the movement of air perpendicular to the propagating optical beam [30]. In Chapter 3, more detail is given about turbulence phenomena. 2. Atmospheric loss – This is mainly due to the fog, aerosol, haze, smoke particles [7], where the induced loss by fog/smoke is dominant [31]. The attenuation is the contribution of molecular absorption and light scattering [32]. Absorption occurs at the molecular level and is resulted from absorption of photon energy by molecules of gases in the atmosphere [33]. Since the dimension of fog particles varies between 0.5 m to 2 m, which is in the range of FSO wavelengths, therefore Mie scattering is the major cause of scattering attenuation [32, 33]. 3. Pointing errors induced fading - Is due to the vibration or small movements of both Tx and/or Rx [34]. In Chapter 4, more detail is given about this phenomena. 4. Link blockage - Mostly due to flying object or birds, which results in a burst error [17]. 5. Ambient noise – This can be considered as the main noise source in many scenarios [35]. The ambient noise is mostly due to sunlight or artificial lighting sources [35]. 252 Chapter 11. Implementing Differential Signalling in Free Space Optical Communication Link In particular, among these effects, the atmospheric loss can lead to attenuation as high as 50 dB∕km for <500 m visibility under the fog condition [40, 41]. As reported in [3] the fog attenuation in moderate continental fog environment (Graz, Austria) in winter season and in dense maritime fog environment (La Turbie, France) in summer months can lead to 120 dB/km and 480 dB/km, respectively [42]. Indeed, the FSO link undergoes a wide range of attenuation in presence of fog and smoke [3], which results in reduced link availability in a significant way so that it was shown in [2] that for the weather condition recorded in Graz, the link availability could drop to ∼67 %. The general FSO system block diagram with more detailed components is illustrated in Fig. 11.4. Note that the FSO link is based on IM/DD technique. The digital input bit stream or the information data is applied to the modulation block. Depending on the application, the modulation could be a simple OOK scheme or more complex multilevel amplitude, frequency and phase scheme [7]. y x S  0,1 Input Bit Stream Modulator w Driver LD Transmitter TIA Free Space Channel Circuit Receiver Aperture Transmitter Aperture Channel z Demodulator O  0,1 Output Bit Stream PD Receiver Fig. 11.4. The block diagram of an IM/DD FSO communication link. LD, PD, TIA are laser diode, photodetector, and transimpedance amplifier, respectively. In this work, the NRZ-OOK data format is adopted, which is the most widely used in commercial FSO systems [36]. NRZ-OOK is used because of its simplicity and a balanced power and spectral efficiency compared to other digital modulation schemes [37]. Note that the research main focus is on the design of the hybrid antenna and less on the modulation schemes. However, the system employing a hybrid antenna could readily be adopted for other modulation schemes. Then the output of the modulator (i.e., a bipolar NRZ-OOK signal in this case) with a bandwidth BW 1⁄ , where is the bit duration is DC-level shifted to convert it into a unipolar format prior to intensity-modulate the light source (i.e., LD in this case). The data rate of NRZ-OOK equals to BW [38]. , divergence angle The output of the LD (i.e., ) has four key parameters, wavelength , beam waist , and output power . The wavelength used in FSO is in the red to infrared range of the spectrum. In this work only two wavelengths of 670 nm and 830 nm have been adopted in the experimental investigation. The visible LD at 670 nm is also used for the alignment of the FSO link. The laser beam is collimated using a lens in order to reduce the geometrical loss [39]. This means that the divergence angle should be kept small, and will be the minimum radius of the propagating laser beam, which can be considered prior to the beam divergence [40]. However, the output power of the LD is subject to the eye and safety regulations. There are different standards for laser safety such as the international 253 Advances in Optics: Reviews. Book Series, Vol. 3 electrotechnical commission (IEC) [41], American national standards institute (ANSI) [42] and European committee of electrotechnical standardization (CENELEC) [43]. In this work one class 3R pointer laser with wavelength of 670 nm and 2 mW output power is used, which is considered to be safe if handled carefully. Also another class 3B pointer laser with wavelength of 830 nm and 10 mW output power which needs eyes protection is used. Note that class 3B lasers are only used for experimental proof of concept in laboratory and in an outdoor real scenario, safe lasers must be implemented [44]. In FSO systems with LOS configuration the link performance will be affected by blocking mostly due to flying objects (e.g., birds) and the atmospheric channel conditions (i.e., fog, smoke, turbulence, etc.). The channel affects are defined as the attenuation (loss) and a random fading. The attenuation is due to: 1. Geometrical loss - The real laser light is not an ideal collimated beam. With even a small divergence angle of , it experiences beam spreading and since the PD has a finite physical size, only a fraction of the laser power is captured and collected at the Rx. Therefore, the geometrical loss is defined by [45]: , 20 log √2 (11.1) where , , and denote the bear radius at transmitter, the link distance and the receiver aperture diameter. 2. Atmospheric attenuation - This loss is due to Rayleigh scattering and molecule absorption, with a typical value of 0.5 dB [39]. 3. Fog attenuation - This is the most important loss in FSO links. Generally speaking, fog and smoke, which are composed of small particles floating in the air, are the main cause of attenuation in FSO systems [32]. The optical beam interacting with fog and smoke particles results in both absorption and Mie scattering, which [46]. The fog attenuation is determined based on the channel's contribute to visibility (Vis) in km and is described by two well-known models of Kim and Kruse [45]. The visibility is define as [32]: Vis , (11.2) where is LD wavelength, and denote the maximum sensitive wavelength for human eye, which is normally set to 550 nm (i.e., the green colour). Based on Kim model is defined as [46]: q 254 1.6 1.6, Vis 50 1.3, 6 Vis 50 Vis 0.34, 1 Vis 6 . Vis 0.5, 0.5 Vis 0.1 0, Vis 0.1 (11.4) Chapter 11. Implementing Differential Signalling in Free Space Optical Communication Link The relation between the total attenuation due to the absorption and scattering of light is given by Beer-Lambert law as [46]: and . . (11.4) Depending on visibility ( Vis ), fog can be defined as thick ( Vis (0.1 Vis 1) or thin (Vis 1). 0.1 ), medium - This includes additional losses due to 4. Miscellaneous attenuation misalignment, devices or other unknown factors [39]. Rain also introduces a loss in FSO systems, which is not significant compared to fog, and smoke [2]. However, rain is a major source of attenuation in RF systems. The received power in terms of the transmit power and all losses is given by [47]: . (11.5) For a system point of view, knowing that the total noise variance at the Rx and the PD's responsivity are and , respectively, the link electrical signal-to-noise ratio (SNR) is defined as [47]: SNR . (11.6) In a clear channel with no fading the system BER is given by [47]: BER √SNR , where ∙ denotes the Gaussian -function defined as Thus in clear channel (i.e., no channel fading), to achieve BER SNR > +13.54 dB. (11.7) exp 10 ⁄2 . one needs In this research work, a channel with fading is considered as outlined below: 1. Scintillation/turbulence - In a clear channel with no turbulence the propagating optical beam only experiences attenuation. Whereas in turbulence channel the propagating beam will experience both attenuation and phase variation due to randomly varying refractive index of the air, thus leading to fading and link failure [7, 48]. Note that, turbulence is caused by the presence of temperature gradient along the laser propagation path and the air movement (wind) perpendicular to the laser beam [17]. Turbulence randomly changes the refractive index along the propagation path, which consequently causes the beam wandering [30]. Depending on the fading intensity, the turbulence can be classified as the weak, moderate, strong and saturated [17]. 255 Advances in Optics: Reviews. Book Series, Vol. 3 2. Pointing errors - This is due to the movement of building, mast, tower, and in general the structures on which the FSO units are mounted [34]. Pointing errors lead to the amplitude fluctuation (or oscillation) of the received optical signal in the transverse plan, thus contributing to deterioration of the link's performance [49]. More details on these fading effects will be given in the next sections. Also mitigation methods to overcome them will be introduced. As shown in Fig. 11.4 in the IM/DD system the received optical signal is captured using a lens and focused onto a PD the output of which is amplified using a transimpedance amplifier (TIA) prior to be demodulated in order to recover the transmitted information. 11.3. Turbulence If hypothetically the turbulent channel is frozen at 0, the channel can be considered as in Fig. 11.5. The input wavefront in this case is planar, which encounters random changes in the refractive index. The refractive index variation is a function of the atmospheric temperature, pressure, altitude and wind speed. The variation is modelled as small cells with different refractive index from the adjacent cells. These cells are called turbulence eddies. The size of eddies might change from a few millimetre to several metres [30, 50]. Turbulence is a slow varying fading channel and has the temporal coherence in the order of 1 to 10 ms [51]. Fig. 11.5. Turbulent channel frozen at t = 0. The turbulent channel consists of eddies with various sizes. A useful parameter is the channel correlation radius where correlation length 256 , which is defined as [52]: , (11.8) . (11.9) is given by [53]: Chapter 11. Implementing Differential Signalling in Free Space Optical Communication Link Also is spatial coherence radius and is defined as [52]: ⁄ 1.46 , (11.10) where 2 ⁄ is the wavenumber and is the index-of-refraction structure is an important parameter to choose the spacing between various parameter. independent Rxs in a system with multiple Rxs. The term ‘independent Rxs’ means that the correlation between the fading effects of two received signals from two independent Rxs is negligible. is a general use parameter and can be employed in different turbulence regimes, whereas correlation length ( ) is an approximation of for the weak turbulence regime [52]. For the index-of-refraction structure ( ) there are a number of models available but the most widely used is the altitude dependent model developed by Hufnagle-Valley, which is given by [54]: 2.7 0.00594 ⁄27 10 exp exp ⁄1000 10 ⁄1500 exp ⁄1000 , (11.11) where , , and represent the altitude (m), the root mean square (rms) wind speed (m/s), at 0 , respectively. Depending on the strength of and the nominal value of might be 1.7 10 m ⁄ during daytime for a 1 km link or turbulence 8.4 10 m ⁄ during night for the same link [54]. Provided that the received optical signal intensity is denoted by , a useful parameter to qualify the effect of turbulence is the scintillation index (SI) , which is defined as the normalized irradiance variance of the optical beam as given by [53]: , (11.12) where ∙ denotes the expected value. The variance of log-intensity signal fluctuation is given by [53]: defined by Rytov variance . 1.23 For the light beam with a spot size of diameter the weak turbulence condition are given by [30]: 1 and Λ (11.13) at the Rx, Rytov variance criteria for 1, (11.14) . Since in this research work Λ ≪ 1 , then only the condition where Λ 2 ⁄ 1 is applicable for weak turbulence. 257 Advances in Optics: Reviews. Book Series, Vol. 3 The distribution of fading coefficient for the weak turbulence regime can be modelled using Log-normal distribution, which is practically valid as long as 0.3. Log-normal probability distribution function (PDF) of the normalized irradiance with mean and is given as [15]: variance where ⁄ exp , (11.15) is the signal light intensity without turbulence. To normalize it is assumed that turbulence and are related as follow [30]: exp 4 [54]. Under the assumption of weak 1≅4 , (11.16) In the literature depending on the light propagation model different expressions are introduced for the variance of Log-normal distribution. For the plane wave propagation, [30]. one has Aperture averaging is a technique that reduces the variation of optical intensity according to the aperture diameter [55, 56]. To benefit from aperture averaging, the size of needs to be larger than in Eq. (11.9) [52, 54], which is valid for the weak turbulence regime. In fact, can be much larger in the moderate-to-strong turbulence regime [52, 56]. With defined as the scintillation index of the Rx with no 0 and aperture and with aperture of , respectively, the aperture averaging factor (AF) is given by [7, 56]: AF 1 ⁄ 1.6682 . (11.17) In the moderate-to-strong turbulence regime one has [30]: 1, for moderate regime, (11.18a) 1, for strong regime, (11.18b) and the received signal optical intensity is based the PDF of Gamma-Gamma (GG) distribution given by [57]: 2 , (11.19) where ≥ 0 and ≥ 0 are known as the effective numbers of large- and small-scale turbulence cells, respectively [58, 59]. Kn (∙), and Γ ∙ denote the modified Bessel function of 2nd kind and order n, and the Gamma function, respectively. 258 Chapter 11. Implementing Differential Signalling in Free Space Optical Communication Link The two parameters of and  that characterize the irradiance fluctuation PDF are related to the atmospheric conditions and , which are given by [30]: α , (11.20a) , (11.20b) 1, exp αβ (11.20c) where σ2lnX and σ2lnY represent the variances of large-scale and small-scale irradiance fluctuations, respectively. As mentioned above the presence of aperture at the Rx will reduce the effect of turbulence. For the plane wave propagation model and considering the aperture size the closed form expressions for σ2lnX and σ2lnY parameters are given by [30]: . . where ⁄4 . [30]. . . . . . ⁄ ⁄ ⁄ ⁄ ⁄ , (11.21a) , (11.21b) 11.4. Channel Model For a single-input single-output (SISO) FSO link with OOK employing a receive can be written as (see Fig. 11.6) [54]: apertures, the received signal at the aperture , (11.22) where ∈ 0,1 represents the information bit, is the optical-to-electrical conversion coefficient, denotes the irradiance received at the aperture, and is additive white . The AWGN can be Gaussian noise (AWGN) with zero mean and variance of considered as a combination of the thermal, shot, and dark noise sources of the Rx and the background ambient light [60]. The subscript m is denoting the fact that in general case an FSO system can consist of several SISO link with the given scheme in Fig. 11.6. It is assumed that the channel for Tx to each Rx is independent, which interprets to the fact that the transversal distance between the adjacent apertures is larger than in Eq. (11.8) [52]. In single-input multiple-output (SIMO), multiple-input single-output (MISO), and multiple-input multiple-output (MIMO), each FSO link (see Fig. 11.6) is equivalent to a 259 Advances in Optics: Reviews. Book Series, Vol. 3 SISO FSO link. For an IM/DD based link with AWGN and assuming an equiprobable data transmission ( 0 1 0.5 ), the probability of error conditioned on the |0 |1 . Note that |0 and |1 are received irradiance is BER 0.5 the conditional probabilities defined by averaging over the PDF of fading coefficient ∙ as [61]: |0 , |1 (11.23) 2 is the noise spectral density. Note that ∙ will depend on the channel where condition. Also note that Eq. (11.23) is only valid for SISO. To wrap up this section, a BER comparison of a SISO link in various turbulence situations with and without aperture averaging is presented to show both the effect of turbulence and using an optical lens at the receiver side, see Fig. 11.7. Fig. 11.6. A simplified illustration of m-th FSO link in a single-input multiple-output scheme. LD and PD are laser diode and photodetector, respectively. 10 10 BER 10 10 10 10 10 0 Clear Weak Weak+AF Mod Mod+AF Str Str+AF -1 -2 -3 -4 -5 -6 0 5 10 15 20 SNR (dB) 25 30 35 40 Fig. 11.7. Comparison of single-input single-output performance (BER versus SNR) over clear, weak, moderate and strong turbulence regimes with and without aperture averaging. ‘Weak’, ‘Mod’ and ‘Str’ refer to weak, moderate and strong regimes, respectively. ‘+AF’ denotes applying aperture averaging. 260 Chapter 11. Implementing Differential Signalling in Free Space Optical Communication Link Note that using Gauss-Hermite quadrature formula as in [79] Eq. (11.23) can be simplified to: √ ∑ , (11.24) where is the order of approximation, is the weight factor for the pth-order approximation, and is the zero of the pth-order Hermite polynomial. For values of and refer to mathematical handbooks such as [102]. 11.5. Differential Signalling The received signal in a FSO communication system is highly sensitive to the atmospheric effects such as fog, smoke, low clouds, snow, rain and the atmospheric turbulence [7, 17, 62, 63] that may result in severe power loss and channel fading. In NRZ-OOK IM/DD based systems, an optimal detection threshold level at the Rx can be used to distinguish the received ‘0’ and ‘1’ bits. However, under atmospheric turbulence the received optical signal will experience random intensity fluctuation as well as fading [64], which can result in the received signal power dropping below the Rx's threshold for a duration of milliseconds. For deep fading simply increasing the transmit power level and using a fixed optimal threshold level at the Rx are not the best options [63]. Most already-proposed detection methods rely on the knowledge of instantaneous or statistical channel state information (CSI). For instance, to resolve the fluctuation of threshold level, in [50] the maximum-likelihood sequence detection (MLSD) scheme was adopted, and it was shown that provided the temporal correlation of atmospheric turbulence is known MLSD outperforms the maximum-likelihood (ML) symbol-bysymbol detection technique. In practical applications ≅ 1 10 ms; then to maximize the link performance needs to be adjusted dynamically. In addition, MLSD suffers from high computational complexity. In [65] two sub-optimal MLSD schemes, based on the single-step Markov chain model, were proposed to reduce the Rx computational complexity; however they still require the CSI knowledge. Employing the pilot symbol (PS) assisted modulation (PSAM) scheme, and assuming that is known, CSI is acquired by inserting PS within the data stream [66]. However, obtaining an accurate-enough instantaneous CSI necessitates a non-negligible pilot overhead. In commercial FSO products, it is desirable to employ low complexity signal detection schemes with simple data framing and packetization structures in order to ensure infrastructure transparency [67]. In outdoor FSO links, a differential signalling scheme, also known as differential detection, was investigated in [35] to remove the effect of background noise. Also the same idea was adopted in [63] that used a pre-fixed optimal threshold level for various atmospheric channel conditions (rain, atmospheric turbulence, etc.). The detection technique did not rely on the CSI (with increased computational load at the Rx) and PS or 261 Advances in Optics: Reviews. Book Series, Vol. 3 a training sequence [63]. However, the simulation based investigation only considered narrow collimated beams without overlapping and with no experimental verification. To mitigate the fluctuation of pre-fixed optimal threshold level, the differential signalling scheme is preferred to AC-coupling (i.e., high pass filtering method) for a number of reasons including (i) no need to increase the transmit power to compensate for the filter attenuation; (ii) no baseline wander effect (i.e., disturbing the DC level of a signal); and (iii) removing the effects of the background noise. The basic concept of differential signalling is to send the signal and its inverted version simultaneously by using two pairs of Txs and Rxs over the same communication channel. Following reception of each signal by the corresponding Rxs and performing a subtraction operation at the final stage, the output signal is regenerated for further processing. With this scheme the challenges are to identify the received signals at the receiver and effectively exploit the potential of differential signalling method under various channel conditions. There are two main motivations for using the differential signalling method. First, it has been shown theoretically that provided Rxs are not saturated by the combined power of the received signal and the background noise level, the effect of ambient illumination can be cancelled out [35]. Second, in [63] a differential signalling scheme was adopted in an IM/DD OOK FSO links to overcome the variation of the threshold level caused by the channel fading. The benefit of differential signalling method in channels with large ambient illumination has been fully investigated in [35]. Therefore, in this research, only the performance of differential signalling in a fading channel is investigated. In addition to turbulence, pointing errors also results in threshold level fluctuations at the Rx, which can affect the link performance as well as making signal detection a challenging task [62, 68]. To mitigate signal degradation researchers have proposed a number of techniques including adaptive detection [62], more complex tracking systems [34], and spatial diversity [69]; also see [7] and the references therein. Adaptive detection techniques (ADT) either imposes computational load at the Rx or reduces the link throughput [65, 66]. When using ADT, the CSI or the temporal correlation of fading is required for data detection at the Rx [62]. FSO links with tracking systems requiring optics, monitoring and controlling circuits [70] are complex and costly to be used in commercial IM/DD NRZ-OOK FSO systems. Using spatial diversity with a number of Txs and Rxs will result in improved system performance for various channel conditions. However, there is still the need for a detection technique to extract the information bits from the combined signal [50]. The differential signalling technique can also be used to mitigate the fluctuation of threshold level due to the pointing errors. In this chapter, first the basic idea of differential signalling method is introduced. Then the challenges associated with the existing differential signalling methods are discussed and a solution is outlined. To the best of our knowledge no research work has been reported on the correlation between two channels in FSO systems with differential signalling. In this work theoretical, simulation and experimental investigation of the correlation are carried out. Correlation is shown to be an important key factor that needs 262 Chapter 11. Implementing Differential Signalling in Free Space Optical Communication Link considering. Also the theory of differential signalling method will be extended into channels with the pointing errors effect. 11.6. Differential Signalling Configuration The differential signalling system block diagram is depicted in Fig. 11.8. The NRZ-OOK signal and its inverted version ̅ are used to intensity-modulate two optical sources at wavelengths of and , respectively. By comparing with the optimal threshold level , where ∙ denotes expected value, the original data bit stream can be recovered (i.e., bit is 0 for and 1 elsewhere). Fig. 11.8. The system block diagram to implement differential signalling in correlated-channels conditions. T is the bit duration. OTx, BC, BS, ORx and TIA refer to optical transmitter, beam combiner, beam splitter, optical receiver, and transimpedance amplifier, respectively. Note that the optimal threshold level for ̅ is also of optical sources are given by: Γ 0 0 Γ . The output intensities Ii (i = 1,2) ̅, (11.25) where Γ denotes the electrical-to-optical conversion coefficient of optical sources. The outputs of optical sources are then passed through a beam combiner to ensure that both beams will be transmitted over the FSO channel of length . Note that the beam combiner is only used for alignment purposes and not for combining signals in the optical domain. The optical signals at the Rx end are given by: 0 0 Γ 0 0 Γ ̅, (11.26) where denotes the channel response including the effect of geometrical and atmospheric losses, pointing errors, and the turbulence. Here only the effect of turbulence is considered. 263 Advances in Optics: Reviews. Book Series, Vol. 3 At the Rx, the optical signal is passed through a 50/50 beam splitter and optical filters with the centre wavelengths of and , prior to being collected by an ORx. The generated photocurrents are amplified by TIA with outputs given by: 0 Γ 0 where is the PD responsivity, is gain of TIA, and variance , . The combined output Γ , ̅ Γ (11.27) is the AWGN with the zero mean is given by: ̅ Γ . (11.28) Note that in [35] it shown that for outdoor applications where the ambient noise effect is also embedded in , the impact of background noise is significantly reduced. 11.7. Differential Signalling and Turbulence 11.7.1. Optimal Detection Threshold Level A sampler with sampling at the centre of bit duration and a threshold detector are used to regenerate the transmit data. From Eq. (11.28), the optimal threshold level for is given by: Using [71] one obtains: Mean Var Mean 2 , . Γ Γ Var , Mean Var Var (11.30a) , 2 Var (11.29) (11.30b) where Mean ∙ denotes the average and Var ∙ introduces the variance. Here, , is correlation coefficient between the channels (i.e., and ). For simplicity, in 2⁄ and , are set. For the weak turbulence Eq. (11.30) Γ , regime follows Log-normal distribution with mean and variance , and , , respectively [71]. For Log-normal distribution one has [54]: 264 Var Mean exp 4 , exp 2 1 , 2 exp 4 , , , (11.31a) 4 , , (11.31b) Chapter 11. Implementing Differential Signalling in Free Space Optical Communication Link where : . Therefore, one has: , Var 2 , Mean exp 4 exp 4 , 0, exp 4 1 exp 4 , (11.32a) , , 1 2 2 . (11.32b) Since optical beams are in parallel and propagating very close to each other over the channel, then both beams will experience the same turbulence effects (i.e., , , ). Considering this approximation, one obtains: Var Mean 2 1 , 0, exp 4 (11.33a) , 1 2 . (11.33b) From Eq. (11.33a), it is seen that to recover the transmit bit stream, the optimal threshold level should be set to 0. This is similar to the work in [63] except for not considering the variance of the detection threshold in Eq. (11.33b) due to turbulence. The method proposed in [63] is effective only under constant fading conditions. However, for randomly varying fading scenario a Rx employing a fixed optimal threshold level is not the optimum and therefore alternative scheme should be considered to ensure improve FSO link performance. 11.7.2. Correlation between Channels From Eq. (11.33b), for 1 (i.e., the highly correlated channels), one has , Var 2 . In other words, turbulence does not affect signal detection provided the channels are highly correlated. According to [51], under the weak turbulence regime and , can be expressed in terms of the transversal distance between the Rx apertures the spatial coherence radius . Here, with parallel optical beams propagating over a LOS link, is in fact the distance between the propagation axes of beams. Thus the correlation coefficient between channels takes the form of [51]: , exp ⁄ , (11.34) where for a plane wave propagation model, the spatial coherence radius is given by ⁄ ⁄ 5 channels are considered uncorrelated Eq. (11.10). From Eq. (11.34), for 0.007 whilst for → 0 one can obtain , → 1. So, by adopting a small , one , can obtain highly correlated channels and, as a result, use an optimal threshold level independent of turbulence. This can be the challenging part of the differential signalling scheme; since the goal is to achieve the following features simultaneously: 265 Advances in Optics: Reviews. Book Series, Vol. 3 1. Both channels have to be highly correlated; 2. Signals are different, which necessitates minimum interference. In the next sections, it will be shown how these two challenges can be addressed by using the scheme illustrated in Fig. 11.8. 11.7.3. Channel Modelling In this section the differential signalling method in more details will be investigated and the effect of signal level, modulation extinction ratio, laser wavelengths, and correlation coefficient on the differential signalling performance will be described. Knowing that superscripts high and low denote corresponding high and low levels of the electrical signal , respectively, the electrical signals of LD in Fig. 11.8 are given by: ⁄2 ⁄2 bit 1 (11.35a) bit 1 (11.35b) Threshold, bit 0 Threshold. bit 0 Each bit in Eq. (11.35) is distinguished by the corresponding electrical signal level. Besides a constant threshold level, which is equivalent to the average signal level, is also included and one can regenerate the information bits by comparing signal to corresponding threshold level. In Section 11.6.1, this threshold level was defined as the optimal threshold level. and are used to internal-modulate two optical sources at and , respectively. The output of optical transmitter (OTx1) is wavelengths of given as: Due (11.36) equiprobable data transmission link the average power level ⁄2 ( 1, 2). By defining the extinction ratio , ⁄ 1 low and high power levels can be expressed as 2 and ⁄ 1 2 , respectively. Therefore, outputs of OTxs ( ) are expressed as: 266 to bit 1 Threshold. bit 0 Chapter 11. Implementing Differential Signalling in Free Space Optical Communication Link 1 1 bit 1 Threshold, bit 0 (11.37a) Threshold. bit 0 (11.37b) , where represents the atmospheric bit 1 The received optical power at the Rx turbulence. The outputs of ORxs are given by: bit 1 Threshold, 2 (11.38a) bit 0 bit 1 Threshold. 2 (11.38b) bit 0 As seen from Eq. (11.38) the threshold levels are affected by turbulence. If only a link with the wavelength is considered then the FSO link in Fig. 11.8 is simplified to a SISO link for which the average value and the variance of the received electrical signal are defined as [71]: Mean Mean Var where Φ , Var Φ bit 1 Threshold, Φ . Thus using Eq. (11.31) one has: (11.39a) bit 0 bit 1 Threshold, (11.39b) bit 0 267 Advances in Optics: Reviews. Book Series, Vol. 3 exp 4 , Threshold, Φ Mean Var bit 1 bit 0 1 , (11.40a) bit 1 Threshold. Φ (11.40b) bit 0 The expression in Eq. (11.40a) shows that the average of threshold level depends on Lognormal variances ( , ). Besides based on Eq. (11.40b), the threshold level fluctuates with the given order as predicted before. The combined output is given as: Φ bit 1 Threshold. Φ (11.41) bit 0 Mean ∙ and Var ∙ will be as given in Eq. (11.42). If laser beams are propagating very close to each other, then they experience the same turbulence strength and , , , therefore Eq. (11.42) leads to Eq. (11.43). Mean Mean Var Φ Var Φ Var Var 268 Var 2 2 Var 2 Var Φ , , , , , Φ , Var Φ Φ Var Var Var bit 1 Threshold, exp 4 bit 0 , (11.42a) bit 0 Mean , Var Mean Threshold, Φ Mean Φ Mean Mean bit 1 Mean 1 Var Var bit 1 Threshold, bit 0 (11.42b) (11.43a) Chapter 11. Implementing Differential Signalling in Free Space Optical Communication Link Φ 2 Φ 2 2 , , , bit 1 Threshold. Φ Φ (11.43b) bit 0 In Section 11.6.1, it is assumed that the links are the same. Here it is good to start with the same concept just to generalize the expressions derived before. Later on the general case where links are not the same be will discussed. By setting Φ Φ or equivalently the average of the threshold level is fixed to ~0 no matter how strong the turbulence is. On the other hand, the variance of the detection threshold under the same condition is defined as: Var 2 exp 4 , 1 Φ 1 , , , , (11.44) denotes the detection threshold level. The derived expression in Eq. (11.44) where is compatible with what was achieved in Eq. (11.33b). From Eq. (11.19), one can formulate the average and the variance of low level of the combined as: high level 2Φ Mean Mean 2Φ Var 4 exp 4 , 1 Φ Var 4 exp 4 , 1 Φ 2 , 2 , and , (11.45a) , (11.45b) , , , , , (11.45c) . (11.45d) Using Eq. (11.45) one can assess the quality of the signal using the Q-factor parameter as given by [72]: Q | | . (11.46) Eq. (11.46) will be used to study the effect of channel characteristics on the received signal in a differential signalling based FSO system in the following sections. 11.7.4. BER Expression For simplicity, the subtracted signal is assumed to be as: ̅ , (11.47) 269 Advances in Optics: Reviews. Book Series, Vol. 3 where is the total AWGN of the Rx. The fading of received intensities is exp 2 where denotes the average signal intensity without turbulence given as with Logand is a distributed normal random variable with mean and variance [54]. To ease the derivation two normal PDF [54]. Also for normalised PDF identical links meanings are considered as . Furthermore it is assumed that exp 2 , where is also a distributed normal random variable with and variance , . To normalize the condition is considered. mean , To obtain , , the following expression is used [71]: , , ln 1 (11.48) where is set to 1 for normalization and Var can be obtained the same as Eq. (11.32b). On the other hand, it can be shown that Var and Var are defined as [71]: Then Var 1 , E . (11.49) will be achieved by the following expression: , , ln 1 , 1 , 1 2 , , 1 , 1 . (11.50) Once equivalent Log-normal variance ( , ) is achieved it is possible to specify the PDF of a differential signalling FSO system by means of Eq. (11.15). Having PDF, one can calculate the average BER of the link using Eq. (11.23). 11.7.5. Numerical Analysis In this section, the effect of link parameters on the performance of differential signalling system will be analysed. The derived expressions will be used and wherever applicable will be supported with Monte-Carlo simulation. In previous section it was shown that by a constant threshold level of 0 can be used for different setting turbulent conditions. Also it was shown that for correlation coefficient ( , ) of 1 the fluctuation of the threshold level reaches its minimum value. In the analysis the adopted wavelengths were 830 and 850 nm and link was 1 km long. To calculate Mean and Var of SISO and differential signalling links, Eqs. (11.40) and (11.43) were used, respectively whereas Eq. (11.46) was used for calculation of the Q-factor. The given values of SNR denote the electrical SNR of the signal before the sampler block box as in Fig. 11.8. From Eqs. (11.40) and (11.43), it is deduced that the threshold level is dependent from extinction ratio ( ). To confirm this, Monte-Carlo simulation was used for both SISO and differential signalling systems for 5 and 10 with Φ Φ 5.7 mV, Φ 8.1 mV, , 1, and 0.5. To obtain each corresponding value, a 1 Mbits of data was transmitted with 270 Chapter 11. Implementing Differential Signalling in Free Space Optical Communication Link 10 independent iterations. The other parameters of the simulations were set according to Table 11.1. The obtained results are summarized in Table 11.2. Note that the fading varying frequency of the channel is in the order of 400 Hz [73], thus in both simulations and experiments, a low data rate was selected to avoid the need for storage of large number of samples. Table 11.1. The summary of FSO system properties used in the differential signalling simulation. Parameter Data rate Link length Turbulence strength Correlation coefficient , Received average optical power PD responsivity Noise spectral density Number of transmit bits Number of iterations Value 1 Mbps 1 km 1, 2.5, and 5 m ⁄ 10 0, 0.5, 0.8, and 1 20 dBm 0.4 102 dB/Hz 1 Mbits 10 Considering the theoretical and simulation results; it is seen that the proposed theory can predict the system behaviour accurately. Besides in agreement with Eqs. (11.40) and (11.43), for the same link condition but different extinction ratio ( ), mean value of threshold detection (Mean ) and standard deviation value of threshold detection ) are the same. ( Var Table 11.2. The theoretical analysis accompanied by simulation of mean of detection threshold (Mean) and standard deviation of detection threshold (√Var) of single-input single-output (SISO) and differential signalling (DS) links for different extinction ratios (ε) of 5 and 10 but fixed Ф1 = Ф2 = 5.7 mV, ФSISO = 8.1 mV, where channels are highly correlated 1,2 = 1, and Rytov variance (R2) is 0.5. SNR (dB) 5 12.2 10 14 Link SISO DS SISO DS a 7.6; [8.1, 0.5] 0.0; [0.0, 0.1] 7.6; [8.1, 0.6] 0.0; [0.0, 0.1] b √ 3.0; [3.3, 0.9] 1.9; [1.8, 0.1] 3.0; [3.4, 0.8] 1.9; [2.0, 0.1] For each case there is a pair of numbers separated by comma. The first number denotes theoretical analysis result while the pair shows the simulation outcome in form of expected value and standard deviation pair, respectively. a, b As discussed earlier for Φ Φ and , 1 , regardless of turbulence conditions, 0 and Var Mean , , . To prove this, another set of analysis 0 to was performed for a range of turbulence strength from almost a clear channel 1 . As shown before, the value of did not affect Mean and 271 Advances in Optics: Reviews. Book Series, Vol. 3 ; then for the simulation, the extinction ratio ( ) was set to 10 and SNR Var was changed by setting Φ . The results of the analysis are represented in Fig. 11.9. From Fig. 11.9 it is observed that for both SISO and differential signalling. For the theory can predict the Mean , there is a slight deviation between the theory and simulation, however Var both theory and simulation show the same trend. As predicted from Eq. (11.40) Mean and Var of SISO link are changing with the turbulence strength. Var of SISO link almost equals to the standard deviation of noise 1.32 mV for a clear channel condition (i.e., Rytov variance ( ) of ∼0) and it , values, which agrees well with Eq. (11.39b). Besides, increases for higher values of and Var of the SISO link. different SNRs results in various Mean Since in this analysis was fixed then SNR was changed by setting appropriately Φ . Thus, the gain of the TIA ( ), PD responsivity ( ), and LD output power ) can change the required threshold level whereas has no effect on it. On the other ( hand for the differential signalling link, Mean is constant for various turbulence conditions and different values of SNRs. This was expected because from Eq. (11.43a) when links have the same parameters (i.e., Φ Φ ) and the optical beams undergo the same turbulence effect; the required threshold level at the Rx is zero for various turbulence of the differential signalling link is also conditions and different SNRs. Var fixed for different turbulence conditions and various SNRs. From Eq. (11.43b) it is known ⁄ 1.9 mV, which agrees well with the simulation that Var , , results as in Fig. 11.9(b). In another set of analysis, the Q-factor for both SISO and differential signalling links are compared under different conditions. From Eqs. (11.40), (11.43), and (11.46) it is seen that in contrary to the Mean and Var , the Q-factor also depends on . Therefore, is set to 5 and 10 for SNRs of 10, 12, 14 dB and the same turbulence strength is used. The theoretical and simulation results are illustrated in Figs. 11.9(c) and 11.9(d). Figs. 11.9(c) and 11.9(d) confirm that the proposed theory predicts the Q -factor for both SISO and differential signalling links. In Fig. 11.9(c) 5, where is 10 for Fig. 11.9(d). For a clear channel ⁄ 10 and as increases, the Q-factor tends to reduce. Changing condition, Q from 5 to 10 has no effect on the Q-factor of the differential signalling link while the SISO 5 in a turbulent channel. link shows a lower Q-factor for So far it has been shown out that for , 1 and Φ Φ , both the Mean and of SISO and differential signalling links and the Q-factor of the differential Var signalling link are independent of and change with Φ . The results also showed that changing Φ has no effect on the Mean and Var of the differential 1. signalling link for , 272 Chapter 11. Implementing Differential Signalling in Free Space Optical Communication Link 12 8 6 p 4 Line+hollow marker Solid marker 2 SISO DS SNR = 10 dB SNR = 12 dB SNR = 14 dB 6 Var(Vthr esh ) mV M ean(Vthr esh ) mV 10 7 SISO DS SNR = 10 dB SNR = 12 dB SNR = 14 dB  i = 10 1,2 = 1 Theory Simulation 5  i = 10 1,2 = 1 Line+hollow marker Solid marker Theory Simulation 4 3 2 0 0 0.2 0.4 0.6 0.8 1 1 0 0.2 0.4 0.6 2 (a) 1 (b) 5.5 7 SISO DS SNR = 10 dB SNR = 12 dB SNR = 14 dB i = 5 1,2 = 1 6 5 Line+hollow marker Solid marker SISO DS SNR = 10 dB SNR = 12 dB SNR = 14 dB  i = 10 1,2 = 1 5 4.5 4 Theory Simulation Line+hollow marker Solid marker 3.5 4 Q Q 0.8 <2R <R Theory Simulation 3 2.5 3 2 1.5 2 1 1 0 0.1 0.2 0.3 0.4 0.5 <2R (c) 0.6 0.7 0.8 0.9 1 0.5 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 <2R (d) Fig. 11.9. Simulation results of: (a) mean of detection threshold Mean(Vthresh), (b) standard deviation of detection threshold √Var(Vthresh), and (c, d) Q-factor versus Rytov variance (R2). The comparison is performed for a range of turbulences and SNRs for: (a) and (b) εi = 10, (c) εi = 5, and (d) εi = 10. SISO and DS refer to single-input single-output and differential signalling, respectively. εi and 1,2 denote extinction ratio and correlation coefficient, respectively. Note that in (c) and (d) the error bars are too small to be seen. It is important to note that Eq. (11.13) gives different results for and , which results in different Log-normal variances (i.e., , , ). Therefore, the simplified expressions given in Eq. (11.34) are no longer valid. Also note that spatial coherence radius ( ) in Eq. (11.10) is a function of wavelength, which leads to different values of correlation coefficient ( , ) for the same FSO system. This issue necessitates us to define a constraint on how distinct the wavelengths can be and also to study the effect of correlation coefficient on the system performance. 273 Advances in Optics: Reviews. Book Series, Vol. 3 To define a constraint for the difference of two wavelengths, which still validates the use of Eqs. (11.10), (11.13), and (11.34), the derivatives of , , , and are taken with respect to . After a series of mathematical simplification, one has: ∆ where ∆ ⁄2 and ∆ , ∆ | , , 2 , ∆ ∆ , (11.51a) , ln (11.51b) , |. ∆ , (11.51c) Considering the rule of thumb that a 10 % tolerance relative to the absolute value is is extracted. This acceptable, from Eqs. (11.51a) and (11.51b) the criteria of ∆ ⁄ and ), means that for criteria, which is independent from the FSO channel (i.e., ∆ ⁄ , Eqs. (11.10) and (11.13) give approximately the results for both wavelengths with 10 % relative deviation. However, in Eq. (11.51c) a fixed constraint cannot be derived. It can be easily shown that assuming the same rule of thumb of ∆ , ⁄ , 0.1 the criteria based on Eq. (11.51c) is given by: ∆ ⁄ . (11.52) Fig. 11.10 shows ∆ ⁄ with respect to ⁄ , a characteristic, which is independent from wavelength, and the link distance or the turbulence strength. It is deduced from ⁄ → 0 the range of selecting and broadens Eq. (11.52) that for ⁄ (i.e., ∆ ⁄ → ∞), whereas for 0 0.26 the range of applicable wavelengths is reduced (i.e., ∆ ⁄ 0.47). Therefore, there is a trade-off between how close the beams have to be and how different the wavelengths can be. The two differential signalling link conditions (i.e., Φ Φ and , → 1) are ideal and in reality there are deviations from the ideal scenario. Thus, the mean value of threshold ), the variance value of threshold detection ( Var ) and detection (Mean the Q-factor are compared for SISO and differential signalling links for the same SNR but different values of Φ over a range of correlation coefficient ( , ). The SNR was set to 12 dB and extinction ratio ( ) was changed to 5, 10, and 20. The turbulence strength of 0.5 was considered and the results are depicted in Fig. 11.11. The value of , spans from uncorrelated channels conditions (i.e., , 0) to fully correlated channels 1). The accuracy of the proposed theory for , range is obvious condition (i.e., , from the good agreement between simulation and predicted results as depicted in Fig. 11.11. 274 Chapter 11. Implementing Differential Signalling in Free Space Optical Communication Link 10 0 " 6 =60 10 1 10 10 -1 -2 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 dr =; 0 Fig. 11.10. Δλ/λ0 plotted with respect to dr/0. The graph shows the relation between the tolerable optical sources wavelengths and the distance between the optical receivers, while the channels are kept independent. The graph is obtained from Eq. (11.28). As expected from Eq. (11.39a) the mean value of the detection threshold (Mean ) of the SISO link in Fig. 11.11(a) depends on the link parameters. Mean of the SISO link decreases when extinction ratio of SISO link ( ) is higher. To keep SNR at 12 dB, the value of Φ was changed to 7.9, 6.4, and 5.8 mV. On the other hand, from Eq. (11.39a) it is observed that for higher Φ the resultant Mean is also higher. The variance value of the detection threshold (Var ) of the SISO link in Fig. 11.11(b) also is dependent on the link parameters and since for a fixed SNR higher requires lower Φ , then Var of the SISO link for higher is reduced. This deduction is in agreement with Eq. (11.39b). Fig. 11.11(c) illustrates the Q-factor for the SISO link, which are obtained from | in Eq. (11.46) is dependent on Mean Eq. (11.46). The numerator |Mean both and Φ , however the effects of Φ and are opposite, therefore for | value. On the same SNR, higher Mean results in the same |Mean the other hand, Var Var is increased for the same SNR and higher Φ . Therefore, the Q-factor of the SISO link is lower for lower values of . Mean of the differential signalling link in Fig. 11.11(a) is also dependent on the link parameters and since Φ Φ the value of Mean is non zero. But as discussed for the SISO link, for the same SNR and higher the value of Mean is reduced. and the Q-factor of the differential signalling link depend on not only the Var link parameters but also on correlation coefficient ( , ). As seen in Fig. 11.11(b) of the differential signalling link reaches the minimum value of total Var standard deviation of noise ( , , ⁄ 1.9 mV) for , 1. Also it can be seen 275 Advances in Optics: Reviews. Book Series, Vol. 3 that lower results in smaller Var . Fig. 11.11(c) shows the Q-factor of the 10 ⁄ at differential signalling link that achieves the maximum value of Q 1. As discussed for the SISO link, the same SNR and higher leads to higher , values of the Q-factor. SISO DS =5  = 10  = 20 SISO 9  = 10  = 20 3.2 Var(Vt hr esh ) mV 7 6 5 4 3 3 2.8 2.6 2.4 2.2 p M ean(Vthr esh ) mV =5 3.4 8 2 2 1 0 DS 1.8 0 0.2 0.4 ; 1;2 0.6 0.8 1.6 1 0 0.2 0.4 (a) ; 1;2 0.6 0.8 1 (b) SISO DS =5  = 10  = 20 2.4 2.2 Q 2 1.8 1.6 1.4 0 0.1 0.2 0.3 0.4 0.5 ; 1;2 0.6 0.7 0.8 0.9 1 (c) Fig. 11.11. Simulation results of: (a) mean of detection threshold Mean(Vthresh), (b) standard deviation of detection threshold √Var(Vthresh), and (c) Q-factor. The comparison is performed between single-input single-output (SISO) link and differential signalling (DS) and for different values of extinction ratios (εi) over a range of correlation coefficient (1,2). Based on the analysis, if achieving higher SNR in a differential signalling link is desirable, then increasing is the preferred option. Of course, increasing needs to be done with respect to the span of laser L-I curve linear region to avoid pulse shape distortion. 276 Chapter 11. Implementing Differential Signalling in Free Space Optical Communication Link To validate the work in Section 11.6.4, the BER results from Eq. (11.50) were compared with the simulation data. The parameters adopted for the FSO system investigated are given in Tabl. Fig. 11.12 shows the predicted and simulated BER versus SNR obtained from the theory as well as the performed simulation based on Eq. (11.50) for the FSO with 0 , and 0.8 and 1 10 and differential signalling for , ⁄ 2.5 10 m . The simulation results are presented by large markers for each case with the error bars to show the tolerance of the simulated values. To obtain each point, 10 iterations were carried out with 1 Mbps of transmit bit stream for each iteration. Note that for BER < 10-6, the simulation results were zero, and therefore are not shown in the graph. It is clear from the figure that the initial assumption that with a Log-normal distribution does indeed lead to a good approximation of PDF of the differential signalling method for the weak turbulence regime. 10 10 BER 10 10 10 10 10 0 -2 -4 -6 C2n =1.0x10-15 ,1,2 =0.0 C2n =2.5x10-15 ,1,2 =0.0 -8 C2n =1.0x10-15 ,1,2 =0.8 C2n =2.5x10-15 ,1,2 =0.8 -10 Line+hollow marker Solid marker -12 0 2 4 6 Theory Simulation 8 10 SNR (dB) 12 14 16 18 20 Fig. 11.12. BER versus SNR in dB of an FSO system with the differential signalling method. The comparison is carried out for various turbulence strengths (Cn2) and correlation coefficients (1,2). Solid lines marked with small markers are based on the derived equations whilst large markers are obtained from the simulation. It was discussed in Section 11.5 that when the channels are fully correlated (i.e., , 1) the effect of the turbulence has the minimum influence on the received signal. In the next 5 10 m ⁄ was considered and the correlation according was changed step based on Tabl. The results in Fig. 11.13 show that when the correlation coefficient increases, the performance of differential signalling method improves in term of 1 the FSO system mitigating the turbulence effect. As shown in Fig. 11.13, for , with the differential signalling scheme offers almost the same performance as in the clear 0 ) in Fig. 11.13, the simulation error channel. For the uncorrelated case (i.e., , increases so that the error bars shown are negative, which are not shown in the logarithmic scale. 277 Advances in Optics: Reviews. Book Series, Vol. 3 10 10 BER 10 10 10 10 10 0 -2 2 Cn = 5x10 -4 -15 Clear -6 1,2 = 0 1,2 = 0.5 -8 1,2 = 0.8 1,2 = 1 -10 Line+hollow marker Solid marker -12 0 2 4 6 Theory Simulation 8 10 SNR (dB) 12 14 16 18 20 Fig. 11.13. BER versus SNR in dB of an FSO system with the differential signalling method for Cn2 = 5×10-15 m-2/3 and a range of correlation conditions (1,2). Solid lines with small markers are based on theory whereas large markers are obtained from simulation. The plus maker denotes the clear channel condition. 11.7.6. Atmospheric Turbulence Experiment To prove the concept of differential signalling technique, the experimental work for the proposed system as given in Fig. 11.14 will be outlined. Chamber Length = 6 m OTx1 OTx2 Mirror BS ORx1 O F 2 Hot Air O F 1 ORx2 Fan Fig. 11.14. Block diagram of the atmospheric turbulence and differential signalling experiment. PATH1 and PATH2 are referring to uncorrelated and correlated paths, respectively. OTx, BS, OF, and ORx are optical transmitter, beam splitter, optical filter, and optical receiver, respectively. 278 Chapter 11. Implementing Differential Signalling in Free Space Optical Communication Link According to proposed scheme shown in Fig. 11.8, an experimental setup for the proposed 0) method was developed to evaluate its performance for both uncorrelated (i.e., , and correlated (i.e., , → 1) channels conditions as depicted in Fig. 11.14. Snapshots of the setup are also shown in Fig. 11.15. The laser beams (see Fig. 11.15(a)) were launched into a chamber of length 6 m, emulating an outdoor uncorrelated FSO channel (see Fig. 11.15(b)). The incident and reflected ray paths are labelled as PATH1 and PATH2, respectively (see Fig. 11.14). In PATH1 optical sources were spaced apart by a minimum ⁄ ⁄ 5 mm to ensure uncorrelated fading conditions (i.e., 5). distance of An adjustable mirror positioned at the other end of the chamber was used to increase the path length by reflecting back the beams. The reflected beams indicated by PATH2 in Fig. 11.14 were kept as close as possible to each other to ensure high correlation between the two paths (note that PATH2 in Fig. 11.14 corresponds to FSO channel in Fig. 11.8). (a) (b) Fig. 11.15. Experimental setup of atmospheric turbulence and differential signalling: (a) OTxs and ORxs at one end of the chamber, and (b) atmospheric camber with temperature sensors to measure temperature gradient, and a pipe to isolate either PATH1 or PATH2 from the turbulent condition of the chamber. OF, OTx, ORx, and BS are optical filter, optical transmitter, optical receiver, and beam splitter, respectively. Heater fans were used to generate turbulence in the chamber, see Fig. 11.14. To measure , the method of thermal structure parameter was used (based on temperature gradient measurement) as in [74]. The temperature gradient was measured using 20 temperature sensors positioned along the chamber, see Fig. 11.15(b). At the Rx end, the reflected beams passed through a 50/50 beam splitter and were applied to two identical PIN PDs after optical filters, see Fig. 11.14 and Fig. 11.15(a). The outputs of PDs were captured using a real-time digital storage oscilloscope for further processing in MATLAB®. First investigated was the effect of turbulence on the uncorrelated path within the chamber. The reflected beams (i.e., PATH2) were passed through a pipe positioned within the chamber. The pipe ensured that propagating beams inside it did not experience any turbulence, see Fig. 11.15(b). Similarly, the effect of turbulence on the correlated path was investigated by isolating the uncorrelated channels (i.e., optical beams in PATH1 279 Advances in Optics: Reviews. Book Series, Vol. 3 propagating through the pipe), see Fig. 11.14. The amplitude of and ̅ were then set in order to ensure that both received electrical signals and had the same amplitude of ~300 mV, which is equivalent to Γ Γ criterion. Fig. 11.16 illustrates the histogram of the detection threshold level obtained from the experiment. Note that due to the hardware dissimilarities, the average of the detection threshold is non-zero. However, since the offset levels are due to ORx1 and ORx2, then the problem can be resolved by adjusting the offset level of the output signal. Table 11.3 shows all the key parameters adopted in the experiment. The recorded data were processed and the detection threshold level was extracted from signals. Fig. 11.17 illustrates the sampled signal as well as the obtained detection threshold level of . The measured mean (indicated by Mean and √Var , respectively) for and standard deviation of correlated and uncorrelated channels are summarized in Table 11.4. 35 30 Uncorrelated Darkness 25 25 Number of Points Number of Points 20 15 10 20 15 10 5 0 -864.35 -864.3 -864.25 Uncorrelated Darkness 30 5 -864.2 -864.15 -864.1 -864.05 Threshold Level (mV) -864 -863.95 0 -887.15 -887.1 -887.05 -887 (a) -886.8 -886.75 (b) 35 35 Uncorrelated Darkness 30 Uncorrelated Darkness 30 25 Number of Points 25 Number of Points -886.95 -886.9 -886.85 Threshold Level (mV) 20 15 20 15 10 10 5 5 0 -915.74 -915.73 -915.72 -915.71 -915.7 -915.69 -915.68 -915.67 -915.66 -915.65 Threshold Level (mV) (c) 0 -949.92 -949.9 -949.88 -949.86 -949.84 Threshold Level (mV) -949.82 -949.8 (d) Fig. 11.16. Histograms of the detection threshold levels of the differential signal threshold (Vthresh) for atmospheric turbulence and differential signalling experiment: (a) uncorrelated channels in dark room; (b) uncorrelated channels in lit room; (c) correlated channels in dark room, and (d) correlated channels in lit room. 280 Chapter 11. Implementing Differential Signalling in Free Space Optical Communication Link Table 11.3. The setup parameters for atmospheric turbulence and differential signalling experimental. 1.05 1 1 0.95 0.95 0.9 0.9 0.85 0.85 0.8 v1 (V) v1(V) Link 2 Link 1 Parameter Data rate NRZ-OOK Chamber length Sampling rate Number of recorded points for each iteration Number of total iterations Optical transmit power Divergence angle PD responsivity Wavelength Optical transmit power Divergence angle PD responsivity Wavelength Optical receiver noise rms 0.8 Value 100 kbps 6m 2.5 M Sample/sec 1 M points 500 10 dBm 9.5 mDeg 0.3 A⁄W 830 nm 3 dBm 4.8 mDeg 0.4 A⁄W 670 nm 1.5 mV 0.75 0.75 0.7 0.7 0.65 0.65 0.6 1 2 3 4 5 Sample Point (a) 6 7 8 9 x 10 5 2.48 2.485 2.49 2.495 2.5 Sample Point 2.505 2.51 2.515 5 x 10 (b) Fig. 11.17. The sampled v1 signal with the estimated detection threshold during atmospheric turbulence experiment. The signal is in blue colour, where the dashed red line with circle markers refers to the estimated detection threshold. As predicted from Eq. (11.33a), for both uncorrelated and correlated conditions the measured mean value is zero. Fig. 11.18 shows an image taken from the oscilloscope screen illustrating how signals and for correlated channels are affected under the same turbulence conditions. Note that turbulence strength and the laser modulation index in Fig. 11.18 were deliberately set to relatively small values in order to better illustrate the correlation between and . 281 Advances in Optics: Reviews. Book Series, Vol. 3 Table 11.4. The summery of the experimental measurement results for turbulence effect on differential signalling. Mean (mv) and √Var (mv), denote the measured mean of detection threshold, variance of detection threshold of the differential signal. Cn2 and 1,2 denote obtained turbulence strength and correlation coefficient. Channels condition Uncorrelated (dark room) -864.2 Uncorrelated (lit room) -886.9 Correlated (dark room) -915.7 Correlated (lit room) -949.9 √ 43.4 45.5 12.9 12.9 ⁄ , 5.11 10 0.08 5.21 10 0.72 Fig. 11.18. Original (v1 in yellow and top) and inverted (v2 in green and bottom) signals captured on the oscilloscope during atmospheric turbulence experiment. It is expected to obtain √Var 2 from the measurements. However, given the rms noise of optical Rx in Table 11.3, the measured √Var in Table 11.4 is different from the predicted value of 2 2.1 mV. This difference might be due to imperfect correlation between channels in PATH2 and using two (not very close) wavelengths of 670 and 830 nm, which could lead to dissimilar . In the experiment 0.17, which corresponds to the weak turbulence regime [30]. Considering Eq. (11.51a) and using the 0.1, it is expected to have ∆ ⁄ 0.09. Note that in rule of thumb, to have ∆ ⁄ the experiment, the accuracy limit of Eq. (11.33b) does not apply perfectly (as 0.21, which corresponds to a maximum wavelength deviation of 160 nm around ∆ ⁄ the central wavelength of 750 nm). Using measured signals correlation coefficient ( , ) was estimated, which are presented in Table 11.4. The estimated , (for the correlated case) are relatively high but do not 282 Chapter 11. Implementing Differential Signalling in Free Space Optical Communication Link correspond to the ideal case of , 1. Other effects that could lead to inaccuracy of the measurement were the noise associated with the oscilloscope and the vibration of the whole setup. However, since it was intended to demonstrate only the difference between uncorrelated and correlated situations and during the entire measurement the same setup was used, these effects are not critical in the final conclusion. In addition to using two wavelengths and spatially closer beams from Eq. (11.10) it is evident that longer transmission spans will lead to larger values of , which in turn helps to achieve a highly correlated channels condition (i.e., , → 1) [75]. In [35], a similar differential signalling based technique was proposed to reduce the effect of background noise in the received signal. The above experiment was carried out in both dark and fully lit environments (with ambient light power level of 45 dBm and 18 dBm, respectively), see Table 11.4. A negligible difference between the standard deviation of detection threshold results in these two cases is noticed. This testifies that under the experimental conditions, the background noise was not dominant. Thus, the reduction in √Var values is due to the theory explained in Section 11.5 rather than the background noise level. Using the derived analytical expression of the variance of the detection threshold, it was shown that the fluctuation in the optimal threshold level highly depended on the correlation between the propagating optical beams. Thus the differential signalling technique is attractive when highly correlated FSO channels can be established. This deduction was validated by means of experimental investigations under uncorrelated and correlated conditions. Also note that to achieve a high correlated channel condition light sources with close wavelengths, spatially closer beams and longer transmission distance are critical to have. 11.8. Differential Signalling and Pointing Errors 11.8.1. Channel Modelling In the previous sections, the effect of differential signalling system in a turbulence channel was investigated and in this section the performance of the same system with the pointing errors is investigated. Considering Eq. (11.30), a clear weather condition and negligible atmospheric turbulence is assumed. Also it is assumed that channel coefficient ( ) includes the geometrical loss and pointing errors effects. The concept outlined in [76] is adopted to describe the pointing errors at Rxs, see Fig. 11.19. Rxs are assumed to have the same aperture diameter as well as the same electrical and optical characteristics. The aperture diameter is and the laser beam spot at the Rx transverse plane has a radius of . Besides the instantaneous radial displacement between the beam centroid and the aperture centre is denoted by . In terrestrial FSO communication systems the fading coefficient due to the geometrical loss and pointing errors is given by [76]: 283 Advances in Optics: Reviews. Book Series, Vol. 3 ; and , exp (11.53) correspond to the geometrical loss and equivalent beam-width, √ ⁄2 respectively. Note that erf and where where [76]. In most practical applications the beam divergence is ≪ 1 Rad), which leads to . Therefore, it is possible to set and . pointing errors , , by selecting the appropriate values for displacement has two components known as the boresight (displacement between beam centre and centre of the detector) and jitter (offset of the beam centre at detector plane) [49]. and erf small (i.e., √ y rb rj r 2wRX ds 2 x Beam Footprint Aperture Fig. 11.19. Rx aperture and a laser beam footprint at the Rx transverse plane. The boresight displacement ( ) represents a deviation originating from thermal expansion of the buildings [34] and determines the mean offset of pointing errors [77], whereas is a random variable originating from building sway and vibration [34]. From the statistics point of view the jitter corresponds to the random variation of the beam [77]. footprint around the boresight direction with the jitter variance of In terrestrial FSO links, the jitter consists of both vertical and horizontal components [34]. Thus, without loss of generality, here the focus will be on the deviation along either vertical or horizontal axis, which can be further extended to the other axis. It is shown in ) and the second moment [34] that has Rician PDF with the average (Mean ) given by: ( Mean 284 exp exp , , (11.54a) (11.54b) Chapter 11. Implementing Differential Signalling in Free Space Optical Communication Link where Var ⁄2 . Therefore the variance of will be: exp . exp (11.55) Since E E ̅ , where is an integer, thereafter by means of Eq. (11.54), the average and variance of Eq. (11.30) are derived as: Var Mean Var Var Mean 2 Mean Var , Var (11.56a) 2 , is the correlation coefficient between two channel coefficients of where 2⁄ . Note that to derive Eq. (11.56), it was assumed that Γ (11.56b) and . The dynamic response of a building with applied live loads depends on the directional stiffness, as well as the height, size and topology. The tip displacements of a tall building can be as large as tens of centimetres due to the normal wind loads. However, irrespective of the height and stiffness of the building, the relative displacement of Rx1 and Rx2 is almost zero. The segment of the Rx mast between the two Rxs can be reasonably assumed to be rigid if the mast is properly designed according to the building standards (i.e., earthquake or wind) and if the distance between the two Rxs is very small compared to in Fig. 11.20), then , . the overall height of the building ( ≅ , Considering the short distance between the two Rxs the relative displacement due to the thermal expansion, which results in , , , can be neglected. The same is true for the relative displacement between the Txs. In view of the above, it can be assumed , , and  ≅ 1. Fig. 11.20. Effect of building movement on FSO Rxs with exaggeration. On the left there is the Tx building hypothetically without movement and on the right the Rxs are installed on top of a building, which is influenced by the effect of building movement. 285 Advances in Optics: Reviews. Book Series, Vol. 3 Considering , , , Eq. (11.56) will be: Mean Var 2Var 1 0,  (11.57a) 2 . (11.57b) Therefore, regardless of the strength of pointing errors the detection threshold level can be set to zero but will experience fluctuation with the given variance. Considering 2 , thus leading to the elimination of  ≅ 1, Eq. (11.57b) simplify to Var pointing errors at the Rx. Therefore for the system shown in Fig. 11.8 provided links are identical, and Rxs are mounted on the same fixture structure, the threshold level could be set to zero for a range of pointing errors strength. To estimate the equivalent parameters of the differential signalling based link, a SISO link with a Rayleigh pointing errors PDF is used as the equivalent link. The pointing errors parameters of the equivalent SISO link is found so that SISO has the same pointing errors variance as differential signalling. For the simplified case where , 0, , Var assuming that and , and simplifying Var , the following equation is derived from which of the equivalent pointing errors PDF can be obtained: 2 1 ⁄ 2 1 2 1  ⁄ , (11.58) when , 0, Rician distribution becomes Rayleigh and it is assumed that PDF , of the equivalent pointing errors to perform comparison is also Rayleigh. 11.8.2. Pointing Errors Experiment To validate the proposed concept practically the experimental setup shown in Fig. 11.21 has been used. Both Txs and Rxs modules were located at one end of a 6 m long indoor atmospheric chamber. In order to double the link length a mirror was used at the other end of the chamber to reflect back the beams (note that the mirror is not shown in the figure). Since the optical beams were in parallel with no overlap at the Rx plane, no optical filters were used. The outputs of Rxs were recorded using a sampling oscilloscope for further analysis via MATLAB®. The amplitude of modulating signal was set such that the Rx output levels were at ~10 mV . Table 11.3 summarizes the key system parameters adopted. Note that in this experiment, the number of iterations was only 50 since the vibration had a fixed pattern rather than being random. To simulate pointing errors condition, both Rxs were positioned on a vibration stand vibrating at a frequency of 5 Hz with the deviation of ~2 mm in the vertical direction (a sinusoidal waveform was used to stimulate the vibrator). In practical scenarios FSO links will experience vibrations in both axes but here the vibration is generated only on the vertical axis. Note that it has the same effect as for vibrations on the horizontal axis. 286 Chapter 11. Implementing Differential Signalling in Free Space Optical Communication Link Fig. 11.22 depicts a captured image from the oscilloscope screen for two signals. Note that the laser modulation index in Fig. 11.22 was deliberately set to relatively small value in order to better illustrate correlation between and . As shown the effect of pointing errors on both signals are highly correlated. The recorded data were processed and the detection threshold level was extracted from signals. Fig. 11.22 illustrates the sampled signal as well as the detected signal . The histogram of the recorded threshold levels for the differentiated signal is illustrated in Fig. 11.23. As in the previous experiment, the dissimilarity of the offset levels added by ORxs results in a non-zero DC offset in the detection threshold level. Fig. 11.24 illustrates the histogram of the detection threshold level obtained from the experiment Fig. 11.21. The pointing errors and differential signalling experimental setup. OTx and ORx are optical transmitter and optical receiver, respectively. Fig. 11.22. An image taken from the oscilloscope screen during the pointing errors and differential signalling experiment. Top yellow signal is v1 whereas bottom green one is v2. 287 Advances in Optics: Reviews. Book Series, Vol. 3 Based on the covariance matrix of the received signals, the correlation coefficient ( ) of 0.92 was obtained from the measurements, which agrees well with the assumption of  → 1 made in the analysis. The measured standard deviation of , and are presented in Table 11.5. -9.846 -9.84 -9.848 -9.85 -9.852 -9.854 -9.86 v2(V) v2(V) -9.85 -9.856 -9.858 -9.87 -9.86 -9.862 -9.88 -9.864 -9.89 -9.866 1 2 3 4 5 Sample Point 6 7 (a) 8 9 4.07854.079 4.0795 4.08 4.08054.081 4.08154.0824.08254.083 4.0835 Sample Point 5 x 10 (b) 5 x 10 Fig. 11.23. The sampled v2 signal with the estimated detection threshold during pointing errors and differential signalling experiment. The signal is blue colour where the dashed red line with circle markers refers to the estimated detection threshold. Table 11.5. The summery of the measurements results for pointing errors and differential signalling experiment. Signal √Var mv 24.75 27.00 4.58 7 6 Number of Points 5 4 3 2 1 0 -9.914 -9.912 -9.91 -9.908 -9.906 -9.904 -9.902 -9.9 Threshold Level (V) -9.898 -9.896 -9.894 Fig. 11.24. Histogram of the detection threshold levels of the differential signal (Vthresh) for pointing errors and differential signalling experiment. 288 Chapter 11. Implementing Differential Signalling in Free Space Optical Communication Link As discussed before for the same pointing errors jitters (i.e., , , ), the variance of ) in Eq. (11.57b) results in almost the same range for both channel coefficient (Var signals. The close match between the measured values of standard deviation of and confirms the deduction. On the other hand, according to Eq. (11.57b), if the effects of pointing errors on both signals are highly correlated, then standard deviation will significantly be reduced by 2 . However the measured value of 4.58 mv slightly differs from the predicted value of 2 2.1 mV. This can be due to the small difference in the values Var and Var . Although the experiment was conducted over a 12 m long of Var FSO link, the investigation can be extended to longer spans. Using the same laser beam, but over a longer link span, the geometrical attenuation and the beam footprint at the Rx will be larger. Higher geometrical loss will reduces whereas larger optical footprints will lead to reduced and increased . However, as seen from Eq. (11.57b), for  → 1 these parameters will have no effect on the resultant variance. , (from Eq. (11.57)) the variance of equivalent differential Using the predicted signalling pointing errors ( , ) can be determined for a range of  . Fig. 11.25 depicts the jitter standard deviation against channel correlation (  ) for the SISO link and equivalent link of differential signalling for receiver aperture radius ( ) of 10 cm, laser beam radius at receiver ( ) of 100 cm , and the jitter variances of 10 cm (i.e., 10 cm ). It is observed that the pointing errors induced fading effect reduces , , with increasing value of  . Also from both Eq. (11.57) and Fig. 11.25 it is seen that for  0.5 one obtains . 14 12 <j (cm) 10 8 6 4 SISO DS PE = 0.5 2 0 0 0.1 0.2 0.3 0.4 0.5 ; PE 0.6 0.7 0.8 0.9 1 Fig. 11.25. Jitter standard deviation of the equivalent differential signalling (DS) pointing errors versus correlation coefficient (PE), for the single-input single-output (SISO) system with differential signalling and for receiver diameter (ds) of 20 cm, beam radius (wRx) of 100 cm, and jitter variances of 10 cm (i.e., j,1 = j,2 = 10 cm). 289 Advances in Optics: Reviews. Book Series, Vol. 3 11.9. Differential Signalling and Manchester Code 11.9.1. System Configuration To the best of our knowledge, Manchester code has not been used to mitigate the fading effect of the channel in a FSO link. In fact, Manchester code is used to achieve clock synchronization. It also can be used to remove the DC component of the signal and to avoid a long stream of logic ‘1’ or logic ‘0’ [78]. In this work Manchester code is adopted to remove the need for two parallel highly correlated links in a differential signalling system. Fig. 11.26(a) illustrates the proposed concept, where the input bit stream and its inverted version are applied to the encoder, which is fed directly to the optical source. The output of the encoder is the Manchester code (also known as phase encoding) word in which the encoding of each data bit has at least one transition at the centre of each bit period, and has a bandwidth twice that of the input signal [78]. Input Bit A Manchester Encoder 1 A 1 B 0 0 1 1 0 1 T Received Signal (a) B C r1 Sample Combiner r2 Received Bit D A 0 1 1 0 A Sampling at t T/4 Sampling at t 3 T/4 1 0 B Output Bit 1 t t+T/4 t+3× T/4 (b) Fig. 11.26. The required signal processing to perform differential signalling using one FSO link. The procedure is shown for a sequence of 01101 bits as an example: (a) the required signal shaping at the Tx, and (b) the required signal recovery at the Rx where two samples (i.e., r1 and r2) are taken at the presented intervals. In the proposed differential signalling scheme, the received raw signal is processed prior to quantization. At the Rx the regenerated bit stream is passed through sampler modules and a combiner, which simply subtracts the sampled outputs, to recover the NRZ data stream as shown in Fig. 11.26(b). As seen so far in the proposed technique the original and inverted versions of signal are transmitted by means of Manchester code and a single FSO link whereas in Section 11.6 and Section 11.7 this was done using two distinct FSO links. Thus, the proposed method ensures that the channels are highly correlated. Besides, it eliminates the need of optics for combining and separating two FSO links as in Section 11.6.3. 290 Chapter 11. Implementing Differential Signalling in Free Space Optical Communication Link 11.9.2. Manchester Code Experiment To prove the validity and benefit of the proposed method an experimental test bed for a SISO FSO link was developed to measure the variation of the threshold level and the Qfactor. The experiment was using the same indoor atmospheric chamber. The method (in unit of m ⁄ ), which is known described in Section 11.6.6 was used to estimate as refractive index structure coefficient and shows the strength of the turbulence strength. For each scenario 250 data sets were recorded. The summary of the experimental setup for the 830 nm wavelength is summarized in Table 11.3. (a) (b) Fig. 11.27. The experimental setup: (a) equipment at the Tx side, and (b) artificial atmospheric channel. AWG and OTx denote arbitrary waveform generator and the optical source, respectively. The experimental setup also is illustrated in Fig. 11.28. Measurements were taken for three different channel conditions, and were also repeated under dark (ambient light power 45 dBm) and bright (ambient light power 18 dBm) environments to ensure that the measurement were not influenced by any undesirable optical signal. Since the results taken under dark and bright room conditions were almost the same only measured data set for the bright room condition are presented, see Table 11.6. The results clearly show the advantage of the proposed method. For example, for 291 Advances in Optics: Reviews. Book Series, Vol. 3 6.06 10 m ⁄ where SISO link did not provide acceptable signal quality (the Q-factor is less than the required value of 4.75), the proposed method results in a Q-factor which is larger than SISO. Besides, compared to a SISO link, the proposed method effectively reduced the variation of detection threshold. The outcome of the experimental result agrees with the deduction in Section 11.6.3 and Section 11.8.1 in the way that in the proposed scheme the channels are highly correlated, therefore a high performance enhancement was expected. To conclude this section, the histogram of the recorded signals for a clear channel as well 6.06 10 m ⁄ are included in Fig. 11.28. In as a turbulent channel for contrast to the previous cases, the combination of Manchester code and differential signalling method results in a signal with a zero offset. 25 25 SISO Threshold Level Turbulence 20 Number of Points Number of Points 20 SISO Threshold Level Clear Channel 15 10 5 15 10 5 0 2853.9932853.9942853.9952853.996 2853.9972853.9982853.999 2854 Threshold Level (mV) 0 2854.001 2508.9 2508.95 2509 2509.05 2509.1 2509.15 2509.2 2509.25 2509.3 Threshold Level (mV) (a) (b) 18 14 12 DS Threshold Level Clear Channel 16 14 Number of Points 10 Number of Points DS Threshold Level Turbulence 8 6 4 12 10 8 6 4 2 0 -3 2 -2 -1 0 1 Threshold Level (mV) (c) 2 3 -3 x 10 0 -4 -3 -2 -1 0 1 Threshold Level (mV) 2 3 4 x 10 -3 (d) Fig. 11.28 (a-d). (a, b) histograms of the detection threshold levels of the single-input single-output link, (c, d) histograms of the detection threshold levels of the differential signalling (DS) link. 292 Chapter 11. Implementing Differential Signalling in Free Space Optical Communication Link 25 18 SISO Q-factor Clear Channel 14 Number of Points Number of Points 20 SISO Q-factor Turbulence 16 15 10 12 10 8 6 4 5 2 0 5 10 15 20 Q 25 30 35 0 0.5 40 1 1.5 2 (e) 3.5 4 DS Q-factor Turbulence 16 14 14 12 10 8 6 12 10 8 6 4 4 2 2 0 38 3 18 DS Q-factor Clear Channel Number of Points Number of Points 16 2.5 (f) 20 18 Q 40 42 44 46 Q 48 50 0 20 25 30 (g) 35 Q 40 45 50 (h) Fig. 11.28 (e-h). (e, f) histograms of the Q-factor of the SISO link, and (g, h) histograms of the Q-factor of the differential signalling link. (a, c, e, f) are for clear conditions whereas (b, d, f, h) are for turbulent channel with Cn2 = 6.06×10-11 m-2/3. Table 11.6. The summery of the measurement for single-input single-output (SISO) differential signalling (DS) link. These results are for DS and Manchester code scheme. SISO Clear 4.25 6.06 a, b ⁄ 10 10 √ DS a Q-factor 0.88 [33.0, 5.3] 40.77 [1.7, 0.6] 7.35 [7.9, 4.1] √ 0.89 0.91 1.09 Q-factorb [45.3, 1.4] [40.3, 2.1] [40.0, 4.1] The pair shows the simulation outcome in form of expected value and standard deviation pair, respectively. 293 Advances in Optics: Reviews. Book Series, Vol. 3 11.10. Summary Using theory, simulation and experiment, the benefits of differential signalling method in turbulence and pointing errors channels were described. Differential signalling has been known as an effective method to mitigate the impact of non-random fading channels (e.g., fog) and cancelling the ambient background noise. In this chapter it was shown that differential signalling also can mitigate the effects of turbulence and pointing errors. In conditions where threshold level of the received signal is varied by turbulence and pointing errors, it was shown that by using the differential signalling method threshold level variation is reduced and the reduction depends on how correlated the channels are. In one experiment the performance of differential signalling under turbulence of 0.17 was evaluated. The measurements showed that for differential signalling in uncorrelated channels condition (i.e., , 0 ) the threshold level had the standard deviation of 43 mV whereas for the same setup under correlated condition (i.e., 0.72 ) the standard deviation reduced to 13 mV . The appropriate conditions of , approaching high correlation were also discussed and it was shown that if the wavelength difference of differential signalling links relative to the central wavelength is less than 0.47, then the channels will have correlation higher than 0.9. In another experiment, the effect of differential signalling under pointing errors fading effect was investigated. The measurement showed that where the standard deviation of threshold level for both links of differential signalling scheme was 25 and 27 mV under the same conditions the threshold level of differential signalling system with correlation coefficient of 0.92 was 5 mV . Finally, an alternative scheme was introduced that in contrary to conventional differential signalling, it only needs one FSO link. The measurements showed that under 6.06 10 m ⁄ using the aforementioned the same turbulence condition of method the Q-factor of the received signal in the new differential signalling scheme was 40 while a SISO link under the same conditions delivers a Q-factor of 1.8. For the future work the following needs investigating: 1. In this chapter, only the performance of differential signalling under a weak turbulence regime was considered. As an extension to the current work, one can look at the behaviour of the FSO system using differential signalling under moderate and strong turbulence regimes. The outcome of such investigation can result in the closed-form expression for BER. Also it will be possible to compare the performance of the differential signalling technique with other existing methods; 2. In this work pointing errors without turbulence was investigated. This work can be extended by investigating the FSO link performance with both pointing errors and turbulence; 3. The differential signalling technique was only studied for a two-level signal (i.e., NRZ-OOK), and showed that in addition to the mean value of the received signal the signal levels were also affected. Therefore, the differential signalling technique should also investigated with higher order modulations such as M-order pulse amplitude modulation ( -PAM). 294 Chapter 11. Implementing Differential Signalling in Free Space Optical Communication Link References [1]. IEEE Standard for Ethernet, IEEE Std 802.3-2015 (Revision of IEEE Std 802.3-2012), March 4, 2016, pp.1-4017. [2]. F. Nadeem, B. Geiger, E. Leitgeb, S. S. Muhammad, M. Loeschnig, G. Kandus, Comparison of link selection algorithms for free space optics/radio frequency hybrid network, IET Communications, Vol. 5, 2011, pp. 2751-2759. [3]. A. Eslami, S. Vangala, H. Pishro-Nik, Hybrid channel codes for efficient FSO/RF communication systems, IEEE Transactions on Communications, Vol. 58, 2010, pp. 2926-2938. [4]. F. Nadeem, E. Leitgeb, M. S. Awan, G. Kandus, FSO/RF hybrid network availability analysis under different weather condition, in Proceedings of the 3rd International Conference on Next Generation Mobile Applications, Services and Technologies (NGMAST'09), 2009, pp. 239-244. [5]. S. Vangala, H. Pishro-Nik, Optimal hybrid RF-wireless optical communication for maximum efficiency and reliability, in Proceedings of the 41st Annual Conference on Information Sciences and Systems (CISS'07), 2007, pp. 684-689. [6]. F. Nadeem, E. Leitgeb, M. S. Awan, G. Kandus, Forward prediction of FSO received signal for FSO/RF hybrid network, in Proceedings of the International Workshop on Satellite and Space Communications (IWSSC’09), 2009, pp. 196-200. [7]. M. A. Khalighi, M. Uysal, Survey on free space optical communication: a communication theory perspective, IEEE Communications Surveys & Tutorials, Vol. 16, 2014, pp. 2231-2258. [8]. F. Nadeem, M. Gebhart, E. Leitgeb, W. Kogler, M. S. Awan, M. S. Khan, et al., Simulations and analysis of bandwidth efficient switch-over between FSO and mmW links, in Proceedings of the 16th International Conference on Software, Telecommunications and Computer Networks (SoftCOM’08), 2008, pp. 356-361. [9]. W. Popoola, Subcarrier intensity modulated free-space optical communication systems, PhD Thesis, School of Computing, Engineering and Information Sciences, University of Northumbria, 2009. [10]. X. Tang, Polarisation shift keying modulated free-space optical communication systems, PhD Thesis, School of Computing, Engineering and Information Sciences, University of Northumbria, 2012. [11]. http://www.itu.int/net/itunews/issues/2011/06/32.aspx [12]. fSONA, http://www.fsona.com/index.php [13]. D. Rodewald, MRV Introduces Industry's First 10G Ethernet Wireless Point-to-Point System, MRV Communications Inc., Chatsworth, CA, USA, 2008. [14]. G. T. Djordjevic, M. I. Petkovic, A. M. Cvetkovic, G. K. Karagiannidis, Mixed RF/FSO relaying with outdated channel state information, IEEE Journal on Selected Areas in Communications, Vol. 33, 2015, pp. 1935-1948. [15]. L. Jeong-Ho, H. Seung-Hoon, Selection diversity-aided subcarrier intensity modulation/spatial modulation for free-space optical communication, IET Optoelectronics, Vol. 9, 2015, pp. 116-124. [16]. Z. Jiayi, D. Linglong, Z. Yu, W. Zhaocheng, Unified performance analysis of mixed radio frequency/free-space optical dual-Hop transmission systems, Journal of Lightwave Technology, Vol. 33, 2015, pp. 2286-2293. [17]. Z. Ghassemlooy, S. Rajbhandari, W. Popoola, Optical Wireless Communications: System and Channel Modelling with MATLAB, Taylor & Francis, Boca Raton, FL, USA, 2013. [18]. H. Willebrand, B. S.Ghuman, Free Space Optics: Enabling Optical Connectivity in Today’s Networks, Sams Publishing, London, 2002. 295 Advances in Optics: Reviews. Book Series, Vol. 3 [19]. A. Sikora, V. F. Groza, Coexistence of IEEE802. 15.4 with other Systems in the 2.4 GHzISM-Band, in Proceedings of the IEEE Instrumentation and Measurement Technology Conference (IMTC’05), 2005, pp. 1786-1791. [20]. L. Rakotondrainibe, Y. Kokar, G. Zaharia, G. El-Zein, 60 GHz high data rate wireless communication system, in Proceedings of the IEEE 69th Vehicular Technology Conference, 2009 (VTC Spring’09), 2009, pp. 1-5. [21]. D. Sinefeld, D. Shayovitz, O. Golani, D. M. Marom, Adaptive rate and bandwidth WDM optical sampling pulse streams with LCoS-based photonic spectral processor, in Proceedings of the Optical Fiber Communication Conference and Exposition and the National Fiber Optic Engineers Conference (OFC/NFOEC’13), 2013, pp. 1-3. [22]. J. I. Saari, M. M. Krause, B. R. Walsh, P. Kambhampati, Terahertz bandwidth all-optical modulation and logic using multiexcitons in semiconductor nanocrystals, Nano Letters, Vol. 13, Issue 2, 2013, pp 722-727. [23]. SMART 2020: Enabling the low carbon economy in the information age, Press Release, The Climate Group on behalf of the Global eSustainability Initiative (GeSI), Brussels, Belgium, Vol. 20, 2008. [24]. A. Chaaban, J. Morvan, M. Alouini, Free-space optical communications: capacity bounds, approximations, and a new sphere-packing perspective, IEEE Transactions on Communications, Vol. 64, 2016, pp. 1176-1191. [25]. M. Bhatnagar, Z. Ghassemlooy, Performance analysis of Gamma-Gamma Fading FSO MIMO links with pointing errors, Journal of Lightwave Technology, Vol. 34, 2016, pp. 2158-2169. [26]. X. Li, N. Bamiedakis, X. Guo, J. McKendry, E. Xie, R. Ferreira, et al., Wireless visible light communications employing feed-forward pre-equalization and PAM-4 modulation, Journal of Lightwave Technology, Vol. 34, 2016, pp. 2049-2055. [27]. A. S. Hamza, J. S. Deogun, D. R. Alexander, Wireless communication in data centers: A survey, IEEE Communications Surveys & Tutorials, Vol. 18, 2016, pp. 1572-1595. [28]. P. Jaedon, L. Eunju, C. Chan-Byoung, Y. Giwan, Outage probability analysis of a coherent FSO amplify-and-forward relaying system, IEEE Photonics Technology Letters, Vol. 27, 2015, pp. 1204-1207. [29]. K. Kikuchi, Fundamentals of coherent optical fiber communications, Journal of Lightwave Technology, Vol. 34, 2016, pp. 157-179. [30]. L. C. Andrews, R. L. Phillips, Laser Beam Propagation Through Random Media, Second Edition, Vol. PM152, SPIE Publications, 2005. [31]. J. Libich, S. Zvanovec, Measurement statistics of three joint wireless optical links, in Proceedings of the International Workshop on Optical Wireless Communications (IWOW’12), 2012, pp. 1-3. [32]. M. Ijaz, Z. Ghassemlooy, J. Pesek, O. Fiser, H. Le Minh, E. Bentley, Modeling of fog and smoke attenuation in free space optical communications link under controlled laboratory conditions, Journal of Lightwave Technology, Vol. 31, 2013, pp. 1720-1726. [33]. M. Ijaz, Experimental characterisation and modelling of atmospheric fog and turbulence in FSO, PhD Thesis, Faculty of Engineering and Environment, University of Northumbria, 2013. [34]. Y. Fan, C. Julian, T. A. Tsiftsis, Free-space optical communication with nonzero boresight pointing errors, IEEE Transactions on Communications, Vol. 62, 2014, pp. 713-725. [35]. M. A. Khalighi, F. Xu, Y. Jaafar, S. Bourennane, Double-laser differential signaling for reducing the effect of background radiation in free-space optical systems, IEEE/OSA Journal of Optical Communications and Networking, Vol. 3, 2011, pp. 145-154. [36]. G. Keiser, Optical Fiber Communications, 4th Ed., McGraw-Hill Companies, New York, 2011. 296 Chapter 11. Implementing Differential Signalling in Free Space Optical Communication Link [37]. http://www.digikey.com.au/Web Export/Supplier Content/RFM_583/PDF/rfm-an-ook-askfsk-comparison.pdf [38]. G. P. Agrawal, Fiber-Optic Communication Systems, 4th Ed., Wiley, New York, 2010. [39]. K. Tsukamoto, A. Hashimoto, Y. Aburakawa, M. Matsumoto, The case for free space, IEEE Microwave Magazine, Vol. 10, 2009, pp. 84-92. [40]. B. E. A. Saleh, M. C. Teich, Fundamentals of Photonics, Wiley & Sons, 1991. [41]. IEC 60825-1:2014, Safety of Laser Products - Part 1: Equipment Classification and Requirements, https://webstore.iec.ch/publication/3587 [42]. Laser Safety Standards, ANSI Z136 Standards, The Laser Institute of America (LIA), https://www.lia.org/store/ANSI+Z136+Standards [43]. Standard: CENELEC - EN 60825-1, Engineering360, http://standards.globalspec.com/std/9861660/cenelec-en-60825-1 [44]. https://www.gov.uk/government/publications/laser-radiation-safety-advice/laser-radiationsafety-advice [45]. P. Ales, Atmospheric effects on availability of free space optics systems, Optical Engineering, Vol. 48, 2009, 066001. [46]. M. Ijaz, Z. Ghassemlooy, J. Perez, V. Brazda, O. Fiser, Enhancing the atmospheric visibility and fog attenuation using a controlled FSO channel, IEEE Photonics Technology Letters, Vol. 25, 2013, pp. 1262-1265. [47]. J. M. Kahn, J. R. Barry, Wireless infrared communications, Proceedings of the IEEE, Vol. 85, 1997, pp. 265-298. [48]. H. E. Nistazakis, A. N. Stassinakis, et al., Performance of quadrature amplitude modulation orthogonal frequency division multiplexing-based free space optical links with non-linear clipping effect over Gamma-Gamma modelled turbulence channels, IET Optoelectronics, Vol. 9, 2015, pp. 269-274. [49]. D. K. Borah, D. G. Voelz, Pointing error effects on free-space optical communication links in the presence of atmospheric turbulence, Journal of Lightwave Technology, Vol. 27, 2009, pp. 3965-3973. [50]. X. Zhu, J. M. Kahn, Free-space optical communication through atmospheric turbulence channels, IEEE Transactions on Communications, Vol. 50, 2002, pp. 1293-1300. [51]. G. R. Osche, Optical Detection Theory for Laser Applications, Wiley, New York, 2002. [52]. C. Zhixiao, Y. Song, W. Tianyi, W. Guohua, W. Shaoling, G. Wanyi, Channel correlation in aperture receiver diversity systems for free-space optical communication, Journal of Optics, Vol. 14, 2012, 125710. [53]. R. L. Fante, Electromagnetic beam propagation in turbulent media, Proceedings of the IEEE, Vol. 63, 1975, pp. 1669-1692. [54]. S. M. Navidpour, M. Uysal, M. Kavehrad, BER performance of free-space optical transmission with spatial diversity, IEEE Transactions on Wireless Communications, Vol. 6, 2007, pp. 2813-2819. [55]. Z. Hajjarian, J. Fadlullah, M. Kavehrad, MIMO free space optical communications in turbid and turbulent atmosphere (Invited paper), Journal of Communications, Vol. 4, 2009, pp. 524-532. [56]. M. A. Khalighi, N. Schwartz, N. Aitamer, S. Bourennane, Fading reduction by aperture averaging and spatial diversity in optical wireless systems, IEEE/OSA Journal of Optical Communications and Networking, Vol. 1, 2009, pp. 580-593. [57]. C. Abou-Rjeily, S. Haddad, Inter-relay cooperation: a new paradigm for enhanced relayassisted FSO communications, IEEE Transactions on Communications, Vol. 62, 2014, pp. 1970-1982. [58]. N. D. Chatzidiamantis, G. K. Karagiannidis, On the distribution of the sum of GammaGamma variates and applications in RF and optical wireless communications, IEEE Transactions on Communications, Vol. 59, 2011, pp. 1298-1308. 297 Advances in Optics: Reviews. Book Series, Vol. 3 [59]. W. Zixiong, Z. Wen-De, F. Songnian, L. Chinlon, Performance comparison of different modulation formats over free-space optical (FSO) turbulence links with space diversity reception technique, IEEE Photonics Journal, Vol. 1, 2009, pp. 277-285. [60]. R. Paudel, Z. Ghassemlooy, H. Le Minh, S. Rajbhandari, E. Leitgeb, Lambertian source modelling of free space optical ground-to-train communications, in Proceedings of the 8th International Symposium on Communication Systems, Networks and Digital Signal Processing (CSNDSP’12), Poznan, Poland, 2012. [61]. M. A. Kashani, M. Uysal, M. Kavehrad, A novel statistical channel model for turbulenceinduced fading in free-space optical systems, Journal of Lightwave Technology, Vol. 33, 2015, pp. 2303-2312. [62]. I. S. Gradshtein, A. Jeffrey, I. M. Ryzhik, Table of Integrals, Series, and Products, 5th Ed., Academic Press, Boston, London, 1994. [63]. I. A. Stegun, M. Abramowitz, Handbook of Mathematical Functions: with Formulas, Graphs and Mathematical Tables, National Bureau of Standards, 1970. [64]. S. Tianyu, K. Pooi-Yuen, A robust GLRT receiver with Implicit channel estimation and automatic threshold adjustment for the free space optical channel with IM/DD, Journal of Lightwave Technology, Vol. 32, 2014, pp. 369-383. [65]. S. Hitam, M. Abdullah, M. Mahdi, H. Harun, A. Sali, M. Fauzi, Impact of increasing threshold level on higher bit rate in free space optical communications, Journal of Optical and Fiber Communications Research, Vol. 6, 2009, pp. 22-34. [66]. M. R. Bhatnagar, Z. Ghassemlooy, Performance evaluation of FSO MIMO links in GammaGamma fading with pointing errors, in Proceedings of the IEEE International Conference on Communications (ICC’15), 2015, pp. 5084-5090. [67]. X. Zhu, J. M. Kahn, Markov chain model in maximum-likelihood sequence detection for freespace optical communication through atmospheric turbulence channels, IEEE Transactions on Communications, Vol. 51, 2003, pp. 509-516. [68]. X. Zhu, J. M. Kahn, Pilot-symbol assisted modulation for correlated turbulent free-space optical channels, Proceedings of SPIE, Vol. 4489, 2002, pp. 138-145. [69]. M. L. B. Riediger, R. Schober, L. Lampe, Fast multiple-symbol detection for free-space optical communications, IEEE Transactions on Communications, Vol. 57, 2009, pp. 1119-1128. [70]. I. S. Ansari, M. S. Alouini, J. Cheng, Ergodic capacity analysis of free-space optical links with nonzero boresight pointing errors, IEEE Transactions on Wireless Communications, Vol. 14, 2015, pp. 4248-4264. [71]. H. Moradi, H. H. Refai, P. G. LoPresti, Spatial diversity for fiber-bundled FSO nodes with limited mobility, Journal of Lightwave Technology, Vol. 30, 2012, pp. 175-183. [72]. K. Yoshida, K. Tanaka, T. Tsujimura, Y. Azuma, Assisted focus adjustment for free space optics system coupling single-mode optical fibers, IEEE Transactions on Industrial Electronics, Vol. 60, 2013, pp. 5306-5314. [73]. A. Leon-Garcia, Probability and Random Processes for Electrical Engineering, AddisonWesley, 1989. [74]. G.976: Test Methods Applicable to Optical Fibre Submarine Cable Systems, ITU-T Recommendation G.976, ITU, 2004. [75]. D. P. Greenwood, Bandwidth specification for adaptive optics systems, Journal of the Optical Society of America, Vol. 67, 1977, pp. 390-393. [76]. Z. Ghassemlooy, H. Le Minh, S. Rajbhandari, J. Perez, M. Ijaz, Performance analysis of ethernet/fast-ethernet free space optical communications in a controlled weak turbulence condition, Journal of Lightwave Technology, Vol. 30, 2012, pp. 2188-2194. [77]. G. Yang, M. A. Khalighi, S. Bourennane, Z. Ghassemlooy, Fading correlation and analytical performance evaluation of the space-diversity free-space optical communications system, Journal of Optics, Vol. 16, 2014, 035403. 298 Chapter 11. Implementing Differential Signalling in Free Space Optical Communication Link [78]. A. A. Farid, S. Hranilovic, Outage capacity optimization for free-space optical links with pointing errors, Journal of Lightwave Technology, Vol. 25, 2007, pp. 1702-1710. [79]. D. K. Borah, D. Voelz, S. Basu, Maximum-likelihood estimation of a laser system pointing parameters by use of return photon counts, Applied Optics, Vol. 45, 2006, pp. 2504-2509. [80]. R. Forster, Manchester encoding: opposing definitions resolved, Engineering Science and Education Journal, Vol. 9, 2000, pp. 278-280. 299 Chapter 12. Fabrications of Holographic Optical Elements in Polycarbonate for Holographic Weapon Sight Application Chapter 12 Fabrications of Holographic Optical Elements in Polycarbonate for Holographic Weapon Sight Application V. Vadivelan and B. Chandar Shekar1 12.1. Introduction A new approach is reported for fabrication of transmission type phase holographic optical elements (HOE) especially for a holographic weapon sight (HWS) for small arms/riffles. Silver Halide (AgH) holographic emulsion has been used for the fabrication of transmission type hologram with reticle image [1-5]. The HWS is having advantages over other types of weapon sight in close quarters combat [6-10] and also it is used as crew optical alignment system [11]. The M/s L-3 EO-tech is pioneer in manufacturing HWS for small arms [12-16] and HWS recent developments and improvements are reported [17-18]. Here, we fabricated two different kinds of HOEs and two different methods for fabrication of HOEs in Polycarbonate [19]. The fabrications of two different types of HOEs are (1) Phase transmission HOE with reticle image and (2) High diffraction efficiency transmission phase holographic collimator. The two different type of fabrications are (1) Fabrication in Silver halide (AgH), conversion into photoresist by contact copying method and transfer to Polycarbonate by electroforming and recombination technique and (2) Direct fabrication of HOEs in photoresist and convert into polycarbonate by electroforming and recombination techniques. We fabricated HOE with reticle virtual image has diffraction efficiency of around 15-20 % in AgH holographic ultra fine grain commercially available photo material but less than 3 % diffraction efficiency of holograms are preferable in HWS due to its higher visible transmission. One of a main drawback of HOEs in AgH is final transmission hologram tends to darken by exposing under ambient light, known as print-out effect. It drastically reduces both the diffraction efficiency and transmission of HOEs. Hence, we V. Vadivelan R&D Department, Ignetta holographic (P) Ltd, Madukkari, Coimbatore, Tamilnadu, India 301 Advances in Optics: Reviews. Book Series, Vol. 3 try to avoid such darkening problem; we fabricated the same kind of HOE in transparent polycarbonate (PC) for HWS application. In our work, we fabricated fairly good diffraction efficiency reticle HOE of nearly 3 % with 80-85 % visible transmission (without antireflection coating) in Polycarbonate for HWS application. First we transferred the hologram from AgH into photoresist (PR) by copying method [20-25]. The crucial and controlled wet chemical process will result a high quality HOE in positive PR. The HOE in PR is once again transferred into PC by using electroforming and recombination techniques [26-29]. The poor transmission of HOE in PR is not suitable for HWS application. Hence, we fabricated HOE in high transmission PC to avoid this specific application. The diffraction efficiency comparison, visible transmission comparison and quality of the output beam of HOE in AgH, PR and PC are analyzed in detail. Our method of fabrication is simple, very fast and accurate in production. This method greatly reduces the cost in terms of mass production. Also we can keep constant diffraction efficiency and transmission for all the fabricated reticle HOE, which is impossible to achieve in conventional AgH emulsion. The HOE in PC can prevent environmental impact; no shatter problem, light weight, easy handling and long life are listed as a few added advantages of HOE in PC. Apart from these regular advantages, in our study we found that AgH transmission holograms have their peak diffraction efficiency by using the same writing wavelength and recorded angle due to Bragg condition [34]. But the holograms transferred into PR and PC shown some interesting results of keep on nearly constant diffraction efficiency with angular variation. Angular response of HOE in PC is much greater compare to the hologram in AgH leads to many applications especially for solar. The HOEs in PC for HWS application is first time reported by author [19]. The experimental method and deep analysis of reticle HOE in PC are reported here. 12.2. Material and Methods The PC received their name [30] because they are polymers containing carbonate groups (−O−(C=O) −O−). The inflexibility and the lack of mobility prevent PC from developing a significant crystalline structure. PC is produced by a polymerization reaction between bisphenol A, a volatile liquid derived from benzene, and phosgene, a highly reactive and toxic gas made by reacting carbon monoxide with chlorine. The resultant polymers (long, multiple-unit molecules) are made up of repeating units containing two aromatic (benzene) rings and connected by ester (CO-O) groups Tetrabromobisphenol A is used to enhance fire resistance. Tetramethylcyclobutanediol has been developed as a replacement for bisphenol A. The study of PC as a substrate for holographic emulsion coating instead of commonly using glass substrate was reported [31]. The PC is one of the strongest and safest materials on the market. Its applications are ranging from feeding bottles to helmet visors of astronauts and for space shuttle windshields. This amorphous nature of the polymer allows the light transmit ability nearly that of glass. Light weight, transparency, excellent toughness, thermal stability, high impact resistance and optical properties makes PC as one of the most widely used engineering thermoplastics [32-34]. There are certain 302 Chapter 12. Fabrications of Holographic Optical Elements in Polycarbonate for Holographic Weapon Sight Application constraints to the use of PC include limited chemical and scratch resistance, photo degradation, birefringence and thermal expansion. However, there are a number of solutions to solve these constraints. The birefringence and thermal expansion are significant problems in PC, but M. Toishi et al, are reported [29] that the PC having great potential in the substrate of a holographic recording medium. The photo degradation process is significantly reduced in the polycarbonate sheet because of the excellent protection offered by the UV co-extruded layer which incorporates UV stabilizers and UV absorbers. Over a prolonged period, a slight yellowing or hazing will be detected in a polycarbonate sheet. He-Ne laser of 25 mW and He-Cd laser of 100 mW are used. A few commercially available holographic materials [35-49] for comparisons are given in Table 12.1. Here, we have used AgH holographic emulsion supplied by M/s Ultimate holography’s ultra fine grain silver halide red sensitive emulsion specially coated for us, which is having spatial resolution of 20,000 lines/mm and the grain size are typical 4 nm. The thickness of the AgH emulsion is 10 µm, its spectral sensitivity is 633 nm and sensitivity is in the range of 150 µJ/cm2 - 200 µJ/cm2. The power of the beam is measured by using power meter from M/s Edmund Optics. The whole experimental setup is placed in vibration free isolation table supplied by M/s Holmarc. The exposed plates are developed in supplier developer of 6 ml mixed with 100 ml de-ionized water for 4 min; the developed HOEs are put into water for 3 min and the same plates were bleached by using R10 bleach [50] until clear. The chemicals combination of bleach formulation are 2 grams of potassium dichromate mixed with 10 ml of hydrochloric acid and added 35 grams of potassium bromide in distilled water to make 1 litre of bleach solution. Isopropyl bath of different concentration level will improve efficiency little more. The recording and chemical processes are done under dark room safe light condition. We have measured the diffraction efficiency of naturally dried HOEs and obtained desired diffraction efficiency. Table 12.1. A comparison of few holographic materials for HOEs. Material/Effect Silver Halide Dichromated gelatin Photopolymers Photoresist LiNbO3 LiTaO3 KNbO3 Sn2P2S6 Photochromics Bacteriorhodopsin gelatin matrix Spectral Range, nm Thickness μm Δn Range ˚C < 1100 Resolving power, lines/mm Up to 10000 7-10 0.02 < 100 < 700 > 5000 15-35 0.022 < 200 514-670 < 450 350-650 300-550 400-900 550-1100 400-700 > 5000 1000 > 2000 > 2000 > 2000 > 2000 > 1600 5-500 1.5-2.4 > 10000 > 10000 > 10000 > 10000 > 100 0.012 2×10-3 10-3 10-4 3×10-4 10-3 < 100 < 500 < 450 > 50 < 66 < 66 520-640 > 1000 30-40 3×10-3 20/40 303 Advances in Optics: Reviews. Book Series, Vol. 3 We have used 3-4 µm thickness blade coated positive PR supplied by M/s Shipley. We have used Microposit developer AZ303 as photoresist developing agent, diluted with deionized water in proportion 1:9. Developing time was 7-10 seconds. It is worth to note that photoresist plates also can be developed with 1.5 % KOH or NaOH solution. However, those chemicals remain in the developed relief and the thin layer of them is transparent. As a result, the holographic image looks great on photoresist, but as soon one will deposit silver on such a relief – silver enters into reaction with KOH or NaOH and the resulting silver relief is much shallower than the photoresist relief. Therefore we stuck with Microposit developer AZ303, which probably has some ingredients preventing developer's remains layer forming. 12.3. Experimental Arrangement 12.3.1. Fabrication Details of Reticle HOE in AgH He-Ne laser wavelength of 632.8 nm is used as a source for recording off- axis transmission type hologram. The unexpanded linear polarized laser beam is divided into two by density variable beam splitter BS, which is helpful to manage desired beam ratio at the holographic recording medium. The mirrors 1 and 2 are directed the laser beams and make them interfere at recording plate with decided angle of interference. The spatial filters SF1 and SF2 are used to expand and spatially clean the beams, the laser beam expanded by SF1 passing through the reticle mask and it is propagate through the imaging lens and referred as object beam, which is finally collected at the recording medium. The mirror 2 is directed the second unexpanded laser beam towards SF2 and it is expanded and spatially cleaned by SF2, which is collimated by collimating lens and referred as reference beam. The reference and object beams are interfered at 56° in AgH emulsion. The transmission hologram fabrication schematic representation is shown in Fig. 12.1. Fig. 12.1. Schematic representation of experimental set up of HOE fabrication. The M/s Uniblitz computer controlled electronic shutter ES is used for precise exposure at the recording plate. We have used ultra fine grain AgH holographic emulsions for this study. AgH supplier developer and bleach used for the fabrication of phase transmission hologram. We have achieved around 28 % diffraction efficiency in this emulsion. But for 304 Chapter 12. Fabrications of Holographic Optical Elements in Polycarbonate for Holographic Weapon Sight Application our study, we used only less than 5 % diffraction efficiency because of its higher transmission in visible spectrum. The final wash of chemically developed phase HOEs and finished HOEs light dispersions are shown in Fig. 12.2. Fig. 12.2. Photographs of Chemically processed HOEs. 12.3.2. Direct Fabrication of HOEs in Photoresist A 442 nm wavelength emitting He-Cd laser source of 100 mW power is used for the fabrication holographic grating recorded directly in photoresist. The exposure of two collimating laser beams interfered at the photoresist is controlled by computer controlled electronic shutter. The interfering angle is decided by two mirrors mentioned in the schematic representation; those are spatially cleaned by spatial filters 1 and 2. The laser beam splitted and intensity ratio is controlled by using variable density beam splitter. The method of fabrication is as same as in Fig. 12.1, only one change made is laser source. It is shown in Fig. 12.3. Fig. 12.3. Schematic representation of holographic reticle fabrication by using He-Cd laser. In Fig. 12.4, the schematic representation for the fabrication of holographic grating is shown. A laser beam is split into two and the beam ratio is adjusted by variable density beam splitter. The two mirrors are used to make desired interference angle at the 305 Advances in Optics: Reviews. Book Series, Vol. 3 holographic plate. Spatial filters 1 and 2 are used to expand and spatially clean the laser beams. Two achromatic doublet collimating lenses are used for collimation. Fig. 12.4. Schematic representation of holographic grating. 12.3.3. Transfer of HOE into PR and PC The HOE fabricated in AgH with 12 % diffraction efficiency is used for transfer process, the image transferred into positive PR by contact copying method by using He-Cd laser of 100 mW power is shown in Fig. 12.5. Fig. 12.5. Schematic representation of copying method. The fabricated HOE in AgH was placed as shown in figure. The gap between the HOE in AgH and HOE to be copy in positive PR is about 1 mm. The primary HOE in AgH diffracts light at an angle and interference takes place between originally separated by a distance. The total power at the recording copy plate is around 6 mW and the exposure time was 14 seconds is the best suitable recording condition for copying HOE in positive PR. When light photons observed by positive PR removed from the surface and unexposed portion will remains at the time of chemical process. Once again the hologram in PR is transferred into transparent PC by using electroforming and recombination technique and the photograph of the electroforming and recombination is shown in Fig. 12.6. The diffraction efficiency of the fabricated HOEs are calculated by general formula, the ratio of the diffracted beam intensity to the incident beam intensity is gives the diffraction efficiency of the HOE. 306 Chapter 12. Fabrications of Holographic Optical Elements in Polycarbonate for Holographic Weapon Sight Application Fig. 12.6. Photographs of electroforming and recombination unit. In electroforming there are two steps followed namely silver spray and electroforming bath. (1) Silver spray is a method of coating a surface with a thin film of pure metallic silver before electroplating. The two different solutions of Silver nitrate 10 grams mixed with Ammonia and Glucose added to Formaldehyde are used to reduce the silver nitrate into pure metallic silver as it falls on the surface of HOE in PR. (2) Silver coated HOE in PR is placed in the tank which is contains nickel sultanate and boric acid added into distilled water. Initially the current level is 5 amp and 5 amp stepwise incremental for every 5 seconds and kept for nearly 60 min. The result is thick layer of deposition of pure nickel on the surface of PR. This is used as a master shim for transfer of HOE into PC by recombination method. By using heat and pressure, the HOE transferred into PC. The study of angular response is very important aspect of this study; it is discussed later. 12.4. Result and Discussion The primary aim of this study is based on fabrication of HOE in PC in order to solve the print-out problem in AgH. We have fabricated phase transmission HOE in AgH holographic emulsion with required 5 % diffraction efficiency and around 90 % transmission at 150 µJ/cm2 exposure dose. It is noted from Fig. 12.7 that the visible transmission of HOE in AgH is comparable to HOE in PC; here we used all the primary visible wavelengths of 632.8 nm, 532 nm, and 442 nm. Higher visible transmission of HOEs is highly recommended for HCS. But the HOE fabricated in AgH have a drawback of became dark under Sun light exposure of few minutes. Here, we try to avoid such a problem by fabricating HOE in PC with the similar effect of HOE in AgH. For this, the HOE in AgH is transferred to positive PR, once again the transferred hologram in PR is replicated into PC by using thin silver coating, electroforming and recombination techniques. We have got the desired optimized diffraction efficiency of around 4 % in PR and around 3 % in PC. The fabricated HOEs in all the three material AgH, PR and PC are shown in Figs. 12.8 (a), (b) and (c) respectively. 307 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 12.7. Visible transmission comparison of all the three holograms at 90°. (a) (b) (c) Fig. 12.8. Fabricated phase transmission holograms in (a) AgH, (b) PR and (c) PC. The interfering angle between two beams is kept 56° and the reconstructed reticle image captured by CCD camera is shown in Fig. 12.9. The diffraction efficiency variation with angle is compared between AgH, PR and PC at three different wavelengths like 632.8 nm, 442 nm and 532 nm. Here it is important to note that holograms in AgH is angular sensitive and its peak diffraction efficiency is obtained at the exact reconstruction angle of 56 degree (Bragg Angle), but it is very interesting and surprise that the HOEs in photoresist and polycarbonate fabricated by 308 Chapter 12. Fabrications of Holographic Optical Elements in Polycarbonate for Holographic Weapon Sight Application copying method from AgH HOE are maintaining almost same diffraction efficiency and it is independent of its angle of incidence. The measurement values of angular variation with diffraction efficiency of AgH, PR and PC are plotted in Fig. 12.10. Fig. 12.9. Photograph of reconstructed reticle image of HOE in AgH. The HOE peak diffraction efficiency of 9.7 % in AgH was obtained in 442 nm at 45°, the same was 4.3 % for both 532 nm at 50° and 632.8 nm at 56°. But considering HOEs in both the PR and PC, the diffraction efficiency was almost has flat response of diffraction efficiency for all incident angles. The diffraction efficiency measuring method is shown in Fig. 12.11. The expanded collimated beam was incident on the hologram in perpendicular direction. The HOEs placed in precise stepper motor control angular rotational stage. The output diffracted beam is focused on power meter by using converging lens. The same setup is used for transmission measurements for all three kinds of holograms. To verify this flat response of diffraction efficiency independent of angular variation, we have fabricated holographic lens with focal length of about 8 cm in AgH and copied into PR. The diffraction efficiency measurement once again confirms the flat diffraction efficiency response and it is independent of angular variation. Hence, Copied HOE from AgH into PC leads to many useful applications. The result gives confidence to fabricate holographic solar concentrators in PC. The microscopic image of reticle HOE in PR and PC confirms the imitation of the same fringe patterns in both holograms and is shown in Fig. 12.12. The HOE in AgH is became fully dark after exposing under direct sunlight for 2 days, the diffraction efficiency reduced about 50 % and laser transmission reduced nearly 70 % but HOE in PC is almost have constant diffraction efficiency and transmission before and after exposure under Sun light for 6 months time. The advantages of PC holograms are easy production, fast and accurate replication, trouble-free handling, preventing from ambient light effect, cost effective, no need of wet chemical process, and better temperature withstand. Maintaining the same diffraction efficiency of all fabricated holograms is impossible in AgH but in PC, we shall keep the same diffraction efficiency for all HOEs to be fabricated. However, PC has few disadvantage, it requires scratch proof coating, UV stabilization need for long time Sun 309 Advances in Optics: Reviews. Book Series, Vol. 3 exposure. The reticle image of the reconstruction of HOE in PC is not sharp as like reticle image in AgH, further work is progressing for the quality improvement. DIFFRACTION EFFICIENCY VARIATION WITH ANGLE AT 632.8 nm 15 10 5 0 0 5 10 15 AgH Hologram 20 25 30 35 40 45 Polycarbonate Holograms 50 55 60 65 70 Photoresist Holograms (a) DIFFRACTION EFFICIENCY VARIATION WITH ANGLE AT 442 nm 20 10 0 0 5 10 15 20 25 30 35 AgH Hologram 40 45 50 55 60 65 70 Polycarbonate Holograms Photoresist Holograms (b) DIFFRACTION EFFICIENCY VARIATION WITH ANGLE AT 532 nm 5 0 0 5 10 15 20 25 30 35 AgH Hologram 40 45 50 55 60 65 70 Polycarbonate Holograms Photoresist Holograms (c) Fig. 12.10. Diffraction efficiency variation with angle at (a) 632.8 nm, (b) 442 nm, and (c) 532 nm. 310 Chapter 12. Fabrications of Holographic Optical Elements in Polycarbonate for Holographic Weapon Sight Application Fig. 12.11. Schematic of diffraction efficiency and transmission measurement set up. (a) (b) Fig. 12.12. Microscopic images of HOE in (a) Photoresist, and (b) Polycarbonate. 12.5. Conclusion We have fabricated desired diffraction efficiency of less than 5 % and visible transmission of more than 90 % in phase transmission HOEs with reticle image in red sensitive ultra fine grain AgH holographic emulsion. To avoid the print-out problem of HOE in AgH, we have transferred HOE from AgH into PR and again in PC, the transfering method was explained in detail. The more attention of the study is that we obtained almost flat response of diffraction efficiency of HOE fabricated in PC for all angle of incidence, which is due to conversion of volume hologram into plane hologram. Here, we compared the diffraction efficiency with angular variation for all the three prime colours of Blue, green and red. We have inspected HOEs in PC and AgH under direct Sun light exposure, The HOE in AgH became completely dark and useless. But HOE in PC was almost having constant value of diffraction efficiency and visible transmission before and after 3 months expousure under direct Sun light. The peculiar result of flat response of diffraction efficiency with variation of angle of PC shows the way to fabrication of HOEs in PC for holographic solar concentration application. Improvement of quality of the image is progessing for HCS and the diffraction efficiency enhancement of HOE in PC for solar applications is our future work plan. Acknowledgment I thank Mr. Thomas Rajan for his encouragement and thanks to Ignetta Holographic pvt Ltd for the use of their laboratory and equipment. 311 Advances in Optics: Reviews. Book Series, Vol. 3 References [1]. M. R. E. La, S. N. Koresheva, M. K. Shevtsov, Optical systems of holographic collimator sights, J. Opt. Technol., Vol. 82, Issue 9, 2015, pp. 592-597. [2]. S. N. Koreshev, G. B. Semenov, Analysis of the intensity of false images formed by volume holographic optical elements, Opt. Spectrosc., Vol. 60, USSR, 1986. [3]. V. K. Salakhutdinov, Holographic Sight, Patent Issue 2034321, Russian Federation, 1995. [4]. S. A. Shodin, Yu. V. Kondakov, Holographic Sight, Patent Issue 2210713, Russian Federation, 2003. [5]. A. M. Tai, E. J. Sieczka, Lightweight Holographic Sight, Patent 6490060, USA, 2002. [6]. S. N. Koreshev, S. S. Kvitko, M. K. Shevtsov, Holographic Collimator Sight, Patent Issue 135426, Russian Federation, 2013. [7]. R. Bhatt, L. Mainali, G. Bhatnagar, A. Sharma, A. K. Gupta, Testing of holographic optical elements for holographic gun-sight, Opt. and Lasers in Eng., Vol. 46, 2008, pp. 217-221. [8]. N. R. Murthy, P. R. Kumar, N. Raghavender, Design and development of holographic sighting system used for small arm weapons in close quarter battle situations, in Proceedings of the 4th IEEE International Conference on Digital Ecosystems and Technologies (DEST’10), 2010, pp. 678-682. [9]. R. Bhatt, L. Mainali, G. Bhatnagar, Development of indigenous holographic gun-sight for small weapons, in Proceedings of the International Conf. on Opt. & Optoelct. (ICOL’05), India, 2005, PP-HDO-3. [10]. P. R. Kumar, N. Raghavender, N. R. Murthy, K. Srimannarayana, M. S. Shankar, R. L. N. Sai Prasad, System Design Aspects of Holographic Weapon Sight, in Proceedings of the 32nd Optical Society of India (OSI) Symposium on Contemporary Optics and Applications, 2007, pp. 164-167. [11]. M. Sarafin, Guidance, Navigation, and Control Section, Crew Optical Alignment Sight, NASA Shuttle Operations Handbook, Vol. 18, NASA, USA, 1998, JSC-18863. [12]. A. Tai, et.al., Compact Holographic Sight, US Patent 5483362, USA, 1996. [13]. J. Upatnieks, Compact holographic sight, Proceedings of SPIE, Vol. 883, 1988, pp. 171-176. [14]. J. Upatnieks, A. Tai, Development of holographic sight, Proceedings of SPIE, Vol. 2968, 1987, pp. 272-281. [15]. A. Tai, E J Sieczka, R. Radler, J. Upatnieks, Holographic gun sights for small arms, Proceedings of SPIE, Vol. 2689, 1996, pp. 210-219. [16]. A. Tai, E. Sieczka, J. Upatnieks, Holographic gun sight – the next generation, Proceedings of SPIE, Vol. 3951, 2000, pp. 84-93. [17]. E. Siecaka, A. Tai, R. Fish, Low Profile Holographic Sight and Method of Manufacturing, US Patent 20060164704 A 1 27, USA, 2006. [18]. T. K. Estes, Holographic Weapon Sight, US Patent D662566, USA, 2012. [19]. V. Vadivelan, B. Chander Shekar, Fabrication of a phase transmission holographic optical element in polycarbonate and its characterization, Applied Optics, Vol. 55, Issue 23, 2016, pp. 6453-6456. [20]. D. Voloschenko, E. B. Arik, F. Dimov, K. Yu, Khin, Swe Yin, N. Rakuljic. Method for Design and Manufacturing of Optics for Holographic Sight, US Patent AG02B532FI, USA, 2015. [21]. G. Saxby, Chapter 21, in Practical Holography, 3rd Edition, 2004, IOP Publishing, pp. 328-332. [22]. P. Hariharan, Copying holograms, Chapter 9, in Basics of Holography, Cambridge University Press, 2002, pp. 78-81. [23]. F. S. Harris, G. C. Sherman, B. H. Billings, Copying holograms, Applied Optics, Vol. 5, Issue 4, 1966, pp. 665-666. [24]. D. B. Brumm, Copying holograms, Applied Optics, Vol. 5, Issue 12, 1966, pp. 1946-1947. 312 Chapter 12. Fabrications of Holographic Optical Elements in Polycarbonate for Holographic Weapon Sight Application [25]. M. J. Landry, The effect of two hologram copying parameters on the quality of copies, Applied Optics, Vol. 6, Issue 11, 1967, pp. 1947-1956. [26]. C. Puech, Copying holograms on photoresist, Applied Optics, Vol. 4, Issue 4, 1971, pp. 279-282. [27]. F. Iwata, J. Tsujiuchi, Characteristics of a photoresist hologram and its replica, Applied Optics, Vol. 13, 1994, pp. 1327-1336. [28]. G. Stojanoff, Review of the technology for the manufacturing of large format DCG-holo grams for technical applications, Proceedings of SPIE, Vol. 3011, 1997, pp. 267-278. [29]. M. T. Gale, Replication techniques for diffractive optical elements, Microelectron. Eng., Vol. 34, Issue 3-4, 1997, pp. 321-339. [30]. F. Iwata, K. Ohnuma, Brightness and contrast of a surface relief rainbow hologram for an embossing master, Proceedings of SPIE, Vol. 523, 1985, pp. 15-17. [31]. A. Fustin, C. Bailly, G. J. Clarkson, P. De Groot, T. H. Galow, D. A. Leigh, D. Robertson, Slawiin, A. M. Zoya, J. K. Wong, Mechanically linked polycarbonate, J. Am. Chem. Soc., Vol. 125, Issue 8, 2003, pp. 2200-2207. [32]. M. Toishi,, T. Tanaka, A. Fukumoto, M. Sugiki, K. Watanabe, Evaluation of polycarbonate substrate hologram recording medium regarding implication of birefringence and thermal expansion, Opt. Comm., Vol. 270, Issue 1, 2001, pp. 17-24. [33]. S. Chakraborty, S. Roy, Structural, dynamical, and thermo-dynamical properties of carbon nano tube polycarbonate composites: A molecular dynamics study, J. Phys. Chem. B, Vol. 116, Issue 10, 2012, pp. 3083-3091. [34]. W. Thaithae, C. Antonio, P. Wattanachai, Properties characterisation of polycarbonate and multi-walled carbon nanotubes composites prepared by solution technique, Asia-Pacific Journal of Chemical Engineering, Vol. 11, 2016, pp. 34-50. [35]. J. F. Lomax, J. J. Fontanella, C. A. Edmondson, M. C. Wintersgill, M. A. Wolak, M. A. Westgate, E. A. Lomax, P. Q. Lomax, X. Bogle, A. Rua, S. G. Greenbaum, Properties of Polycarbonate containing BaTio3 nanoparticles, J. Appl. Phys., Vol. 115, 2014, 104103. [36]. Y. N. Denisyuk, Photographic reconstruction of the optical properties of an object in its own scattered radiation field, Sov. Phys. Dokl., Vol. 7, Issue 158, 1962, pp. 543-545. [37]. P. B. Palacios, S. A. Alvarez, J. M. Saez, M. V. Collados, D. Chemisana, J. Atencia Broadband behavior of transmission volume holographic optical elements for solar concentration, Optics Express, Vol. 23, Issue 11, 2015, pp. A672-A681. [38]. A. Fimia, L. Carretero, A. BelIndez, Holographic optical elements recorded on spherical surfaces with photopolymers, Applied Optics, Vol. 33, Issue 17, 1994, pp. 3633-3634. [39]. A. Fimia, N. López, F. Mateos, R. Sastre, J. Pineda, F. Amat-Guerri, A new photopolymer useful as holographic recording material, Applied Optics, Vol. 32, 1993, pp. 3706-3707. [40]. V. Tigran, J. Galstyan, F. Viens, A. Villeneuve, K. Richardson, M. A. Duguay, Photoinduced self-developing relief gratings in thin film chalcogenide As2S3 glasses, Journal of Light Wave Technology, Vol. 15, Issue 8, 1997, pp. 1343-1347. [41]. G. Stojanoff, H. Schutte, P. Froning, Evaluation of holographic materials from applications point of view, Proceedings of SPIE, Vol. 3294, 1998, pp. 2-13. [42]. T. K. Gaylord, M. G. Moharam, Analysis and applications of optical diffraction by gratings, Proceedings of IEEE, Vol. 73, Issue 5, 1985, pp. 894-937. [43]. J. N. Cederquist, J. R. Fienup, Analytic design of optimum holographic optical elements, J. Opt. Soc. Am. A, Vol. 4, 1987, pp. 699-705. [44]. E. Hasman, A. A. Friesem, Analytic optimization for holographic optical elements, J. Opt. Soc. Am. A, Vol. 6, 1989, pp. 62-72. [45]. T. S. Yu, Effect of emulsion thickness variations on wavefront reconstruction, Applied Optics, Vol. 10, 1971, pp. 1324-1328. [46]. R. D. Rallison, Incoherent multifocus hololens design and fabrication, Proceedings of SPIE, Vol. 1183, 1989, pp. 663-668. 313 Advances in Optics: Reviews. Book Series, Vol. 3 [47]. M. Chang, N. George, Holographic dielectric grating: theory and practice, Applied Optics, Vol. 9, 1970, pp. 713-719. [48]. T. K. Gaylord, F. K. Tittel, Angular selectivity of lithium niobate volume holograms, J. App. Phys., Vol. 44, 1973, pp. 4771-4773. [49]. J. N. Latta, Computer-based analysis of hologram imagery and aberrations, Applied Optics, Vol. 10, 1971, pp. 599-618. [50]. D. H. Mcmahon, A. R. Franklin, Efficient, high quality, R-10 bleached holographic diffraction gratings, Applied Optics, Vol. 8, Issue 9, 1969, pp. 1927-1929. 314 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators Chapter 13 Optical Methods for the Characterization of PV Solar Concentrators Antonio Parretta1 13.1. Introduction The concentration of solar radiation plays a key role in the field of renewable energies, as it can be effectively applied to thermal, thermodynamic, photovoltaic (PV) and even hybrid thermal/photovoltaic technologies [1-10]. In concentrating photovoltaic systems (CPV) the size of the photovoltaic receiver (solar cell) is reduced by a factor equal to the geometric concentration ratio, and this has a strong, positive impact on the cost of the total PV concentrator, opening perspectives for the use of more sophisticated and more efficient devices. The concentrating optics is one specific component of the photovoltaic concentrator. It must be designed to transfer the incident solar radiation to the receiver searching the maximal optical efficiency achievable within an angular range limited by physical constrains [11]. The concentrating optics should produce a concentrated flux with reduced non-uniformity on the receiver to minimize ohmic losses [12, 15], and should be designed with great attention on many aspects related to its final industrial application. These are: compactness, tolerance on assembling errors, low cost of manufacturing processes, optimal placement of the receiver for electrical and thermal issues, use of materials of high reliability, high durability and low cost, high efficiency at the module and array level. All the previously indicated characteristics must be considered in the design of photovoltaic concentrators; many optical configurations have been proposed during the last years [15]; a large spectrum of possible designs, with different levels of effectiveness, can be achieved by applying the “nonimaging” optics [16-20]. The fundamental quantities of a PV solar concentrating optics usually considered are, from one side, the geometric concentration ratio and the optical efficiency, giving the optical concentration ratio, and, from the other side, the spatial and angular flux distribution on the receiver. These quantities are defined based on the irradiation conditions that is on the angle-resolved radiance of the light source. At mid-high Antonio Parretta Physics and Earth Science Department, University of Ferrara, Italy 315 Advances in Optics: Reviews. Book Series, Vol. 3 concentration ratios that is at collecting angles of the order of few degree or less, only the direct component of solar radiation is effective, and the input flux can be approximated by a parallel beam of known spectral irradiance and direction. At low concentration ratios, it is useful to consider also the contribution from a portion of diffuse light, which can be modeled as a lambertian source with defined angular divergence. We have in this way introduced the concept of “optical characterization” of a solar concentrator [21-45]. Different approaches can be followed to perform it; here we will focus our attention on two of them, directly derived by our recent research on this subject: the “direct method” and the “inverse method”, distinguished by the modality in which the concentrator is irradiated, if from the input or the output aperture, respectively. In the “direct method” [21-35], the angle-resolved transmission efficiency is obtained irradiating the input aperture by a suitably oriented parallel beam, of known irradiance, and measuring the output flux; this must be repeated for all the significant directions of incidence, which are strictly dependent on the geometrical symmetry of the concentrator. From the transmission efficiency curve obtained for the different azimuthal directions the “acceptance angle” is derived; it is a parameter that defines the angular limit within which the incident radiation is collected. An alternative way to obtain the angle-resolved optical efficiency is the “inverse method” [36-45], a very effective method where the concentrator can be tested irradiating it from the output aperture, therefore reversing the light path which occurs during the normal operating conditions. It is characterized by a remarkable rapidity of measurements and by a very simple apparatus with respect to the direct method. The main features of this method are illustrated throughout the course of this work and compared with another inverse method, derived from a modification of the original one [36, 45]. The direct method and the two inverse methods are here applied to recently developed nonimaging photovoltaic concentrators of reflective and refractive optics nature. The reflective SCs were of the type CPC (Compound Parabolic Concentrator), developed by Winston [17, 18, 20] starting from his pioneering work [46]. As already discussed, solar concentrators (SC) are usually investigated to know their optical transmission properties when they are irradiated by a uniform, quasi-collimated light beam simulating the direct component of the solar radiation. The most important result of this study is the curve of optical transmission efficiency drawn as function of the angle of incidence of the collimated beam respect to the optical axis of the concentrator. For nonimaging CPC SCs, it has the aspect shown in Fig. 13.1. The transmission curve is 50 characterized by an acceptance angle,  acc , corresponding to 50 % of the efficiency measured at 0°, valid for generic applications, whereas, for photovoltaic (PV) 90 applications, an acceptance angle,  acc , corresponding to the 90 % of the efficiency measured at 0°, is usually adopted. The other important properties which are largely investigated in a solar concentrator are the spatial and angular distribution of the flux on the receiver, of minor importance in thermodynamic solar concentrators, but of crucial importance in photovoltaic solar concentrators [12-15, 47]. The optical transmission curve determines how efficient is the optical transmission and how accurate must be the pointing of the solar tracker, to keep always the efficiency on the top of the curve. The spatial and angular flux density distributions establish if the system is suitable for a photovoltaic receiver, or if a secondary optical element (SOE) must be added to it [48, 49]. These are 316 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators the basic information which are generally pursued, both theoretically and experimentally, when working with photovoltaic solar concentrators. Fig. 13.1. Typical optical transmission curve of a nonimaging solar concentrator. In the next section, we will discuss about theoretical models of irradiation of the solar concentrator; these models are just simplifications of the real irradiation conditions which can be applied in practice with indoor experiments or outdoor expositions. For each model, we derive a specific method of characterization of the SC that can be applied by optical simulations at a computer or by experimental measurements. The acronym assigned to each model of irradiation is the same of that assigned to the corresponding characterization method. We distinguish, for example, between “direct irradiation” and “inverse irradiation” of the SC, depending on the direction of the incoming light, or between “local irradiation” and “integral irradiation”, depending if the irradiation is limited to a small area of the SC aperture, or if it is extended to the entire aperture area; we finally distinguish between a “quasi-collimated irradiation” by a far light source, in contrast to a “diffuse irradiation” by a lambertian source. In the last case, we speak about a “lambertian irradiation” that is an irradiation with constant radiance from all directions within a maximum value of solid angle. In the theoretical section of this work, we focus our attention to generic 3D solar concentrators, regardless if refractive or reflective, imaging or nonimaging, and they are studied as generic optical components for which reflection, absorption or transmission properties are defined respect to specific models of irradiation. In the further sections, instead, we discuss the practical application of the characterization methods, introducing real prototypes of nonimaging, 3D photovoltaic solar concentrators of the type CPC (Compound Parabolic Concentrator), which were realized and extensively studied at Ferrara University and other refractive PV concentrators realized at ENEA labs. In what follows, a generic PV SC is schematized as a device confined between an entrance aperture (ia) with area Ain and an exit aperture (oa) with area Aout, where Ain > Aout, as the definition of solar concentrator requires. A solar concentrator operates, in practice, under 317 Advances in Optics: Reviews. Book Series, Vol. 3 “direct irradiation”, that is under irradiation directed to the entrance aperture (ia) and with a receiver, the energy conversion device (the solar cell), at the exit aperture. In our models, however, when required, we replace the receiver by any detector suitable to measure the total output flux, or its spatial and angular distribution; we also use the exit aperture to put there any source of light for inverse irradiation. The same considerations are valid when we consider the input aperture of the SC; we can measure the total flux exiting from it, or its angular distribution, when operating in reverse direction, and we imagine also to use the entrance aperture to put there any source of light for direct irradiation. What is there between the two apertures depends on the specific fabrication technology, and will not be considered, discussing the theoretical models, because not relevant for a general discussion on its overall optical properties. The main question relative to the operation of a solar concentrator is its ability to transfer light to the output (here with light we intend the full spectrum of the sunlight or any portion of it). The simplest question to ask is: how an elementary beam, incident on (ia) at point P(x, y) from (, ) direction, is transmitted by the concentrator? This question introduces the first and simplest method of characterization of the solar concentrator: the “Direct Local Collimated Method” (DLCM) [31, 32]. To apply this method in the most general form we should consider also the polarization of the beam and its spectrum. In what follows, however, we simplify our discussion by considering always unpolarized and monochromatic light at input. The role played by unpolarized light, in fact, has a significant importance in this work. On the one hand, a solar concentrator works mainly with direct sunlight, which is strictly unpolarized, on the other one, in the following, all the presented methods of SC characterization require the use of unpolarized light. With DLCM irradiation, the elemental collimated beam impinging on the elementary area dAin of the entrance aperture is transmitted to output with an efficiency expressed by the quantity  ( P , dA ,  ,  ) , the local optical transmission efficiency. The beam can be dir in totally reflected backwards, or totally absorbed inside the concentrator: these are extreme cases in which we cannot draw a path for light from the entrance to the exit aperture or vice versa. In all the other cases, we can follow the beam from one aperture to the opposite one. We distinguish therefore between “connecting” and “not connecting” paths, when the paths connect or not the two apertures, respectively. The attenuation that an elementary beam experiments inside the SC is the result of all the interactions with the surfaces and the interfaces met during its travel. In the hypothesis that the beam undergoes only reversible processes [50], as reflections and/or refractions at planar surfaces, excluding surface diffusion or diffraction phenomena, the total attenuation of the beam can be derived by applying repeatedly the Fresnel equations. If  the incidence angle and ’ the transmission angle (by reflection or refraction), it can be found by Eqs. (13.1a) and (13.1b) that the transmission factors Trefl for reflection and Trefr for refraction do not change at exchanging  and ’ angles, that is inverting the direction of travel of the light path, establishing in this way a “reversibility principle”: “the attenuation undergone by an unpolarized beam on the same path, but at opposite direction, is the same” [51]. Trefl  318  cos 2 (   ' )  cos 2 (   ' )  1  sin 2 (   ' )   , 2 2 2  sin (   ' )  cos (   ' )  (13.1a) Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators   1  cos2 (  ' ) Trefr  2  sin  sin' cos  cos' 2 . 2  sin (  ' )  cos (  ' )  (13.1b) The identical connecting paths, as A→B and B→A between the two apertures of the SC, show the same transmission factor TAB = TBA when the starting beam is unpolarized. When the methods described in the following theoretical section are applied, the condition of depolarization of incident beam must be accurately satisfied, both for the “direct sources” and the “inverse sources”. If the irradiation of the SC by a collimated beam is extended to the entire area of input aperture, we talk about the “Direct Collimated Method” (DCM). Fig. 13.2 illustrates the basic scheme of DCM, where Edir is the input cross-section irradiance, in and out the corresponding input and output flux. Fig. 13.2. Basic scheme of the Direct Collimated Method (DCM). The condition TAB = TBA is the basis of the so called “Inverse Lambertian Method” (ILM) [36-45], initially named ILLUME (Inverse ILLUmination MEthod) and known as P-Method (Parretta-Method), to distinguish it from the PH-Method (Parretta-Herrero Method), which will be discussed later in this section. The ILM has been conceived for deriving the absolute transmission efficiency of DCM by analyzing, instead of the flux collected at the receiver (the output aperture) with direct irradiation, the flux collected at input aperture with the inverse irradiation. To apply this concept, it is necessary that the rays analyzed with the “direct irradiation” overlap those analyzed with the “inverse irradiation”, that is, that the respective optical paths be identical. Now, in the direct irradiation by DCM, the input beam should be varied in the 0°-90° range of polar angle. To deduce, therefore, the attenuation undergone by the direct rays inclined at any polar angle respect to the optical axis, it is necessary to analyze all the inverse rays emitted by the concentrator in any direction from the input aperture. The source of the inverse rays must be placed in correspondence of the receiver (the output aperture) and must be able to emit rays, from each point and in any direction inside the SC, at constant radiance, to not discriminate any direction. Only in this way it will be possible to produce, in the reverse mode, all the connecting paths which will overlap with those that are produced in direct mode by a collimated beam inclined at different polar 319 Advances in Optics: Reviews. Book Series, Vol. 3 angles between 0° and 90°. To apply the “inverse method” ILM in a correct way, therefore, we need to put a spatially uniform lambertian source at the output aperture, as shown in Fig. 13.3, where Linv is the constant radiance of the inverse lambertian source. Fig. 13.3. Scheme of the “Inverse Lambertian Method” (ILM). The limit polar angle, at exit aperture (oa), is m = 90°. As we will see, the direct transmission efficiency,  dir ( , ) , is obtained by ILM from the radiance of the inverse light, Ldir ( ,) . If we are interested in knowing the efficiency of light transmission from the input opening to a specific receiver area, dAin or Ain , around point P, we can apply the ILM method to this area, obtaining the quantity dir ( P, dAin , , ) or dir (P, Ain , , ) , introducing in this way the “Inverse Local Lambertian Method” (ILLM). Recent developments of the “inverse lambertian method” have been proposed by Herrero et al. [52, 53]. They modified the P-Method with the aim to test the real optical properties of a photovoltaic solar concentrator as a whole, that is as optical unit + receiver (the solar cell). In this approach, the electroluminescence (EL) light emitted by the forward biased solar cell (the receiver of concentrator) acts as reverse light. Here, the “generalized” Kirchhoff’s law must be applied, which was derived by Wurfel [54] by applying a thermodynamic treatment to both thermal and non-thermal radiation, and which is based on the concept of the chemical potential of the radiation. From the generalized Kirchhoff's law, the solar cell contributes to the reverse light with the same efficiency with which contributes to the absorption of light under direct irradiation [55]. The second change made by Herrero et al. to the P-Method was the use of a parabolic mirror to focus the EL light on the lambertian screen (see Fig. 13.4). This choice allows to get a linear polar distribution of the transmission efficiency directly on the screen, avoiding the need to keep the screen far from the concentrator. The peculiarity of the “luminescence” method is that it operates with “real” receivers (the solar cells), not the “ideal” lambertian ones (with unitary absorptivity); while this is a good feature for the characterization of specific CPV optics, it limits the method at the photovoltaic solar concentrators. To have general methods for testing any type of solar 320 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators concentrator, independently from the nature of receiver (photovoltaic or thermodynamic), a source of lambertian properties for the reverse light is required. The use of a parabolic mirror to focus the reverse light on the screen is generally useful, so we have exploited the idea of adding a parabolic mirror to the optical path of reverse light in the original inverse method (ILM). The result is a method referred to as the “Parretta-Herrero method” (PH-Method), to keep memory of the two contributions [37, 45]. The schematic of PH-Method is reported in Fig. 13.4. Here, the raytracing of rays emerging parallel to the axis of the concentrator and focused on the origin of the x/y frame fixed on the screen is shown. The solar concentrator (sc) is not visible because it is much smaller of the parabolic mirror (pm) and of the planar screen (ps). In Fig. 13.4 the polar diagram built on the screen (ps) is also shown. The point of coordinates ( , ) is the target of any ray exiting from the concentrator at the same polar and azimuthal angles, independently from the starting position from the input aperture. In this way, it is solved the problem of angular resolution that ILM suffers when the planar screen on which is projected the inverse light is not far enough (see Section 13.5.2). Fig. 13.4. Schematic principle of the Parretta-Herrero (PH) method used for simulating the optical properties of a generic solar concentrator. Inverse rays exiting from different points of the solar concentrator (sc) input aperture at the same polar angle  and azimuthal angle , converge, thanks to the parabolic mirror (pm) on the same point on the screen (ps). If we want to analyze the SC, activating simultaneously all the connecting paths in direct mode, we should consider an infinity of beams impinging on the input aperture at different polar angles. This is achieved as well by using a spatially uniform lambertian source placed at the input aperture, as schematized in Fig. 13.5, where Ldir is the constant radiance of the direct lambertian source. We introduce in this way the “Direct Lambertian Method” (DLM). The DLM can be applied to simulate the SC when it is entirely irradiated by a diffused light. The DLM operates with Lambertian light with a divergence of 90°. If we reduce this divergence to an angle m, we talk about the “Direct Lambertian 321 Advances in Optics: Reviews. Book Series, Vol. 3 Methodm)”, DLM (m) (see Fig. 13.6). DLM (m) can be applied to simulate the SC partially irradiated by diffused light. If m is the angular divergence of the direct solar radiation, m = S = 0.27°, then DLM (S) simulates in a correct way the SC irradiated by the quasi-collimated direct component of solar radiation. The DLM (m) can simulate the irradiation of a SC by a near lambertian light source, as illustrated in Fig. 13.7. Fig. 13.5. Scheme of the “Direct Lambertian Method” (DLM). The limit polar angle, at input aperture (ia), is m = 90°. Fig. 13.6. Scheme of the “Direct Lambertian Method (m)”, DLM (m). Fig. 13.7. Irradiation of the SC by a nearby lambertian source (ls). 322 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators If the concentrator is irradiated simultaneously in “direct” and “inverse” modes by lambertian sources, we talk about a mixed Lambertian irradiation and we can introduce the “Mixed Lambertian Method” (MLM) (see Fig. 13.8). In MLM all the connecting paths will overlap and, putting Ldir = Linv , also the elementary flux flowing through any connecting path will be the same along the two directions. Then, with Ldir = Linv , also the total flux flowing through the concentrator from one aperture to the other will be the same in the two directions. The application of MLM allows to look at the SC as a passive optical component with optical features which can be find an equivalence in the traditional electrical quantities. Fig. 13.8. Scheme of the irradiation of the SC by both “Direct Lambertian Method” (DLM) with radiance Ldir and “Inverse Lambertian Method” (ILM) with radiance Linv. 13.2. Theoretical Aspects of SC Irradiation and Definition of New Optical Quantities 13.2.1. Direct Collimated Irradiation An elementary beam, incident on the point P of (ia) and flowing inside the SC in the direct , ,  1. Alternatively, mode, will be transmitted to the output with an efficiency the elementary beam will be reflected backwards or totally absorbed inside SC. In practice, the DLCM is easily applied by utilizing a laser beam as it illustrated in [26, 31, 32, 42]. The DLCM method can be applied to a finite Ain area of input aperture; in this case we obtain the efficiency of direct transmission to this area dir ( P, dAin , , ) . If  dir ( P, ,  ) is averaged over a uniform distribution of points P on the input aperture, the transmission efficiency  dir ( , ) at collimated light of the SC can be approximately estimated. The precise estimation of  dir ( , ) , however, requires the full irradiation of input aperture by a collimated and uniform light beam, and this is obtained by the application of the “Direct Collimated Method” (DCM), the most appropriate method to simulate the behavior of a SC operating under the direct solar irradiation. As we have seen in the Introduction, the most important quantity summarizing the properties of light 323 Advances in Optics: Reviews. Book Series, Vol. 3 collection of a solar concentrator (SC) is its “absolute” transmission efficiency  dir ( , ) , expressed as function of the polar and azimuthal angles of direction of the collimated beam, characterized by a constant irradiance Edir on the wave front (see Fig. 13.1):  dir ( , )   out ( , )  out ( , )  ,  in ( , ) Edir  Ain ( , ) (13.2) where Ain ( ,  ) is the area of input aperture projected along (, ) direction. If the contour of input aperture is contained on a plane surface, Eq. (13.2) simplifies as: dir ( , )  out ( , ) . Edir  Ain  cos (13.3) The “absolute” transmission efficiency dir ( ,) can be expressed as:  dir ( ,  )   dir (0)  dir , rel ( ,  ) . (13.4) where  dir ,rel ( , ) is the “relative” transmission efficiency of the SC and  dir (0) is the transmission efficiency at 0° (see Fig. 13.1). It is clear that dir ( , ) is the average value of  dir ( P , ,  ) , the local, collimated optical efficiency, when  dir ( P, , ) is calculated for all the points of the entrance aperture. We have therefore for the output flux:  out ( , )   dS  E dir  cos  dir ( P, , )  ... Ain  ...  Edir  cos  dS   dir ( P, , )  Edir  cos  Ain  dir ( P, , ) , (13.5) Ain and for the transmission efficiency:  Edir  cos   dS  dir ( P, , )  dir ( , )  ...   Ain Edir  Ain  cos   ... dS  dir ( P, , ) Ain Ain (13.6)   dir ( P, , ). The basic scheme of the “direct collimated method” (DCM) is shown in Fig. 13.2. The representation of a perfectly parallel beam, as in Fig. 13.2, is purely ideal and cannot be achieved in practical experiments. The flux at input can be written in fact as: in  Edir  Ain  cos  Ldir ( ,)   Ain  cos . 324 (13.7) Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators To have a perfectly parallel beam, we should have  = 0, that is Ldir ( ,) should be infinite to have a finite flux at input, and this is impossible to reach in practice. In an ideal experiment, we could imagine collimating light from a dimensionless source placed in the focus of a parabolic mirror, or to collect light from a source of finite dimension placed at infinite distance: in both cases the radiance of the light source becomes infinite. So, we can write more precisely for the direct optical efficiency under a collimated beam: dir ( , )  d out ( , ) d ( , )   ... d in ( , ) din ( , ) lim d ( , ) 1  ...  , Ldir ( , )  Ain  cos d  0 d (13.8) where Ldir ( ,) is the radiance, from ( ,  ) direction, of a finite source at finite distance and the symbol  means “transmitted”. To explore the full properties of light collection of the SC, the collimated beam must be oriented respect to the optical (z) axis of concentrator varying  in the 0°-90° interval and  in the 0°-360° interval. If the SC has cylindrical symmetry, it is sufficient to fix a  value and to vary only . Generally, the SCs have squared or hexagonal input apertures, because these geometries allow to pack better them in a concentrating module, then the  angle can be limited, in these cases, to the 0°-90° or the 0°-60° interval, respectively. Despite this limitation, however, the number of measurements required by the application of DCM is very high, both for simulations and for experimental measurements. This is indeed the very strong limit of DCM applied to the determination of dir ( ,) . This limit can be overcome by using the “Inverse Lambertian Method” (ILM) of irradiation, as it will be demonstrated in the following section. Dealing with “nonimaging” SCs [16-20], whose transmission curve has a step-like profile (see Fig. 13.1), their characterization by DCM can be simplified; it is sufficient in fact to vary the input  angle from 0° to a little more than the acceptance angle at 50 % of 0° 50 50 efficiency,  acc . Rays incident at    acc , in fact, will be rejected back by the SC before reaching the output aperture. The quantity  dir ( , ) (see Fig. 13.1) represents the fraction of flux transferred to the output, and then it represents the “direct transmittance” at collimated light, or “direct collimated transmittance” of the SC, when it is viewed as a generic optical component. In like manner, we can speak of a “direct collimated reflectance”  dir ( ,  ) or of a “direct collimated absorptance”  dir ( , ) of the SC for the fraction back reflected or absorbed of the input flux, respectively: 325 Advances in Optics: Reviews. Book Series, Vol. 3  dir ( , )  d  ( , )  ... d in ( , ) lim d  ( , ) 1  ...  , Ldir ( ,  )  Ain  cos d  0 d  dir ( , )  d ( , )  ... d in ( , ) lim d ( ,  ) 1  ...  . Ldir ( , )  Ain  cos d  0 d (13.9) (13.10) We have for the conservation of energy:  dir ( , )   dir ( , )   dir ( , )  1 . (13.11) We observe that, while the fraction of transmitted flux at output is generally measured, the reflected or absorbed ones are not; only the total lost flux is deduced from Eq. (13.11). The measure of  dir ( , ) and  dir ( ,  ) is however possible and is a simple task by simulation. A typical curve of dir ( ) for a 3-D nonimaging concentrator like a CPC (Compound Parabolic Concentrator) is illustrated in Fig. 13.1 [16-20]. Here the  angle is not represented as the CPC has a cylindrical symmetry. We distinguish the 0° efficiency 50 dir (0) , the acceptance angle at 50 % of 0° efficiency  acc and the relative transmission 50 angle, as it can be easily demonstrated. curve  dir , rel ( ,  ) , characterized by the same  acc Respect to solar concentrators like the nonimaging CPCs, the “imaging” solar concentrators show a very different transmission curve, with a long tail and a short flat portion at small angles [17]. For these concentrators, the DCM must be applied by varying 50 . The transmission the polar angles from 0° to a limit  angle, m , well higher than  acc curves dir ( ) or  dir , rel ( ) are defined for a perfectly collimated light and can be drawn easily by simulations with an optical code. With experimental measurements, the beam will be quasi-collimated, with a divergence angle which can be easily controlled (see Section 13.3). In addition to transferring light from the entrance to the output, a concentrator also concentrates the light. The two functions, transfer and concentration, are closely linked, because there can be no concentration if there is no good light transfer, and the transfer of light is not significant in the absence of concentration. As we have seen, a 3D solar concentrator is characterized by an input aperture of area Ain , an output aperture of area Aout , then by a “geometric concentration ratio” given by: 326 Cgeo Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators Cgeo  Ain / Aout . (13.12) Cgeo cannot be chosen arbitrarily big, for example by reducing arbitrarily Aout , because that shouldn’t give any benefit in terms of the optical concentration, expressed by the dir “optical concentration ratio” Copt : dir Copt  Eout / Ein , (13.13) where Ein is the flux density (irradiance) at input and E out is the “average” irradiance at output of the concentrator. As we shall see soon, the maximum value of Cgeo depends on the geometry of the light source. To find the expression for the optical concentration ratio we start from the definition of the generalized Étendue, applied to 3D concentrators, which establishes the invariant quantity: n2  dx dy dL dM  const, (13.14) where L, M are the cosines directors of light rays respect to the x, y axes of the reference frame, and n is the index of refraction. Eq. (13.14) expresses the Liouville theorem [17] establishing the invariance of the volume occupied by the system in the phase space. By applying this invariance to the input and output apertures of the concentrators (see Fig. 13.9), we have, respectively: n 2  dx  dy  dL  dM  n'2  dx'  dy'  dL'  dM '. (13.15) Fig. 13.9. Irradiation of a generic optical system. The invariance expressed by Eq. (13.15) implies of course that there are no optical losses for the light beam during its path from the entrance to the exit. Integrating Eq. (13.15) we have: n 2  Ain  sin 2  in  n'2  Aout  sin 2  out , (13.16) 327 Advances in Optics: Reviews. Book Series, Vol. 3 Ain n '2  sin 2  out .  2 Aout n  sin 2  in (13.17) Eq. (13.17) represents the geometric concentration ratio C geo , but, having assumed the absence of optical losses inside the concentrator, it also represents the optical concentration ratio C opt , or better, the maximum optical concentration ratio achievable (see Appendix 13.A). We have in fact: Copt  Eout  out Ain out Ain     . Ein Aout in in Aout (13.18) The ratio  out  in represents the well-known “optical transmission efficiency” opt :  opt   out ,  in (13.19) that we have assumed equal to 1 when is valid the Eq. (13.15). We have therefore that, in the absence of optical losses, Eq. (13.17) expresses the ratio of both geometric and optical concentration. We can then write, when opt = 1, that: C opt  Ain n'2  sin 2  out .  2 Aout n  sin 2  in (13.17a) In general, however, opt < 1, and then Eq. (13.18) becomes: Copt  Eout  out Ain   opt  Cgeo ,  Ein in Aout (13.18a) and Eq. (13.17a) becomes: 3D C opt   opt  C geo   opt  Ain n '2  sin 2  out .   opt  2 Aout n  sin 2  in (13.17b) Of course, the same formulas, making the square root, can be adapted to the case of 2D concentrators: C geo  lin / lout , 2D  opt  Cgeo  opt  Copt 328 lin n' sin out  opt  , lout n  sinin (13.20) (13.21) Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators where lin and l out are, respectively, the widths of the input and of the output openings of the linear concentrator. The quantities n and n’ as the indices of refraction of the medium at the entrance and at the exit of the concentrator, respectively. In the practical case of a concentrator exposed to the Sun, we have n = 1, because the medium to which is exposed the input aperture is the air. The index of refractive n’ is that of the medium in which is placed the receiver (the solar cell) at the exit of the solar concentrator. 13.2.2. Direct Lambertian Irradiation The “Direct Lambertian Method” (DLM) [26, 29, 41, 42] allows to study the transmission efficiency of a concentrator when the irradiation at input is integrated over all the directions in space. DLM simulates the behavior of the concentrator under diffused light, for example the diffuse solar radiation in a totally covered sky. A clear sky, in fact, contrary to what one might think, is not a good example of isotropic light, because the polarization of solar light by Rayleigh scattering produces a radiance strongly dependent on the direction of diffuse light respect to the direction of Sun [56]. Fig. 13.4 shows the scheme of DLM applied to a 3D-CPC concentrator, with Ldir constant radiance of the diffused light source. The total incident flux is: 2  /2 0 0   indir  Ldir  Ain  d   d  sin   cos    A in  Ldir , (13.22) where   Ain is the Étendue. In the following, we will consider, for simplicity, only concentrators with cylindrical symmetry, then  dir ( , ) will be set equal to  dir ( ) . The equations can be easily extended, whenever necessary, to the general case by reintroducing the dependence on the azimuthal angle . The flux “transmitted” to the output aperture becomes:  out dir    dir  2  Ldir  Ain   /2  d  sin   cos  ). dir ( (13.23) 0  The optical losses due to the “reflected” flux  dir and to the “absorbed” flux  dir are expressed respectively as:   dir  2  Ldir  Ain   /2  d  sin   cos   ), (13.24) ). (13.25) dir ( 0 dir  2  Ldir  Ain   /2  d  sin   cos   dir ( 0 329 Advances in Optics: Reviews. Book Series, Vol. 3  In such a way that:  dir   dir   dir   indir . Here  dir ( ) is the “direct collimated reflectance” and  dir ( ) is the “direct collimated absorptance” of the concentrator, as previously defined. lamb DLM gives the “direct lambertian transmission efficiency”  dir , also called “direct lamb lambertian transmittance”  dir , defined as the ratio of output to input flux:  lamb dir  lamb dir   indir  2   dir ...  2   dir (0)   /2  d  sin   cos   dir ( )  ... 0  /2  d  sin   cos (13.26)   dir , rel ( ) . 0 In a similar way, we define the other two quantities related to DLM: the “direct lambertian lamb lamb reflectance”  dir and the “direct lambertian absorptance”  dir : lamb lamb  dir   dir  ...  2   dir (0)    dir  2  indir  /2  d  sin   cos   dir ( )  ... 0  /2  d  sin   cos (13.27)   dir , rel ( ) , 0  lamb dir  lamb dir   indir  2   dir ...  2   dir (0)   /2  d  sin   cos   dir ( )  ... 0  /2  d  sin   cos (13.28)   dir , rel ( ) . 0 The output radiance, in general, is not constant like the input radiance, so we speak about an average output radiance: out Ldir  2  Ldir  Ain dir   Aout   Aout ...  2  Ldir  C geo   /2  /2  d  sin   cos   dir ( )  ... 0  d  sin   cos (13.29)   dir ( ) , 0 where C geo is the geometric concentration ratio. Now we define a new quantity, lamb C opt , the ratio between average output and input radiance: 330 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators lamb Copt  out Ldir  2  C geo  Ldir ...  2  C geo   /2  d  sin   cos    dir ( )  ... 0  /2  d  sin   cos  (13.30)  [1   dir ( )   dir ( )] . 0 From Eqs. (13.26), (13.30) we find the relationship: lamb  Copt out  Ldir A E A A E lamb   dir  C geo  indir  in  out out  in  out . Ldir Ein  Ain Aout Ein  dir Aout (13.31) Eq. (13.31) has the same form of the relationship defining the optical concentration ratio of a SC under collimated irradiation [17] (see also Eq. (13.18a)): coll  Copt Eout  A   dir  C geo  out  in . Ein  in Aout (13.32) lamb We define therefore the quantity Copt as the “optical concentration ratio under direct lambertian irradiation” or “direct lambertian concentration ratio”. The direct lambertian model can be applied also reducing the angular extension of the lambertian source from π/2 to a limit polar angle  m (see Fig. 13.5). The corresponding method, DLM ( m ) , is particularly useful when we analyze the behavior of nonimaging SCs. Because of the step-like profile of their optical efficiency (  dir ( )  const for 50 50 ;  dir ( )  0 for    acc ), in fact, the characterization of these SCs under direct   acc 50 lambertian irradiation can be limited to angles    m   acc , reducing in this way the time of computer elaboration or simplifying the experimental measurements. The theory of DLM ( m ) is just that developed until now for the DLM, modified in the limit polar angle of the diffuse irradiation. We have therefore for the input and output flux, respectively:   indir ( m ) 2 m 0 0    Ldir  Ain  d  d  sin   cos    Ain  Ldir  sin 2 m , (13.33) m  dir ( m )  2  Ldir  Ain  d  sin   cos  dir ( ) . (13.34) 0 The direct transmission efficiency of this method becomes: 331 Advances in Optics: Reviews. Book Series, Vol. 3 m lamb ( m )   dir dir ( m ) 2   d  sin   cos  dir ( ) . in  dir ( m ) sin 2  m 0  (13.35) The average output radiance becomes: m out Ldir ( m )  dir ( m )  2  Ldir  C geo  d  sin   cos  dir ( ),   Aout 0  (13.36) lamb and the optical concentration ratio C opt ( m ) is: m  lamb ( m) Copt out ( m ) Ldir   2  C geo  d  sin   cos  dir ( ). Ldir 0  (13.37) 13.2.3. Inverse Lambertian Irradiation We saw in the Introduction that, for the reversibility principle, the optical loss reported by a direct ray is the same as that shown by an inverse ray if the optical path is the same and if both starting rays are unpolarized. The attenuation factor for the radiance of the direct beam incident at point P in direction ( , ) represents the local direct transmission efficiency dir ( P, , ) , while the attenuation factor for the radiance of the ray emitted by the SC from point A in the reverse direction ( , ) represents the local inverse transmission efficiency inv ( P, , ) . We extend now these concepts to all points of Ain directly irradiated in direction ( , ) (DCM, see Fig. 13.2) and to the same points of Ain that emit light in the reverse direction ( , ) (ILM, see Fig. 13.3). If the inverse radiance at output aperture Linv (Fig. 13.3) is constant for all directions, that is, if the source at output aperture is Lambertian, then the inverse output radiance, Lout inv ( ,  ) , averaged over all points of Ain , must have the same angular distribution of the inverse transmission efficiency  inv ( , ) , averaged over all points of Ain . But the average inverse transmission efficiency  inv ( , ) must have the same angular distribution of the average direct transmission efficiency  dir ( , ) , because the transmission of the single connecting paths is invariant respect to the direction of travel of light. As consequence, we can affirm that the inverse radiance of the concentrator Lout inv ( ,  ) , irradiated on the output aperture with a uniform and non-polarized Lambertian source, is proportional to the efficiency of the direct transmission  dir ( ,  ) of a non-polarized collimated beam, that is the two corresponding relative quantities coincide. We have therefore that: Lout inv , rel ( ,  )  dir , rel ( ,  ), where: 332 (13.38) Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators Lout inv , rel ( ,  )  Lout inv ( ,  ) , Lout inv (0) (13.39)  dir , rel ( , )   dir ( , ) .  dir (0) (13.40) Eq. (13.38) establishes the equivalence between the “relative” inverse radiance and the “relative” direct transmission of the SC. The above discussion establishes therefore the suitability of the inverse lambertian method (ILM) to provide all information concerning the relative efficiency of transmission of the concentrator under direct irradiation,  dir , rel ( ,  ) (see Fig. 13.1). The simulated and experimental measurements of relative inverse radiance Lout inv,rel ( , ) of a solar concentrator is discussed in Sections 13.3, 13.5.2, 13.6. Here we recall that, to perform these measurements, it is sufficient to project the inverse light of concentrator towards a far planar screen and to record the image produced there; a simple elaboration of the image gives Lout , and so dir,rel ( ,) . Here we want to emphasize another inv , rel ( ,  ) fundamental aspect of ILM, that is the fact that it provides also the value of dir (0) , and so the “absolute” transmission efficiency dir ( , ) (see Eqs. (13.4), (13.40)), without recourse to any direct measure by DCM [36, 39, 41, 57], as it will be demonstrated by the following considerations. When the SC is irradiated in the reverse way (see Fig. 13.3), the exit aperture (oa) of area Aout becomes a Lambertian source with constant and uniform radiance Linv . The total flux, injected into the SC and function of radiance L inv , becomes: 2   ininv  Linv  Aout  d  0  /2  d  sin   cos    A out  Linv . (13.41) 0 The inverse flux transmitted to output, the input aperture (ia) of area Ain of the SC, supposed of cylindrical symmetry, is given by: out  inv  inv  2  Ain   /2  d  sin   cos  L  ), out inv ( (13.42) 0 where  is the direction and Lout inv ( ) is the radiance of inversely emitted light. We now define the inverse lambertian transmission efficiency, or “inverse lambertian lamb transmittance”,  inv , defined as the ratio of output to input flux: 333 Advances in Optics: Reviews. Book Series, Vol. 3      inv  ininv lamb inv ...  ...  2  C geo Linv 2  Ain   /2  d  sin   cos  L 0  /2  ...   Aout  Linv  d  sin   cos  L  ) out inv ( (13.43)  )  ... out inv ( 0 2  C geo Linv  Lout inv (0)   /2  d  sin   cos  L  ). out inv , rel ( 0 lamb Let us compare the inverse lambertian transmittance  inv of Eq. (13.43) with the direct lamb lambertian transmittance  dir of Eq. (13.26) by taking their ratio: 2  C geo   lamb inv lamb dir  Linv  Lout inv (0)   /2  /2  d  sin   cos  L 0  d  sin   cos  2   dir (0)  ) out inv , rel (  ... ) dir , rel ( (13.44) 0 ...  C geo  Lout inv (0) .  dir (0)  Linv This ratio is just a property of the SC and should not depend on radiance quantities as it lamb lamb appears in Eq. (13.44). To clarify this situation, we calculate the ratio  inv by  dir applying the simple condition Ldir = Linv , at which the total integral flux transmitted in the out “direct” and the “inverse” directions is the same:  out dir = inv , because such is the flux transmitted through the elementary connecting paths in the two directions. By putting out  out dir =  inv and using Eqs. (13.23) and (13.42) we find: 2  Ldir  Ain   /2  d  sin   cos   )  ... dir ( 0 ...  2  Ain   /2  d  sin   cos  (13.45a)  ). Lout inv ( 0 Putting Ldir = Linv and applying Eqs. (13.39), (13.40), we have: Linv   dir (0)   /2  d  sin   cos   )  ... rel dir ( 0 ...  Lout inv (0)   /2  d  sin   cos 0 334 (13.45b)  ).  Lout inv .rel ( Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators We finally find:  dir (0)  Lout inv (0) . Linv (13.45c) Eq. (13.45c) allows us to calculate  dir (0) by ILM measuring Lout , the average on-axis inv ( 0) inverse radiance of SC, and Linv , the radiance of the inverse lambertian source [39, 41]. From Eq.s (13.44), (13.45c) we find moreover that the ratio between the inverse and direct lambertian transmittances is equal to C geo , that is independent on radiance, as foreseen. out This could be also deduced by considering that, if  out dir =  inv , we have: lamb out  inv  indir  indir   Ldir  Ain  inv A    in   in  C geo , lamb in out  inv  dir  inv   Linv  Aout Aout  dir (13.46) that is: the “inverse lambertian transmittance” of a SC is C geo times its “direct lambertian transmittance”, or equivalently: the “input direct lambertian flux” needed to sustain an equal transmitted flux in the opposite directions is C geo times the “input inverse lambertian flux”. This result is not surprising; it is a direct consequence of the geometrical asymmetry of the concentrator and disappears when C geo = 1, that is Ain = Aout . It is interesting to note that this result does not require any information about the internal features of the SC, but is only dependent on the sizes of the lateral apertures. Eq. (13.46) tell us that the optical “transparency” of the SC to lambertian light is not symmetric. 13.2.4. Mixed Lambertian Irradiation Let us imagine now to irradiate both apertures of the SC by two different lambertian sources with Ldir  Linv (see Fig. 13.7). If  L  L dir  L inv is the difference of incidence radiance between input and output, then we have for the net flux through SC, in the direct direction: out out lamb in lamb in    net dir   dir   inv   dir   dir   inv   inv  ... lamb ...   dir  [ indir  C geo   ininv ]  ... lamb ...   dir  [  Ldir  Ain  C geo    Linv  Aout ]  ... (13.47) lamb ...  (  Ain   dir )  L. From Eq. (13.47) we deduce, for example, that a CPC immersed in an integrating sphere has no net flux flowing through it (the radiance is constant inside the integrating sphere, so  L = 0). Eq. (13.47) has a strong similarity with the Ohm’s law: I  G  V , where (W) has the role of current,  L (W/srꞏm2) the role of potential difference and  net dir 335 Advances in Optics: Reviews. Book Series, Vol. 3 lamb (  Ain   dir ) (srꞏm2) the role of conductance. The net flux inside the SC, indeed, is the natural optical partner of the electric current and the choice of the radiance as the optical partner of the electric potential is the only one which allows to put Eq. (13.47) in the form of the Ohm’s law. Attempts to assign the role of potential to the total input flux (  A L) or to the Étendue (  A) are in fact unsuccessful. From Eq. (13.47) we define the “direct conductance under lambertian irradiation” or “direct lambertian optical conductance” lamb : Gdir lamb lamb Gdir  (  Ain   dir ). (13.48) The surprising result is that, if we reverse the SC keeping fix the radiance gradient, now the flux flows in the inverse direction with the same conductance. We have in fact, changing the sign to both members of Eq. (13.47) and using Eq. (13.46): net out lamb in lamb in   inv  inv   out dir   inv   inv   dir   dir  ... lamb ...   inv  [  Linv  Aout    Ldir  Ain / Cgeo ]  ... ...  (  Aout   (13.49) lamb inv )  L, with L  Linv  Ldir . From Eq. (13.49) we define the “inverse conductance under lamb lambertian irradiation” or “inverse lambertian optical conductance” Ginv : lamb lamb Ginv  (  Aout   inv ). (13.50) From Eqs. (13.48), (13.50) we conclude that the two conductances are equal: lamb lamb Gdir  Ginv . (13.51) The result of Eq. (13.51) is surprising in the fact that the optical asymmetry of the SC has disappeared when the conductance of the SC is considered. Eqs. (13.48) and (13.50) show that the “optical conductance” can be put in the form: G lamb  (  A)   lamb , (13.52) that is: “conductance” = “Étendue” x “transmittance”. Now the equivalence between the two opposite conductances is direct consequence of the fact that the “direct étendue” is Cgeo times the “inverse étendue” and that the “inverse transmittance” is Cgeo times the “direct transmittance”. From Eq. (13.47) we derive the density of the net flux through the net input aperture J dir (the average net flux flowing through the unit area of the input aperture inside the SC in direct way): 336 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators net J dir   net lamb dir  (   dir )  L , Ain (13.53) net where L  Ldir  Linv , and the density of the net flux through the output aperture J inv (the average net flux flowing through the unit area of the output aperture inside the SC in the reverse way) is: net J inv  net  inv lamb  (   inv )  L , Aout (13.54) where  L  Linv  Ldir . If l SC is the length of the SC, we can introduce the quantity  L / l SC , the average gradient of radiance through the concentrator, a quantity which cannot be measured in practice, but which can be imagined existing inside the concentrator from a theoretical point of view. Eq. (13.53) becomes net lamb J dir  (   dir  lsc )  L lamb  (   dir  lsc )  grad L. lsc (13.55) In Eq. (13.55) gradL is intended as the average gradient of L. Eq. (13.55) is optically net equivalent of the Ohm’s law expressed in local form: J    E , with J dir (W/m2) with the lamb role of current density, gradL (W/srꞏm3) with the role of electric field and (  dir  lsc ) (srꞏm) with the role of electrical conductivity. The equivalent expression for the inverse current density is given by: net lamb J inv  (   inv  lsc )  L lamb  (   inv  lsc )  grad L. lsc (13.56) lamb We can define therefore a “direct lambertian optical conductivity”  dir and an “inverse lambertian optical conductivity”  lamb inv of a solar concentrator as follows: lamb lamb  dir     dir  lsc , (13.57) lamb lamb  inv     inv  lsc . (13.58) From Eqs. (13.57), (13.58) and (13.46) we find that: lamb lamb  inv  C geo   dir . (13.59) 337 Advances in Optics: Reviews. Book Series, Vol. 3 Eq. (13.59) tell us that the “inverse optical conductivity” of a SC is Cgeo times its “direct optical conductivity”, so it restores the asymmetry of the concentrator, as we have found for the transmittance efficiency. As we have seen for the direct local collimated method (DLCM), which applies only to a portion of input opening, also the local inverse method can be applied fruitfully to small or large areas of the exit opening. We talk in this way of areas Aout , and of efficiency of direct transmission to these areas as:  dir ( P ,  Aout ,  ,  ) . The  dir ( P ,  Aout ,  ,  ) efficiency can be obtained by applying the ILLM method to measurements of , when the receiver is inversely irradiated by a lambertian source L out inv ( P ,  A out ,  ,  ) placed in the  A out area and centered on point P. The new situation is like that which would occur if the concentrator could be amended as follows: the new receiver is the selected area of the old receiver; the new concentrator is the old concentrator plus the excluded part of the receiver. This new way of looking at the receiver is very powerful. In this way, in fact, we can study the efficiency of collection of any portion of the optical receiver, and since the radiation on the receiver is generally not uniform when the concentrator is directly irradiated, it happens often to be wonder about the direction of the direct rays arriving in a certain area of the receiver. Through the ILLM method, therefore, we can know from which direction the rays in excess in a certain area of the receiver arrive, or from which direction they are failing to arrive in a certain area of it. In a forthcoming part of this work, the applications of the ILLM method to nonimaging CPCs will be shown. 13.3. Equivalence between DCM and ILM In the theoretical Section 13.2.3 we have discussed the ILM method, demonstrating that it can be very effectively used to find the optical transmission efficiency of a solar concentrator. We have shown that the transmission efficiency has the same angular profile of the inverse radiance measured looking at the entrance opening of the SC, which in this way acts as a light source. The only condition required is that the SC be illuminated by the side of the exit aperture that is where it is placed the receiver (solar cell), by a uniform and unpolarized Lambertian source. Here we demonstrate, by simulation, that DCM and ILM are equivalent in giving the transmission efficiency of the concentrator. At this purpose, we have chosen two concentrators. The first is an ideal 3D-CPC with the following characteristics (see Appendix 13.B): diameter of entrance aperture 50 = 5°; 2a = 114.8 mm; diameter of exit aperture 2a’ = 10 mm; length L = 712.9 mm; acc focal length f = 5.44 mm; wall reflectivity Rw = 1.0. The maximum divergence of output 50 is  out = 90°. The second concentrator is like the previous one after rays when in = acc halving its length from the entrance side, then it is called HT-CPC (Half-Truncated CPC). Its characteristics are: diameter of entrance aperture 2a = 104 mm; diameter of exit aperture 2a’ = 10 mm; length L = 356.4 mm; focal length f = 5.44 mm; wall reflectivity 50 is almost unchanged, as we shall see below and as Rw = 1.0. The acceptance angle acc discussed in (see Appendix 13.B). The optical simulations were carried out by using the software for opto-mechanical modelling TracePro of Lambda Research [58]. To apply the 338 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators ILM method to the ideal 3D-CPC, we have irradiated the exit aperture by a uniform Lambertian source and projected the light onto a faraway ideal absorbing screen (with unitary absorptance). The screen was circular with D = 4 m diameter, placed at a distance d = 20 m. The use of large distances between CPC and screen is mandatory with the ILM method, because it is necessary to reduce as much as possible the polar angle resolution associated with each point on the screen (this aspect is discussed in detail in Section 13.5.2). Fig. 13.10 shows an example of raytracing with 500 k inverse rays. The CPC is hardly visible, being very small compared to the CPC-screen distance. These conditions assure an angular resolution better than 0.2° for all the points on the screen (see Eq. (13.63)). Fig. 13.10. Example of raytracing of the 3D-CPC with 500 k inverse rays. Fig. 13.11a shows the map of irradiance (W/cm2) produced on the screen, and Fig. 13.11b the irradiance profiles along x and y axes. The simulation of ILM at a PC requires long acquisition times, sometimes of the order of hours. Starting from the map of Fig. 13.11a, the following procedure was applied to get the transmission efficiency curve. Being the CPC a cylindrical symmetry concentrator, the irradiance map was first symmetrized with respect to the azimuthal angle, to get most of information from the raytracing data. Then we have converted the distance r into the polar angle :  = tan-1(r/d), and normalized the irradiance profile E() to the = 0° value, obtaining Erel ( ) ; the irradiance Erel ( ) was finally multiplied by the factor (cos)-4 to obtain the normalized radiance Lrel ( ) (see Eq. (13.C3) in Appendix 13.C). The normalized radiance profile Lrel ( ) is equal to the normalized transmission efficiency of the CPC, rel ( ) , obtained with DCM. The DCM simulations were carried out using the same TracePro software [58], by preparing a perfectly collimated beam at input of the 3D-CPC and an ideal absorber at the output aperture. The collimated beam was oriented at different polar angles respect to the optical axis, from 0° to 6° with 0.5° steps, and the flux absorbed by the receiver measured. The efficiency of transmission was simply the ratio between output and input flux (see Eq. (13.2)) and is reported in Fig. 13.12. Here, the transmission efficiency profiles obtained by the two methods, DCM and ILM, are reported. Without any doubt, the two methods are equivalent. The simulation of the second concentrator, HT-CPC, is here discussed (see Fig. 13.13a). The DCM and ILM were applied to HT-CPC as made for the ideal 3D-CPC. Fig. 13.13b shows an example of inverse raytracing. The map of irradiance on the screen is shown in Fig. 13.13c, whereas Fig. 13.13d shows its average cross section. Fig. 13.14a shows the 339 Advances in Optics: Reviews. Book Series, Vol. 3 curve of relative transmission efficiency measured with the DCM and the curve of relative inverse radiance measured with the ILM. As the wall reflectivity is unitary, the curve of relative transmission efficiency is equal to the curve of absolute transmission efficiency. The extreme coincidence between the two methods is clearly evident. Fig. 13.14b compares the DCM transmission efficiency of HT-CPC with that simulated for the original ideal 3D-CPC (with double length). The effect of halving an ideal CPC is clear: the acceptance angles remain quite constant, but a tail appears on the halved CPC just in correspondence of the 50 % acceptance angle (see Appendix 13.B for a discussion). (a) (b) Fig. 13.11. Map of irradiance distribution on the far screen (a) and cross section x/y profiles (b). Fig. 13.12. Comparison between the profiles of transmission efficiency and inverse radiance, simulated with DCM and ILM, respectively. 340 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators Fig. 13.13. Simulation of HT-CPC by ILM. (a) HT-CPC model; (b) Example of raytracing; (c) Irradiance map, and (d) Irradiance profile. (a) (b) Fig. 13.14. (a) Relative transmission efficiency measured vs. the incidence angle (DCM) and relative inverse radiance measured vs. the emission angle (ILM) of the HT-CPC concentrator; (b) The transmission efficiency of HT-CPC simulated by DCM is compared to that simulated for the original 3D-CPC. 341 Advances in Optics: Reviews. Book Series, Vol. 3 From the practical point of view, it is useful to repeat here that the great advantage of the ILM method with respect to DCM, when measuring the transmission efficiency, is that it requires only the recording of the reverse image produced on a screen by the SC, to obtain the angular profile of efficiency (the relative transmission efficiency), and the recording of the front image of the input aperture to calculate the efficiency on the optical axis, i.e. at 0 ° of polar angle (see Section 13.5.2). Conversely, the DCM method requires dozens of measures for different polar angle values in the case of SC with cylindrical symmetry, multiplied by the number of azimuthal angles of interest in the case of non-cylindrical symmetry. The ILM method contains, in two images, all the information of efficiency for all the polar and azimuthal angles, and this is really a great achievement. The ILM requires long acquisition times, but this is a problem for the computer and not for the operator. 13.4. Real Prototypes of Nonimaging Solar Concentrators Here we present some nonimaging SCs that we have used for the simulated and the experimental measurements. The first three SCs are of the type 3D-CPCs, but are not ideal (see Appendix 13.B), as they were obtained after a substantial change to their shape. An ideal 3D-CPC, in fact, although the many advantages mentioned in the Introduction, shows, as main drawback, to be too long respect to the linear dimensions of its input aperture. For example, an ideal 3D-CPC suitable for the use in PV, with a solar cell of around 1 cm diameter and a Cgeo  100x, has a length 7 times the diameter of the entrance aperture (see below). So, the first change to make on a CPC is to shorten it. This operation does not change too much its optical properties (see Appendix 13.B), so it’s a change that should be done. The second change to make on a CPC is the squaring of its input aperture, because this allows the efficient packing of the single optical units in a CPV module. This last operation produces four planar lateral walls which converge at the entrance opening. These walls are useless, then the last, very effective operation, is the removal of these walls. This was the ingenious idea of Antonini et al. [36, 37, 59-66], which led to the creation of a small, but efficient company (CPower), spin-off of Ferrara University, to produce medium concentration CPV modules made by nonimaging, very innovative optical units. The three modifications just discussed introduce three types of SC prototypes which we propose below. The last prototype we present is a nonimaging refracting concentrator realized at ENEA laboratories. The first prototype is the Truncated-CPC (T-CPC) that is an ideal CPC which was cut from the entrance side. The actual dimensional parameters are: 1-cm diameter exit aperture 2a’, 14-cm diameter input aperture 2a, 35.8-cm length L. Fig. 13.15a shows the CAD model of the T-CPC and Fig. 13.15b shows the T-CPC during an experiment with inverse light. As it can be seen, the inner wall surface was smooth, but deliberately left to the natural state after the forming process of a polyurethane prism, with the aim to study the effects of not mirroring. Fig. 13.16a shows the CAD model of the Truncated and Squared CPC (TS-CPC), the second presented prototype. The TS-CPC has a squared input aperture of 10-cm side, a circular output aperture of 1-cm diameter and a 35-cm length. Fig. 13.16b shows the prototype of TS-CPC during the characterization by a laser beam (DLCM). It was realized by forming a polyurethane prism and coating the internal walls 342 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators by adhesive strips of the VM2002 Radiant Mirror Film of 3M (see Fig. 13.17a). The reflectance of the film resulted quite insensitive to incidence angle and equal to 951 % in the 400-110 nm spectral interval (see Fig. 13.17b). The mirroring of the internal wall determined some imperfections on the TS-CPC wall shape, which caused some loss of light by scattering, despite the high intrinsic reflectivity of the film (see Section 13.6.1.1). (a) (b) Fig. 13.15. (a) CAD of the truncated CPC (T-CPC); (b) Photo of the T-CPC during experiments with illumination by inverse light. (a) (b) Fig. 13.16. (a) CAD model of the truncated and squared CPC (TS-CPC); (b) Photo of the TS-CPC during an experiment with laser illumination. As anticipated, the “Rondine” SC [36, 37, 59-66] is an evolution of the TS-CPC, due to a further reduction of length, the removal of the four adjacent planar faces and, most important, the deformation of its surface, following a patented design. The concentrator length has been defined to obtain about only one reflection for the rays entering parallel to the optical axis of the concentrator and striking the surface, to reduce the optical losses 343 Advances in Optics: Reviews. Book Series, Vol. 3 due to multiple reflections. The elementary concentrator has a squared input window and lateral apertures; this gives to the “Rondine” a profile that resembles the swallows, hence the Italian name assigned to the concentrator. Due to the squared input window, many elementary units can be closely packed in a dense array, without losing active front surface. This approach gives the same angular tolerance as that from flat-mirrored surfaces placed on the cut lateral planes. The absence of a symmetrical rotational axis allows a quite uniform irradiance distribution on the solar cell, reducing in this way possible losses in FF. Two different designs of optical units were realized: Rondine-Gen1 and RondineGen2, differing in dimension and shape (see Fig. 13.18). Specular reflectance (%) 100 96 94 Unpolarized light  = 633 nm 92 90 (a) 3M Film on plastic 98 0 10 20 30 40 50 60 Incidence angle (°) 70 80 90 (b) Fig. 13.17. (a) Photo of the internal wall of the TS-CPC. It is visible the output aperture closed by a solar cell; (b) Data of specular reflectance of the 3M film/substrate sample at  = 633 nm as function of the incidence angle of the laser beam. The average reflectance is 95  1 %. Fig. 13.19 shows the CAD models, not the same scale, of the two “Rondine” concentrating units. As explained before, the two Rondine units work in practice with the four lateral apertures opened. When operating the simulation of these concentrators, therefore, the application of the DCM method required the reintroduction of the four walls (see Fig. 13.19) to properly set-up the parallel beam at input, because its entrance aperture has a non-planar profile. This arrangement was not necessary when working with the ILM method. In this case, in fact, the presence or absence of the four walls was indifferent, as verified by simulation. This result establishes another point in favor of the ILM method with respect to DCM, namely that ILM can be applied, without any modification, to concentrators in which the profile of the input opening is not contained in a plan, which often occurs in non-imaging concentrators. This is not possible with DCM. The last tested prototype is the “PhoCUS” concentrating unit, an optical element used in the CPV system PhoCUS (Photovoltaic Concentrators to Utility Scale), a project of ENEA laboratories [48]. The primary optics were manufactured as PMMA refractive elements. A secondary optical element (SOE) was included to minimize the optical losses due to module assembly and tracking inaccuracies. In this work, the tested optical unit is a rugged “Mock-up” containing the primary refractive optics, the secondary optical element (SOE) 344 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators and a receiver. The “Mock-up” assembly is shown in Fig. 13.20. Two types of lenses were used: a “prismatic” lens and a “hybrid” lens (see Fig. 13.21). The SOE is a truncated inverted pyramid made of polycarbonate substrate coated by the VM2002 Radiant Mirror Film of 3M. This combination of materials assured a > 95 % reflectivity in the spectral range of the Silicon cell. x y (a) (b) Fig. 13.18. (a) Rondine-Gen1 single optical units based on NIO (Non-Imaging Optics); it is visible the exit aperture with its quasi-rectangular shape (the direction of x-axis is defined by the longer side); (b) Rondine-Gen2 single optical units; it is visible the exit aperture with its quasi-squared shape. Fig. 13.19. CAD models of Rondine-Gen1 (left) and Rondine-Gen2 (right) after the addition on the front aperture of four planar ideal mirrors. The two prototypes are shown with a different scale. 345 Advances in Optics: Reviews. Book Series, Vol. 3 (a) (b) Fig. 13.20. Schematic (a), and photo (b) of the Mock-up assembly comprising the primary lens, the secondary optics (SOE) and the solar cell. (a) (b) Fig. 13.21. Photos of the prismatic (a), and the hybrid (b) lens. 13.5. Practical Application of the SC Characterization Methods 13.5.1. Application of the DCM Method The application of the DCM method to the characterization of a SC would require, in principle, the use of a perfectly collimated light beam. This imply the use of a source with infinite radiance (see Eq. (13.7)), impossible to achieve in practice. On the other hand, the use of a perfectly parallel beam would be motivated by the need to draw a transmission efficiency curve with the maximum angular resolution possible. This need can only be justified if we want to have a perfect comparison between DCM and ILM, since ILM provides the transmission efficiency with an angular resolution practically of 0°. All this can be done through optical simulations on the computer. In practice, as a SC operates with the direct component of solar radiation, with a small angular divergence (±0.27°), the application of the DCM method can be made using a quasi-parallel beam having this angular divergence. To be precise, in this way we are applying the DLM(S) method, with S = 0.27° the angular divergence of the Sun, but since S is so small, we can consider the DLM(S) method practically equivalent to the DCM method. 346 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators The basic experimental setup adopted for measurement of the optical efficiency of a light concentrator is schematically reported in Fig. 13.22. In Fig. 13.22a, the light source (ls) illuminates the integrating sphere (is1) which acts as a Lambertian source that is a source of diffuse light with constant radiance at its exit aperture, the window (w) with diameter D. A portion of light emerging from the sphere is collected by the parabolic mirror (pm1), placed slightly off-axis with respect to window (w) and at a distance d from it equal to the focal distance f of (pm1). The mirror (pm1) produces in this way a parallel beam whose maximum angular divergence can be controlled by varying the diameter D of (w). To obtain a beam with solar divergence (±0.27°), it is necessary to keep the ratio d /D equal to ~100, the same ratio between Sun-Earth distance and Sun diameter. The parallel beam, spatially filtered by the diaphragm (di), illuminates the solar concentrator (sc). The light at the exit aperture of concentrator (sc) is directed to a second integrating sphere (is2) and its flux measured, through the photodetector (pd), placed inside (is2), by the radiometric unit (ra). Fig. 13.22. (a) Schematic of the experimental setup used with the direct method for measuring the light collected by a solar concentrator (sc) at different incidence angles; (b) Reference setup used to measure the flux incident at input of the concentrator. 347 Advances in Optics: Reviews. Book Series, Vol. 3 To perform the angle resolved measurements required to draw the optical efficiency curve, the concentrator (sc) must be oriented with respect to the parallel beam at different angles (, ). The polar incident angle  is sufficient to define the orientation of the concentrator when it has a cylindrical symmetry; for other concentrators, it is necessary to consider also the azimuthal incident angle . If I ( , ) is the photodetector signal measured at the different incidence angles, the transmission efficiency of the SC relative to the on-axis direction (0° polar incidence angle) is obtained by the configuration of Fig. 13.22a and is given by: rel opt ( ,  )  I ( ,  ) . I (0) (13.60) rel To get the absolute optical efficiency, opt ( , )  opt (0) opt ( , ) , it is necessary to make a further measurement, that of the incident flux at entrance of the concentrator. This is done by removing the concentrator (sc), decoupling it from sphere (is2), and orienting the collimated beam from mirror (pm1) towards a second parabolic mirror (pm2) (see Fig. 13.22b), which will provide to re-focalise the beam inside the same integrating sphere (is2). For this measurement, it is required the knowledge of the spectral reflectance Rpm of (pm2). If I ref is the photodetector signal measured with the reference setup of Fig. 13.22b, the angle resolved absolute efficiency of the concentrator becomes: opt ( ,  )  I (0) I ( ,  ) I ( ,  )    R pm . I ref / R pm I (0) I ref (13.61) Fig. 13.23 shows the photos of the experimental set-up of Fig. 13.22, realized at ENEAPortici laboratories, applied to the characterization of the prismatic lens PhoCUS. (a) (b) Fig. 13.23. Photos of the experimental set-up of the direct method for measuring the optical efficiency of the prismatic lens PhoCUS, at ENEA-Portici laboratories. 348 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators Alternative configurations of the experimental set-up are shown in Fig. 13.24 and Fig. 13.25. Here we use two integrated spheres instead of one to have a better integration of light inside the sphere used as source of diffused light. We distinguish between two different experimental configurations. In the first of Fig. 13.24a (configuration A), we have used a continuous light; in the second one of Fig. 13.24b, (configuration B), we have used a modulated light. In both configurations, the light source (ls) is a fluorescent lamp with a spectrum near to that of the Sun (T = 5500-6000 K), facing the inside of the integrating sphere (is1). In the configuration A, as detector for flux measurements, we have used a CCD camera, facing the inside of the integrating sphere (is3), operating like a normal photodetector; the collected flux is measured by averaging the intensity of the digital CCD image. Based on our experience, the CCD, due to its high sensitivity, is the only means of measuring the continuous, very low flux transmitted by the (sc) to the integrating sphere (is3). Fig. 13.24. Schematic principle of the direct method used for the measure of the output flux. (a) Configuration A: the source (ls) is a continuous light and the radiometer is a CCD. (b) Configuration B: the source (ls) is a chopped light and the radiometer is a lock-in amplifier. The diameter of window (w) is 5 mm and the focal length of (pm1) is 500 mm. 349 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 13.25. Schematic of the experimental configuration A applied to the measure of the flux at 0° incidence. A similar schematic applies for the experimental configuration B. In the configuration B, the light from the source (ls) is modulated by the chopper (ch). The Lambertian light from window (w) is then modulated and produces a modulated signal on the photodetector (pd), recorded by the lock-in amplifier (li). The use of the modulated-light configuration of Fig. 13.24b allows to increase the sensitivity of measurements made by using a photodetector (pd). Fig. 13.26 shows a photo of the experimental set-up used with configuration A. The parabolic mirror (pm1) was a low cost, commercial mirror with a focal length of 50 cm. Fig. 13.26. Photo of the apparatus for DCM measurements with configuration A. The tested sample is a nonimaging concentrator Rondine-Gen1. Photos of configuration B are shown in Fig. 13.27. Here a black wall separates the light source section from the receiver section. The mirror (pm1) is a high-quality, high-cost, Pyrex parabolic mirror of 1500 mm focal length. The use of a uniform quasi-collimated beam is very important in these measurements. The intensity distribution (irradiance) of light on the cross section of the collimated beam was measured by projecting the parallel 350 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators beam on a lambertian diffuser [67] and recording the intensity distribution by means of a CCD. The intensity profile shows a flatness modulation of around ±8 % for the commercial parabolic mirror used for the configuration A of Fig. 13.26. This, as we shall see, will produce a small distortion of the efficiency curve and a small alteration of the acceptance angle (see Fig. 13.69). The projection on a lambertian diffuser of the parallel beam from the high-quality parabolic mirror is shown in Fig. 13.28a. The intensity map shows a flatness better than ±0.5 % within 100 mm width (Fig. 13.28b). ls+is1 sc fan is2 w ch sw pd li a) b) pm1 c) d) Fig. 13.27. Photos of the apparatus for DCM measurements with configuration B. Light source section with the separating wall (sw) (a). Receiver section with the lock-in amplifier (li) (b). L. Zampierolo is aligning the Rondine-Gen1 with the parabolic mirror (c). Parabolic mirror (d). The flux measurements at the exit aperture of the (sc) can be carried out in two modes. In the first one, that we have called SPHERE mode, we have coupled the exit aperture of the (sc) to the (is3) integrating sphere, provided with an internal photodiode (pd) (a high efficiency solar cell), as illustrated in Fig. 13.29a; in the second one, that we have called CELL mode, we have closed the (sc) exit aperture directly on a high efficiency solar cell, the same used in the CPV module (see Fig. 13.29b). 351 Advances in Optics: Reviews. Book Series, Vol. 3 (a) (b) Fig. 13.28. (a) Projection of the parallel beam on a lambertian diffuser after reflection from pm1. Plot of the intensity of light measured on a 100×100 mm2 square cross section of the parallel beam after reflection from pm1. solar cell a) b) Fig. 13.29. The Rondine-Gen1 solar concentrator is closed on the integrating sphere (is3) (SPHERE mode) (a); The Rondine-Gen1 solar concentrator is closed directly on the high efficiency solar cell (CELL mode) (b). In both cases, the photodiode (pd) inside (is3) and the high-efficiency solar cell were connected to the lock-in amplifier (li) shown in Fig. 13.27b. The reason why we adopted two different receivers for measuring the flux transmitted by the (sc) is the following: in the SPHERE mode, the integrating sphere behaves as an ideal receiver, because we haven’t losses for reflection at its exit window; then the measure gives the transmission efficiency of the (sc) itself, that is of the optical unit alone, regardless of the type of receiver used; in the CELL mode, on the other hand, the measure gives the transmission efficiency of the system (sc) + PV receiver, and then the more realistic angle-resolved optical performance of the PV solar concentrator when it is mounted on a CPV module. For a correct application of the CELL mode, however, we need to use a solar cell with the same optical and electrical characteristics of the true PV receiver used in practice. 352 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators Fig. 13.30 shows some photos recorded during DCM measurements on the refractive, nonimaging concentrator PhoCUS. Fig. 13.30a shows the phase of assembly of the mockup with the prismatic lens. Fig. 13.30b shows the square shaped image of the focused light produced by the prismatic lens. The illuminated area corresponds to the area of the PV receiver, a high-efficiency SunPower cell. Fig. 13.30c shows the (sc) during measurements of the output flux by the integrating sphere coupled to a photodiode (SPHERE mode), and Fig. 13.30d shows the assembled mock-up with mounted the prismatic lens. (a) (b) (c) (d) Fig. 13.30. Experimental set-up of the DCM method applied to the PhoCUS concentrator. (a) F. Aldegheri is working on the assembly of the (sc) at the receiver section. (b) Square-shaped image of the focus produced by the prismatic lens. (c) DCM Flux measurements by the SPHERE mode. (d) Photo of the assembled PhoCUS mock-up. 13.5.2. Application of the ILM Method (Parretta-Method) The “inverse method”, initially known as ILLUME (Inverse Illumination Method) [44], was later revisited and improved, assigning the new name of “Inverse Lambertian Method” (ILM). The ILM was born as an alternative to the DCM just to obtain the relative 353 Advances in Optics: Reviews. Book Series, Vol. 3 (to 0°) or normalized transmission efficiency of the SC. We have demonstrated, in fact, in the theoretical Section 13.2.3, and by simulation in Section 13.3, that the two methods are equivalent. Subsequently, we could find the procedure to get also the “absolute” transmission efficiency. Later, we realized that ILM could become a powerful tool for analyzing other SC properties, by applying the ILLM (see Sections 13.6.1.3 and 13.6.2.1). ILM greatly simplifies the experimental apparatus for measuring the angle-resolved transmission efficiency, both relative and absolute, and drastically reduces the number of measurements. The method consists in irradiating the concentrator (sc) in a reverse way by placing a planar Lambertian light source (ls) of uniform radiance Linv at the exit aperture, and in measuring the radiance Linv () of the light emitted by the concentrator from the input aperture as function of the different orientations in space, characterized by the polar and the azimuthal emission angles  and  (see Fig. 13.31) (here we use for simplicity the same symbols for the angular direction of the rays incoming to and outcoming from the input aperture). When inversely illuminated, the concentrator becomes a light source whose radiance will no longer be constant, because the concentrator changes the angular distribution of the rays emitted by the lambertian source (ls) before they are emitted from the entrance opening. Fig. 13.31. Basic scheme of the inverse lambertian method (ILM). A Lambertian light source is applied at the exit aperture of the (sc). Differently from DCM, where measurement of output flux out (, ) every time requires changing the orientation of the concentrator with respect to the quasi-parallel beam, the measure of Linv () is now easy and immediate, because it can be obtained projecting the inverse light on a far planar screen, recording the intensity of its image and elaborating it at a computer (see Fig. 13.32). The processing procedure depends, however on the type of measurement, if simulated or experimental, and the details of this procedure are given in the Appendix 13.C. We summarize here this procedure. If the inverse method is simulated at a computer, the planar screen (ps) is assumed as an ideal absorber, and the irradiance E () of the absorbed light is easily measured and transformed into the inverse radiance Linv (). If the inverse method is applied experimentally, on the contrary, the planar screen must be a white diffuser with lambertian properties and a CCD, or a webcam, must be used to record the image on the screen. Still, from the intensity map of the image, the relative inverse radiance is calculated. In this case, the irradiance of incident light is measured 354 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators indirectly, deriving it from the intensity of the CCD image, ICCD (). The only foresight to follow when applying the inverse method is that the planar screen, oriented perpendicular to the optical axis of the concentrator and illuminated by the inverse light, must be placed at a distance d from the concentrator much higher than the linear dimensions of the input aperture, to have an adequate angular resolution for the profile of optical efficiency (see Fig. 13.33a). Fig. 13.32. Schematic of the experimental ILM apparatus. For example, for a circular input aperture of diameter D, the angular resolution (uncertainty) res for all the points (in a circle) of the screen characterized by the polar emission angle  is given by:  res ( , D, d )  tg 1 [ D  cos2  (2d  D  sin   cos )]  ... ...  tg 1 [ D  cos2  2d ]. (13.62) The angular resolution is the worst on the optical axis:   res , max  tg 1 ( D 2 d ), (13.63) and improves at increasing  (see Fig. 13.33b) Fig. 13.33c shows, as an example, the angular resolution calculated for ILM applied to the TS-CPC when its distance from the screen is 360 cm. Since the angular range of light emission is generally small, we can assign to the points of the emission radiance the angular resolution obtained by Eq. (13.63). The need to have d >> D implies some limits to the practical application of the ILM in laboratory, where distance d is of the order of some meters. For example, a concentrator of 10-cm aperture size requires a screen at 5 m to have a resolution of at least ~0.5° on the optical axis. This is the only drawback of the ILM when applied experimentally. When ILM is simulated on the computer, on the other hand, it is easy to set a sufficiently high value of d /D to obtain the desired resolution, of course by using a ray-tracing with number of rays, and then with processing time, greater at greater resolution. The choice of d, then of the angular resolution, must be made mainly considering the expected value for the angle of acceptance. A resolution of ~0.5°, for example, cannot be tolerated for an acceptance angle (at 50 %) of about 1° or less, while it is acceptable for an acceptance angle of at least 5°. 355 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 13.33. (a) The drawing shows the effect of the SC-screen distance on the angular resolution at point P. (b) The drawing shows the effect on the angular resolution of the position on the screen of the point P. (c) Angular resolution as function of the emission angle, calculated for a TS-CPCscreen distance of 360 cm. 356 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators Referring to the measurement of E ( ,  ) , it may happen that the optics of the CCD is not always adequate to capture the entire inverse image at the distance d between concentrator and screen. In this case, that we have effectively experienced, it is possible to double the optical path between the screen and the CCD interposing a mirror (mi) between the CCD and the screen, as shown in Fig. 13.34. Fig. 13.34. When the full inverse image on the screen cannot be captured by the CCD objective, a mirror (mi) can double the optical path between screen and CCD. With simulation measurements, the irradiance E (, ) on the planar screen is transformed into the radiance Linv (, ) of the concentrator by the following expression (see the Appendix 13.C): L inv ( ,  )  1 d2 E ( ,  )  , Ain cos 4  (13.64) where Ain is the input aperture area and d is the on-axis distance between the planar screen and the centre O of the input aperture of the concentrator. To obtain the relative profile of the inverse radiance, L rel inv ( ,  ) , that is the radiance normalized to the 0° value, it is sufficient to measure the normalized irradiance on the screen, E ( , ) : rel rel Lrel inv ( ,  )  E ( ,  )  1 . cos 4  (13.65) The factor (cos)-4 considers the fact that the screen is a flat surface rather than spherical, and this determines a (cos)-2 factor, and the points on the screen (ps) are not located at the same distance from the centre of the opening entrance, which is the point from which we measure the angles, and this determines another (cos)-2 factor. When the inverse method is applied experimentally, Eq. (13.64) modifies and becomes (see the Appendix 13.C): L inv ( ,  )    f 2 d2 R  Ain  I CCD ( ,  )  1 , cos 8  (13.64a) 357 Advances in Optics: Reviews. Book Series, Vol. 3 where f is the focal length of CCD and R the reflectivity of the lambertian screen. We emphasize here the fact that the screen must be a Lambertian reflector for the Eq. (13.64a) to be valid. The normalized radiance now becomes: rel Lrel inv ( ,  )  I CCD ( ,  )  1 . cos8  (13.65a) We have demonstrated on several occasions that the relative inverse radiance profile rel L rel inv ( ,  ) coincides with the relative optical efficiency profile  opt ( ,  ) of the concentrator operating in the direct mode (DCM) when the direct beam is parallel or quasi-parallel. When the direct beam is not strictly parallel, we should take account of its divergence and to put it as the uncertainty on the measured polar angle. We have therefore for the optical efficiency of the concentrator: rel dir ( , )  dir ( , ) dir (0)  L rel inv ( , )  dir (0), (13.66) where L rel inv ( ,  ) is obtained from Eq. (13.65) or (13.65a) depending if we are operating with a simulation program or with experimental measurements. We conclude this part highlighting the fact that the radiance L rel inv ( ,  ) summarizes all information relating to the light collection properties of the concentrator in relation to its orientation with respect to the solar disk. The inverse method as discussed until now seems to allow determining only the relative angle-resolved optical efficiency of the concentrator, by processing the intensity image produced on the planar screen by the concentrator irradiated in the inverse way. To obtain the absolute efficiency as indicated by Eq. (13.66), we should ask the direct method for help to measure the on-axis optical efficiency dir(0) . If it were so, we should need to setup, besides ILM, also the direct method apparatus just for one measurement, and the advantages of the inverse method in terms of simplicity of the experimental apparatus should be lost. Fortunately, new developments of the theory [22, 26, 36, 38, 57] provided a way of obtaining dir(0) also by the inverse method. It was shown in fact that the on-axis efficiency dir(0) can be obtained by the expression: dir (0)  L inv (0) , LREC (13.67) where Linv (0) and LREC are radiances measured on the image recorded by the CCD camera, or the web camera, oriented towards the input aperture of the concentrator irradiated in the inverse mode (ILM) (see the schematic of the apparatus in Fig. 13.35). Linv (0) is the average radiance of the whole input aperture, whereas LREC is the radiance of the receiver, the exit aperture, corresponding to the radiance of the Lambertian source. 358 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators Fig. 13.35. Experimental set-up for determining the on-axis optical efficiency of the concentrator. The CCD camera is turned towards the concentrator irradiated in inverse mode and the image of the input aperture is recorded. It should be emphasized that Eq. (13.67) is easily applied to concentrators for which the exit aperture is visible from the input aperture, like any concentrator realized only by reflective optics, like the 3D-CPCs. For concentrators with refractive components in the front, like the PhoCUS, the measure of LREC becomes independent of that of Linv (0) and requires the removal of the primary lens for measuring the radiance of the lambertian source. As the optical efficiency dir (0)  1, we will always have Linv (0)  LREC . At its first appearance [44], the inverse method, at that time called ILLUME, was applied by illuminating, with a laser beam from the inside of the TS-CPC concentrator, a Lambertian white diffuser placed on the exit aperture (see Fig. 13.36a). This configuration, indeed, was the evolution of an accidental experiment that brought me to the discover of the ILLUME method (now ILM or P-Method (Parretta-Method)). On that occasion, the illumination of just the centre of the diffuser with a laser beam produced a back-reflected light which appeared as a well-defined image, with the same symmetry (square) of the concentrator input aperture (see Fig. 13.36b). Subsequently, I understood that the diffuser should be full illuminated to apply the ILLUME method in a correct way (see Fig. 13.36c), and this was made coupling the laser with a beam expander (see Fig. 13.36d). However, the laser beam was coming from the same side towards which the inverse light was reflected, so it was necessary to use a screen with a hole to allow the laser beam to pass and, at the same time, to allow the reflected light to be projected (see Fig. 13.37). Subsequently, another configuration was attempted to create a Lambertian source that is by applying a semi-transparent diffuser to the CPC exit aperture and illuminating it externally by means of a lamp (see Fig. 13.38). This configuration, however, did not work well, as a transmission diffuser never produces a Lambertian light, but a light where the most divergent rays are penalized. To produce the Lambertian source on the back of the concentrator (sc), therefore, the best solution remains the use of a lamp (lp) coupled to a pair of integrating spheres, (is1) and (is2) (see Fig. 13.39a). A Xenon arc lamp is the best choice to simulate the direct solar spectrum, whereas the integrating sphere is the best choice to obtain a lambertian light source [68-75]. The concentrator is grafted on to the output window of (is2) as shown in Fig. 13.39b, where it is also shown the “baffle” (ba) at the centre of the sphere. The baffle has the function to filter the light coming from sphere (is1) and to assure a lambertian distribution of light at the exit window of (is2). The planar screen is placed in laboratory at a proper distance from the source and there the 359 Advances in Optics: Reviews. Book Series, Vol. 3 characteristic light image, leaved by the inverse beam as a fingerprint, is produced (see Figs. 13.39c, d). Fig. 13.36. Application of ILM with a Laser beam. (a) Closing of exit aperture with a Lambertian diffuser. (b) Inverse light projected on a nearby screen. (c) The diffuser is fully illuminated. (d) The laser beam is expanded. Fig. 13.37. Schematic principle of the inverse illumination method (ILLUME), utilizing a laser and a Lambertian diffuser to produce the inverse Lambertian light. 360 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators (a) (b) (c) Fig. 13.38. (a) Schematic principle of the inverse illumination method (ILLUME), utilizing a lamp (lp) and a semi-transparent diffuser (ld) to produce the inverse Lambertian light. (b) Back illumination of the diffuser (ld) by the lamp (lp). (c) Inverse illumination of the TS-CPC by a lamp (lp). The inverse image produced on the screen is then analysed by means of the software operating with the CCD camera (in our case the HiPic software operating with the Hamamatsu CCD camera). Fig. 13.40 shows, as an example, the intensity profiles recorded along the horizontal (x-axis) and vertical directions (y-axis) of the inverse image produced on the screen (ps). Some black dots are placed at known distances on the screen (ps) and used as reference points of a Cartesian frame for measuring the distances on the screen and then calibrating the polar and azimuthal angles associated to any point. During the calibration of distances, the horizontal and vertical profiles were traced in such a way to take in the dots, which appear in the profiles as thin peaks (see Fig. 13.40). To get the correct irradiance profile on the screen, it could be needed to adjust the intensity and shape of the image for possible effects of perspective; this depends on the actual position of the CCD camera respect to the optical axis and is not necessary when the camera is placed very close to concentrator. If necessary, the perspective correction can be applied by using a specific program [67, 76-79]. Once known the distances between the dots, the calculation of angles is straightforward. The photo of the front side of the Rondine irradiated in the inverse mode is shown in Fig. 13.41a. The CCD camera is placed very close to the concentrator and on the plane of its input aperture. 361 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 13.39. (a) Photo of the ILM (ex ILLUME) apparatus during characterization of the RondineGen1. (b) Photo of the Rondine-Gen1 "grafted "on the integrating sphere (is2). (c) The planar screen and the ILM image produced on it by the Rondine-Gen1. (d) ILM image produced on the screen by the TS-CPC. (a) (b) Fig. 13.40. Analysis by HiPic software (Hamamatsu) of the ILM image produced by the RondineGen1 concentrator. The irradiance profiles, traced along the horizontal (x-axis) (a) and the vertical (y-axis) (b) directions, are used to calibrate the polar angle of the points along the corresponding axis. 362 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators Fig. 13.41. (a) Photo of the Rondine-Gen1 during ILM measurements. The CCD is placed just below at short distance. (b) Image of the front side (on-axis) input aperture of Rondine-Gen1. The blue frame surrounds the area of the receiver (REC), that is of the lambertian source; the red frame surrounds the total input area. Fig. 13.41b shows the detail of the entrance aperture sight in front, outlined by the red frame, with the central region REC, outlined by the blue frame, corresponding to the Lambertian source (the exit window of the integrating sphere). The central region is the most lit of the image, because it is the direct source of inverse light, which does not undergo attenuation inside the concentrator. If the lambertian source is made well, moreover, the image of REC is very uniform, as it represents the constant radiance of an integrating sphere. Following Eq. (13.67), the on-axis efficiency dir (0) is equal to the ratio between the average intensity of the red region and the uniform intensity of the blue region. Indeed, if we would take this measure at different CCD orientations, we would precisely get the absolute transmission efficiency: dir ( , )  Linv ( , ) . LREC (13.68) By doing so, however, we would fall back into the same drawbacks of direct method, because we would need again to carry out a measure for each orientation of the CCD. However, when the ILM method cannot be applied by projecting the reverse light on a screen, the use of a CCD which is oriented at different angles towards the concentrator becomes a very interesting alternative to the DCM method, because it is much cheaper (we have seen that the application of DCM involves the use of a quality parabolic mirror if we want to get a uniform parallel beam) (see Figs. 13.76 (b, c)). This application of ILM would require only a lamp, an integrating sphere and a webcam, all very economical components (the issue of integrating spheres is discussed apart in Appendix 13.D). If you consider Eq. (13.68), in fact, the use of the Lambertian screen is no longer necessary. 363 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 13.42a shows the input aperture of the refractive, nonimaging PhoCUS concentrator as it appears in the dark when irradiated in the inverse way, and Fig. 13.42b the corresponding image produced on the planar screen. The image is now smaller than that produced by the Rondine concentrator, because the acceptance angle is very low, 1°. In Fig. 13.42b are visible the black dots on the light image, used to calibrate the distances of the points on the screen, and then to assign the correct  and  angles to each point P. (a) (b) Fig. 13.42. In (a) the input aperture of the PhoCUS concentrator as it appears in the dark when irradiated in the inverse way; in (b) it is shown the corresponding image produced on the planar screen. To measure the absolute on-axis (0° incidence) optical efficiency of the Mock-up,  dir (0) , we had to use a different procedure respect to that used for the Rondine concentrator. In this case, in fact, we have a refractive concentrator, that needs the recording of two images for obtaining  dir (0) . We have oriented the CCD towards the concentrator, aligned with its optical axis, and we have taken one image of the full input aperture with the lens; after, we have taken a second image of the input aperture after removing the lens. In the first image, we have taken the mean radiance of the lens, Linv(0) ; in the second image, we have taken the mean radiance of the sphere cavity, LREC , corresponding to the lambertian source. The ratio between these two quantities has given the absolute on-axis optical efficiency of concentrator  dir (0) and, from Eq. (13.66), the absolute angle-resolved optical efficiency: dir (0)  Linv (0) / LREC . 13.5.3. The Application of DLCM Method As we have discussed in Section 13.2.1, the DLCM is a method that examines the local properties of the solar concentrator, then it can be effectively applied by using a laser for the irradiation of the input aperture. The method, even though constrained by lengthy measurements, gives nevertheless interesting information on local mirror surface defects or manufacturing defects, like internal wall shape inaccuracies. It is very useful to 364 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators investigate CPC-like concentrators, which concentrate light through multiple reflections on their internal wall. The measurements with the laser can be supported by optical simulations with commercial codes. The method, simple to apply, requires just a laser to scan the CPC input aperture following a matrix-like path (see Fig. 13.43a), at a controlled orientation of the beam. For an electronically driven movement of the source, the laser can be fixed on a x/y table and moved following a matrix-like pattern of points (typically 25×25 points with 4 mm steps). In a simpler setup, the laser is moved manually along the vertical (y) direction on a column and translated horizontally along x direction on a rail (see Figs. 13.43b and 13.44). In this way, the square aperture of the concentrator is entirely scanned along x and y directions. Fig. 13.43. (a) Representation of the scanning process of the entrance aperture of a CPC with the laser beam; (b) Schematic experimental setup of the “laser method”. The output flux is measured directly by a photodetector (pd) coupled to a radiometric unit (ra), or through an integrating sphere (see box). (a) (b) Fig. 13.44. (a) The experimental DLCM apparatus with the manually driven laser. A photodiode is visible on the back of the TS-CPC; (b) Photo of the TS-CPC during the centering of the laser beam. 365 Advances in Optics: Reviews. Book Series, Vol. 3 Particularly important is the laser beam polarization state. It must be unpolarized, but in general the available laser beams are partially polarized. To overcome this, it is necessary to double the measurements, repeating them with the laser rotated of 90° respect to the previous ones, and making the average of the flux values. The light collected at output of the CPC can be measured by a photodetector (pd), controlled by a voltmeter (ra). The reference flux at input is obtained by a calibration step performed sending the laser beam to impinge directly on the (pd) surface. 13.5.3.1. Optical Efficiency Measurements The map of (pd) signals is then transformed in a map of optical efficiency of the CPC by the ratio between test signals and average reference signal. The photodetector (pd) plays an important role in the precision of efficiency data. Its response in fact must be as much as possible insensitive to variations of incidence angle of laser beam at the exit aperture of the concentrator. Regarding this, the results of previous investigations on the angleresolved absorbance of photovoltaic devices have been considered [47]. A preferred way to measure the flux at output of the CPC is to match it to an integrating sphere provided with a photodetector (pd) inside it (see the box in Fig. 13.43b), but it depends on the power of the laser in use, because the low integrating sphere sensitively reduces the irradiance on the photodetector compared to that of direct irradiation. If the entire CPC surface is scanned, the integration of the maps gives the value of transmission efficiency for a specific orientation of the laser, so a curve of optical transmission efficiency can be drawn, and the acceptance angle measured, after joining more maps of flux obtained at different polar angles. The analysis of the single maps allows to obtain interesting information on light collection by the different regions of CPC input area. It reveals, moreover, how the efficiency of light collection depends on several factors like surface reflectivity, number of reflections of the single beam, local angle of incidence, local surface defects, and so on. By comparing the simulated maps with the experimental ones, then is possible to emphasize the effects directly related to manufacturing defects. Fig. 13.44a shows a photo of the simple apparatus with a manually driven laser with = 633 nm wavelength and 5 mW power. The TS-CPC is positioned on a rotating support, provided with a goniometric scale [80], by which it is possible to fix the azimuthal angle of incidence of the laser beam. The incidence angle is adjusted by first aligning the laser beam with the TS-CPC, then projecting the beam on a far screen for adjusting the desired incidence angle. Fig. 13.44b shows a photo of the TS-CPC during the centering of the laser beam. 13.5.3.2. Beam Exit Angle Measurements The measurement of exit angles of the laser beam at the output aperture of a CPC, can be carried out by using the simple apparatus schematized in Fig. 13.45a and shown in Fig. 13.45b. A plastic hemispherical globe (hg), made from a garden lamp, is centered on the 366 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators output aperture of the TS-CPC and drawn with parallels and meridians to indicate the polar and azimuthal angles of the laser beam spot. The opalescence of the globe, realized by sandblasting, allows to visualize the impact point of the laser beam and then, with a good approximation, its exiting direction. Fig. 13.45a shows how the two angles are measured: the polar angle  (0°-90°) on the parallels and the azimuthal angle  (0°-360°) on meridians, where  is measured starting from y axis and rotating towards x axis. Fig. 13.45. (a) Schematic of the experimental configuration for measurement of beam exit angle. It is illustrated a ray entering the TS-CPC and exiting at the polar angle . The azimuthal angle  is measured by rotating the y axis towards the x axis; (b) Photo of the apparatus with the hemispherical globe (hg) fixed to the TS-CPC and the laser. 13.5.4. The Application of the PH-Method (Parretta-Herrero Method) The schematic configuration of PH-Method is reported in Fig. 13.46a. In the figure, we have placed the Rondine concentrator deliberately enlarged to better show the path of the rays. The parabolic mirror (pm) has an aperture diameter D, a length L, a focal length f, and the square absorber (ps) has a side l. The solar concentrator (sc) coupled to the lambertian source (ls) is placed just below the screen, with its input aperture planar to the screen surface. The position of the concentrator must be accurately calculated placing it at a suitable distance from the optical axis to avoid any interference with rays reflected by the mirror (pm). This distance is found after fixing the maximum angular divergence, m (with m > 0) of inverse rays to be recorded along the x/y axes. Angle m must be chosen at a value where the optical efficiency is sufficiently low (  20 % of the maximum value) and the optical efficiency curve sufficiently well defined. Fig. 13.46a simulates red rays emerging from the concentrator with divergence m, which converge to the bottom of the screen y(m,), just at the upper edge of the Rondine, after being reflected by mirror (pm); this is highlighted in the enlarged detail of Fig. 13.46b. Fig. 13.46a shows also that all the rays (green) with direction parallel to the z axis are focused at the center of the screen. The y coordinate of rays emitted on the y/z plane at angle m can be expressed as: 367 Advances in Optics: Reviews. Book Series, Vol. 3 y ( m )  2 f (1  cos  m ) . sin  m (13.69) (a) (b) Fig. 13.46. (a) Schematic principle of the Parretta-Herrero (PH) method. (pm): parabolic mirror, (ps): planar screen, (sc): solar concentrator, (ls): lambertian source. D and f: diameter and focal length of (pm), respectively; l: side of the screen (ps). The concentrator lies below the coordinate y ( m ) in such a way to avoid any interference with the rays exiting at    m ; (b) It is shown a parallel beam emitted at    m on the y/z plane, which focuses at the bottom edge of the screen (ps). When the PH-Method is applied by simulation, it is necessary first to calibrate the angle coordinates in terms of x and y coordinates on the square screen (ps). This can be done by setting a uniform and parallel source of light exiting from the input aperture of the Rondine concentrator (see Fig. 13.46b) and oriented towards the mirror (pm) at some calibrated polar angles measured on the y/z incident plane, and on a plane parallel to the x/z plane and crossing the center of concentrator. Once a certain number of polar angles for the parallel beam are chosen, the x/y coordinates of the point on the screen, where the beam 368 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators is focused, are recorded. The calibration provides the angle coordinate along x and y axes, as function of the pixel number npx (for example 128 for both axes covering the entire screen). We report here an example of the calibration made for the Rondine Gen1 concentrator [37, 45]:  X ()  14.94  0.232  n px , (13.70a) Y ()  14.75  0.226  n py . (13.70b) The calibration must provide an excellent linear, and pratically equal behavior for both axes. Fig. 13.47 shows, as an example, the calibration curve found for the Rondine Gen1. Fig. 13.47. Calibration curves of the emission angle vs. pixel number relative to the x and y axes of the screen. 13.6. Experimental Results 13.6.1 The Truncated and Squared CPC (TS-CPC) 13.6.1.1. Local Optical Efficiency by the Laser Method (DLCM) Fig. 13.48 shows some of a series of experimental local efficiency maps obtained at  = 0° azimuthal angle (input aperture edge horizontal), orienting the laser beam towards left (looking at the CPC input aperture), at different incidence angles  with respect to z optical axis. Besides the data of local absolute efficiency, each map brings also the information about average efficiency and standard deviation. Some of the  = 0° maps have been also simulated and are shown in Fig. 13.49. Other maps were obtained by measurements at  = 45° and are shown in Fig. 13.50. The summary of all the measured data for the total aperture, left-side aperture and right-side aperture, is reported in Table 13.1. 369 Advances in Optics: Reviews. Book Series, Vol. 3 =1° (>Left) ; Eff=73,4% (=11,6%) 10 1,100 0,9625 0,8250 0,6875 0,5500 0,4125 0,2750 0,1375 0 8 6 4 Y coordinate (a.u.) Y coordinate (a.u.) 10 2 2 4 6 8 X coordinate (a.u.) =0° ; =633nm; Eff= 78.9% (=7,4%) 1,100 0,9625 0,8250 0,6875 0,5500 0,4125 0,2750 0,1375 0 8 6 4 2 2 10 4 (a) 8 10 (b) =2°(>Left) ; Eff=56,4% (=23,6%) 10 1,200 1,050 0,9000 0,7500 0,6000 0,4500 0,3000 0,1500 0 8 6 4 Y coordinate (a.u.) Y coordinate (a.u.) 10 6 X coordinate (a.u.) =3°(>Left) ; Eff=35,4% (=31,9%) 1,200 1,050 0,9000 0,7500 0,6000 0,4500 0,3000 0,1500 0 8 6 4 2 2 2 4 6 X coordinate (a.u.) 8 2 10 (c) 4 6 8 10 X coordinate (a.u.) (d) Fig. 13.48. Some experimental maps of the local optical efficiency of the TS-CPC concentrator, obtained with the “laser method” (DLCM). Azimuthal angle: = 0°. Incidence angle: (a) = 0°; (b) = 1.0°; (c) = 2.0°; d) = 3.0°. (a) (b) (c) Fig. 13.49. Some simulated maps of the local optical efficiency of TS-CPC, obtained with the “laser method” (DLCM). Azimuthal angle: = 0°. Incidence angle: (a) = 0°; (b) = 1.5°; (c) = 2.5°. 370 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators 10 10 1,200 1,200 1,050 0,9000 0,7500 6 0,6000 0,4500 4 0,3000 0,1500 2 4 6 X Coordinate (a) 8 0,9000 0,7500 6 0,6000 0,4500 4 0,3000 0,1500 2 0 2 1,050 8 Y Coordinate Y Coordinate 8 10 0 2 10 4 6 X Coordinate (b) 8 10 10 1,200 1,050 0,9000 0,7500 6 0,6000 0,4500 4 0,3000 0,1500 0 2 2 4 6 X Coordinate (c) 8 1,400 8 Y Coordinate Y Coordinate 8 1,225 1,050 0,8750 6 0,7000 0,5250 4 0,3500 0,1750 2 0 2 10 4 6 X Coordinate 8 10 (d) Fig. 13.50. Some experimental maps of the local optical efficiency of TS-CPC, obtained with the “laser method” (DLCM). Azimuthal angle: = 45°. Incidence angle: (a) = 0°; (b) = 1.0°; (c) = 2.0°; (d) = 3.0°. Table 13.1. All the experimental efficiency data of the TS-CPC, obtained with the “laser method” at = 0° and 45° azimuthal angles, after integration of the local efficiency maps. (°)  (°) Eff (%) sd (%) 0 0 0 0 0 0 0 45 45 45 45 45 0 0.5 1.0 1.5 2.0 3.0 4.0 0 1.0 2.0 3.0 4.0 78.4 75.2 73.4 63.9 56.3 35.4 25.3 75.8 72.9 45.7 29.9 22.3 21.3 19.3 21.3 28.7 32.7 37.3 32.8 23.1 27.6 39.2 36.7 33.3 Eff (%) (left) 75.6 71.8 68.0 49.9 55.6 11.7 12.5 sd (%) (left) 20.5 18.3 20.5 28.1 33,3 21.4 22.4 Eff (%) (right) 81.0 78.5 78.9 78.1 57.1 59.0 38.2 sd (%) (right) 21.7 19.6 20.9 21.2 32.1 35.2 36.3 371 Advances in Optics: Reviews. Book Series, Vol. 3 A look at the maps of Fig. 13.48 immediately gives a lot of information. Fig. 13.48a shows that the TS-CPC efficiency is not perfectly homogeneous, unlike what we would expect from Fig. 13.49a. The right side of the map is slightly more efficient than the left side, then the right-side mirror film was realized better. The central region of map is considerably inefficient. This is due to a wrong mechanical shaping of the prism surface near the bottom of the CPC, and to an overlapping of film strips near to the exit aperture. Is also visible the effect of joining the two portions of the prism by the less efficient vertical row at the center of the CPC. The average efficiency of the TS-CPC at  = 0° is 78.9 %, to be compared to the theoretical 94.9 % efficiency calculated by simulation the DCM with TracePro, with the assumption of a 95 % wall reflectivity (see Fig. 13.17b). The simulated results for  = 0° give an average number of reflections of light rays inside the TS-CPC of about one, as the loss of flux at each reflection is 5 % and ~5 % was the total calculated loss. The real prototype shows a real efficiency far smaller. This fact demonstrates the unsuitability of the manually surface coating process, despite the high reflectivity of 3M film. At increasing incidence angle at  = 0°, the development of optical maps is quite clear. Both left and right side of input aperture show a decreasing efficiency, but a remarkable loss of efficiency is observed on the left side of the input aperture map, because there the angle of impact of rays on the left side wall decreases (with respect to normal direction), with the consequent increase of the rejection probability for the rays (see Appendix 13.B). At  = 3° incidence, the loss of rays on left side aperture is remarkable, indicating that we are in the proximity of the acceptance angle. In practice, the left side is efficient only at  = 11.7 % (see Table 13.1). At 4° incidence, a large part of input window is not efficient at all. Due to the strong difference found for the efficiency between left side and right side, we have plotted separately the two integral efficiencies (see Fig. 13.51a). Relative efficiency (%) Relative Efficiency (%) 80 60 40 Right side Left side 20 0 TS-CPC, Incidence>Left, Azimuth = 0° 100 100 0,0 0,5 1,0 1,5 2,0 2,5 Angle of incidence (°) (a) 3,0 3,5 4,0 80 60 40 Absolute Relative 20 0 0 acc = 2.8° 1 2 3 Angle of incidence (°) 4 (b) Fig. 13.51. Experimental integral optical efficiency curves TS-CPC for 0° azimuthal angle. (a) Relative integral efficiency calculated separately for the left side and the right side of input aperture; (b) Absolute and relative integral efficiency for the entire input aperture. 372 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators After 1° incidence, the two efficiency curves sensibly diverge. At  = 3.0°, near the acceptance angle, there is the maximum difference between them. We note incidentally a sort of on/off optical switch behavior for light entering in the TS-CPC at ~3° incidence: “on” for light impinging on the right side, “off” for light impinging on the left side. The integrated efficiency data of Table 13.1 give rise to the angle-resolved optical efficiency curves, absolute () and relative rel (), for = 0° azimuth and  = 633 nm wavelength 50 90 (see Fig. 13.51b). From them we derive the acceptance angles  acc = 2.8°, acc = 1.0°. The experimental efficiency is clearly affected by several factors: imperfections of TSCPC wall shape; imperfections of 3M coating surface and loss of light by scattering; wall reflectivity of dependent on incident angle of impinging rays, photodetector response dependent incidence angle of collected rays; laser characterization realized on a matrix of points of the input aperture (input aperture area not fully illuminated). The experimental local efficiency maps of TS-CPC obtained by measurements at the azimuthal angle  = 45° show a behavior like that of the  = 0° maps (see Fig. 13.50 and Table 13.1). By integrating these maps of local optical efficiency, we obtained a curve of efficiency with the following acceptance angles: acc ~ 2.3° at 50 % efficiency and acc ~ 1.2° at 90 % efficiency. 13.6.1.2. Beam Exit Angle Measurements The lengthy manual measurements suggested us to draw only maps of azimuthal and polar angle at = 0° incidence angle at the input of TS-CPC (laser beam parallel to the z optical axis) (see Fig. 13.45). Each point of the TS-CPC input aperture is represented by a value of the correspondent angle. The maps of Fig. 13.52 are very interesting for the beauty of the representation. Both have a well-defined symmetry. The polar angle map (see Fig. 13.52a) shows that rays incident on the periphery of input aperture (ia) exit from the TS-CPC with a small divergence. The divergence increases at decreasing the distance between impact point and center of (ia). The regularity of this result is assured by the fact that the number of reflections is virtually one for any point of (ia), as results also by simulations with TracePro. The central region of polar map is rather confused. Here the beam spot on the screen (hg) is very dispersed and so impossible to measure. The cause is the same affecting the efficiency maps at the central region (see Fig. 13.48a and 13.50a): the wrong shaping of the prism surface near the bottom of TS-CPC, and the pronounced overlapping of film strips near the exit aperture, which produce a strong scattering of the beam towards unpredictable directions. Fig. 13.52b shows that the azimuthal angle also varies very regularly with the coordinate of input beam. It is easy to note that at an entrance point on (ia) corresponds an exit direction opposite with respect to the center of aperture. Then, if we let the input ray move on a circle clockwise around the center, the exiting azimuth will regularly increase and the polar angle will remain constant. At the center of the map we note the same irregularities as for the polar angle map. 373 Advances in Optics: Reviews. Book Series, Vol. 3 Another way of describing the change of direction of the exit beam is to draw its path on a planar screen as function of the y coordinate of a line scanned on the input aperture. An example is shown in Fig. 13.53, where horizontal lines are scanned with a 1 mm diameter laser beam on the (ia) of the TS-CPC at different y coordinates, with steps of 4 mm. Only the upper half of TS-CPC has been explored, starting from y = 4.8 cm and going down to y = 0.4 cm. The path of exit beam is ring shaped, has an azimuthal excursion less than , and its divergence increases progressively when the y coordinate of scanned line approaches the center of input aperture. 10 400,0 350,0 300,0 250,0 200,0 150,0 100,0 50,00 0 8 6 4 2 2 4 6 8 X coordinate (a.u.) Y coordinate (a.u.) Y coordinate (a.u.) 10 10 80,00 8 70,00 60,00 50,00 6 40,00 30,00 4 20,00 10,00 2 0 2 (a) 4 6 8 X coordinate (a.u.) 10 (b) Fig. 13.52. Input aperture maps of polar (a) and azimuthal (b) angles of the laser beam exiting from the output aperture of the TS-CPC, at = 0° incidence angle. (a) (b) (c) Fig. 13.53 (a-c). Optical path of exit beam as obtained scanning different horizontal lines on the input aperture of the TS-CPC. (a) y = 4.8 cm; (b) y = 4.4 cm; (c) y = 4.0 cm. 374 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators (d) (e) (f) (g) (h) (i) (l) (m) (n) Fig. 13.53 (d-n). Optical path of exit beam as obtained scanning different horizontal lines on the input aperture of the TS-CPC. (d) y = 3.6 cm; (e) y = 3.2 cm; (f) y = 2.8 cm; (g) y = 2.4 cm; (h) y = 2.0 cm; (i) y = 1.6 cm; (l) y = 1.2 cm; (m) y = 0.8 cm; (n) y = 0.4 cm. 375 Advances in Optics: Reviews. Book Series, Vol. 3 13.6.1.3. Optical Efficiency by DCM and ILM Fig. 13.54a shows the irradiance map on the planar screen, obtained by simulating the inverse method (at that time ILLUME, applied by illuminating by a laser beam a lambertian diffuser placed at the exit aperture of the concentrator) with a beam (ib) of 5-mm cross section radius. Fig. 13.54b shows the average x/y profile at the center of the screen. The HT-CPC was first investigated by simulating its relative transmission efficiency with the DCM and ILM methods. A wall reflectivity of 0.9 was used for both simulations. Fig. 13.54c shows the comparison between the results of relative efficiency (DCM) plotted as function of the polar angle of incidence of the collimated beam, and those of the relative radiance (ILM) plotted as function of the emission angle of inverse light (azimuthal angle  = 0°). Note that in Fig. 13.54c the DCM curve contains less points than the ILM curve, because DCM is a step-by-step process made manually at the computer, with intentionally smaller steps at around the acceptance angle, where the slope of the curve is higher. The number of points in the ILM curve, on the other hand, depends on the number of inversed rays used and then can be increased as desired. The two curves coincide almost perfectly, and give the following results: (a) (b) (c) Fig. 13.54. (a) Irradiance map of the TS-CPC concentrator, obtained by simulating the inverse method with a beam (ib) of 5-mm cross section radius. (b) Average x/y profile at the center of the screen. (c) Comparison between the angle-resolved relative efficiency (DCM) and relative radiance (ILM) of the TS-CPC. 376 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators DCM: acc (90°) = 1.3°, acc (50°) = 1.9°; ILM: acc (90°) = 1.3°, acc (50°) = 1.8°. This result was obtained thanks to the fact that, in ILM simulations, we have placed the screen at a very high distance from the TS-CPC, obtaining an angular resolution   0°. The results of Fig. 13.54 confirm the equivalence of the DCM and ILM methods for what concerns getting the relative transmission efficiency. The experimental measurements were made by applying the ILM method. Fig. 13.55a shows the TS-CPC apparatus illuminated by the lambertian light from an integrating sphere, and Fig. 13.55b shows the inverse light projected on the screen 360 cm far apart. At this distance, we have an angular resolution on the screen at the optical axis   0.8°. Fig. 13.56a shows the obtained ILM experimental transmission efficiency curve. It is rather deformed, due to the poor angular resolution, and is like that of the laser method (see Fig. 13.56b), with the same acceptance angles. Fig. 13.56b shows, by comparison, the simulated ILM curve of Fig. 13.54, which surely represents the true transmission efficiency of the TS-CPC. (a) (b) Fig. 13.55. Photo of the ILM apparatus and image of the inverse light on the screen. (a) (b) Fig. 13.56. (a) Experimental efficiency curve (relative) obtained by ILM. (b) Comparison between the three simulated and experimental relative efficiency curves. 377 Advances in Optics: Reviews. Book Series, Vol. 3 13.6.1.4. Local Optical Efficiency by ILLM In the Introduction, we have seen that the ILM method becomes a powerful tool of investigation of the optical properties of a concentrator if the receiver is locally illuminated by inverse light, and this is the ILLM method [44]. In this way in fact we are able to derive the relative efficiency curve rel() specific of the illuminated region of the receiver, and to establish the range of incidence angles, in “direct irradiation”, at which that region collects light (is illuminated) by a beam at input with less or more efficiency. A demonstration of this procedure is given here simulating the inverse illumination of the center of the diffuser of the TS-CPC by a collimated beam with variable cross section (here we apply the old ILLUME method, where the inverse lambertian light is created by illuminating, by a parallel and centered beam, a lambertian diffuser placed close to the exit aperture of the CPC (see Fig. 13.57b)). The irradiance profiles, E ( d , x ) , E ( d , y ) , recorded on a squared screen of 2000-cm side placed very far from the TS-CPC (3000 cm), have been averaged and are reported in Fig. 13.58 for different values of the beam cross section radius. The x and y directions correspond to the TS-CPC input aperture edges (see Fig. 13.57a). The E (d ) profiles of Fig. 13.58 are shown at increasing beam cross section radius R. The higher radius (5 mm) is that required to illuminate the entire diffuser surface. In this case, we obtain the same irradiance profile from which the relative efficiency rel() of Fig. 13.54c was derived, by applying Eq. (13.C5). The angular interval spanned by the profiles is 18.4°. Fig. 13.57. (a) CAD model of the TS-CPC. (b) Example of raytracing by TracePro of the truncated and squared CPC (TS-CPC), illuminated in the inverse way by a collimated inverse beam (ib) of radius cross section R. The (ib) is evidenced in blue color. It is interesting to note that reducing the beam cross section the efficiency curve reduces in width and increases in height (about twice); as consequence, also the acceptance angle is reduced. This means that concentrating the inverse beam at exit aperture towards the center of the diffuser (of the CPC receiver) produces a direct beam at input aperture more aligned with the optical axis. This also means something well known from the science of nonimaging optics, that is a beam well collimated with the optical axis of the CPC (= 0°) produces an intense irradiation in the very center of the receiver [17]. This is demonstrated here by the simulation of Fig. 13.59, where it is shown the irradiance profile produced on the receiver of the TS-CPC by a plane wave aligned with the optical axis z (= 0°). 378 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators (a) (b) (c) (d) (e) (f) Fig. 13.58. Irradiance profiles on the planar screen obtained by simulating (20k rays) the local inverse method (ILLM) for the TS-CPC concentrator. Input beam radius: (a) R = 0.05 mm (scale 10 W/m2); (b) R = 0.5 mm (scale 10 W/m2); (c) R = 1.0 mm (scale 10 W/m2); (d) R = 2.5 mm (scale 10 W/m2); (e) R = 3.5 mm (scale 7 W/m2); (f) R = 5.0 mm (scale 5 W/m2). (a) (b) Fig. 13.59. (a) Simulated irradiance map on the 1-cm diameter receiver of TS-CPC, illuminated by a plane wave aligned with the optical axis. (b) Irradiance profiles measured along x and y axes (see Fig. 13.57a). 379 Advances in Optics: Reviews. Book Series, Vol. 3 The thinning of the central profile in Fig. 13.58 is accompanied to an interesting result, that is the appearing of two satellite peaks at 7.2°. This angle corresponds to the angle  shown in Fig. 13.57b, made with z axis by a ray tangent to the upper edges of the TS-CPC exit and entrance apertures, along x or y directions. Angle  is also the minimum angle at which the square aperture starts to shadow the circular receiver respect to an incoming direct beam. To study the optical efficiency of regions of the receiver (diffuser) placed at different distance from the centre, we have simulated the inverse illumination of the TS-CPC diffuser with collimated beams having a constant cross section ( mm2), but the shape of annulus, with Rint and Rext internal and external radius, respectively. The irradiance profiles E ( d , x ) and E ( d , y ) have been recorded on the same squared screen of 2000 cm side at 3000 cm from the TS-CPC, as in the previous simulations. They have been averaged and are reported in Fig. 13.60 at increasing values of the internal radius Rint. (a) (b) (c) (d) (e) (f) Fig. 13.60. Irradiance profiles obtained by simulating (20k rays) the inverse local method ILLM for the TS-CPC concentrator. The profiles were taken along x and y axis directions, for different values of the internal radius Rint of the annulus (the external radius is in parenthesis) with constant area  mm2. (a) Rint = 0.5 (1.12) mm (scale 10 W/m2); (b) Rint = 1.0 (1.414) mm (scale 10 W/m2); (c) Rint = 2.0 (2.245) mm (scale 10 W/m2); (d) Rint = 3.0 (3.165) mm (scale 10 W/m2); (e) Rint = 4.0 (4.123) mm (scale 5 W/m2); (f) Rint = 4.9 (5.0) mm (scale 3 W/m2). The Rint = 0.0 (1.0) mm profile has not been shown as it corresponds to that with R = 1.0 mm of Fig. 13.58. 380 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators The shape of the profiles in Fig. 13.60 changes quite noticeably at high values of Rint, when the illuminated region is more peripherical. Here we note the disappearing of a central peak and the formation of two strong satellite peaks near  = 0°. This is easily explained by the same argument used in occasion of the previous simulations, that is the direct illumination of the TS-CPC by a z axis aligned beam is not able to illuminate the periphery of the receiver, as the profile of Fig. 13.59b demonstrates. Then, if we illuminate the periphery of the receiver, we cannot extract from the input of the TS-CPC a beam aligned with the z axis. As consequence, the optical efficiency at near  = 0° for the peripheral region must be very small (see Fig. 13.60f). 13.6.2. The (Virtual) Half-Truncated CPC (HT-CPC) 13.6.2.1. Local Optical Efficiency by ILLM We describe here the optical properties of the Half-Truncated CPC (HT-CPC) virtual concentrator introduced in Section 13.3 [44]. The HT-CPC is similar to the TS-CPC in the dimensional parameters, but the different shape of entrance aperture, in particular the presence in TS-CPC of the four planar surfaces (see Fig. 13.57a), renders its response to inverse method very different. As for the TS-CPC, we have simulated the inverse illumination of the center of the diffuser, placed at the exit aperture of HT-CPC, by a beam with variable cross section (ILLUME method). The screen (ps) is of 1000-cm diameter at 3000-cm distance from the HT-CPC entrance window (ia). The irradiance profiles, E(d, x) and E(d, y), at the center of the planar screen (ps) have been averaged and reported in Fig. 13.61 for different values of the beam cross section radius R. They show the existence of satellite peaks similar to those obtained with the TS-CPC (see Fig. 13.58). The profiles of Fig. 13.61 span an angle of incidence of 9.5°, and the more pronounced satellite peaks are centered at 4.7°. Differently from those of the TS-CPC, they change drastically in shape at increasing the beam cross section radius, and the intensity decreases sevenfold. Increasing beam radius, the satellite peaks, well visible in Fig. 13.61 at R = 0.05 mm disappear at R ~ 0.25 mm, leaving a broad background; the central peak decreases progressively up to disappearing, and the broad peak transforms in the large and top-flat peak typical of ideal CPCs (see Fig. 13.12 and Fig. 13.14). This means that, even halved, an ideal CPC maintains quite unchanged its optical behavior. The simulation with R = 5 mm (see Fig. 13.62), gives the relative optical efficiency of the 90 50 =4.5° and acc =5.1°, practically the HT-CPC concentrator, with acceptance angles acc same value of the not halved CPC. The irradiance profiles on the planar screen (ps) relative to the inverse illumination of HT-CPC (ILLUME) with beams of constant cross section area (π mm2) and shape of annulus are reported in Fig. 13.63. These profiles differ strongly from the correspondent ones of TS-CPC concentrator. This demonstrates that, even though similar dimensionally, the TS-CPC and HT-CPC concentrators show a strong difference in optical behavior due to the different shape of the input aperture and of the internal wall. 381 Advances in Optics: Reviews. Book Series, Vol. 3 (a) (b) (c) (d) (e) (f) Fig. 13.61. Irradiance profiles obtained by simulating the inverse method for the HT-CPC. The profiles were taken at different cross sections of the input beam (ib), centered on the diffuser. (a) R = 0.05 mm (scale 0.4 W/m2); (b) R = 0.5 mm (scale 0.2 W/m2); (c) R = 1.0 mm (scale 0.12 W/m2); (d) R = 2.5 mm (scale 0.08 W/m2); (e) R = 3.5 mm (scale 0.06 W/m2) (f) R = 5.0 mm (scale 0.05 W/m2). (a) (b) Fig. 13.62. (a) Irradiance map obtained by applying the ILLUME method to the HT-CPC with a collimated beam of 5.0 mm radius incident on the center of the diffuser. The 1000-cm diameter and 3000 cm distant screen (ps) spans an angle of incidence of 9.5°; (b) The average x/y irradiance 50 = 5.1°. profile. The average value of acceptance angle is acc 382 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators (a) (b) (c) (d) (e) (f) Fig. 13.63. Irradiance profiles obtained by simulating (50k rays) the inverse method (ILLUME) for the HT-CPC concentrator (50k rays). The profiles were obtained by averaging the profiles taken along x and y directions, for different values of the internal radius Rint of the annulus (the external radius is in parenthesis). (a) Rint = 0.5 (1.12) mm (scale 0.1 W/m2); (b) Rint = 1.0 (1.414) mm (scale 0.1 W/m2); (c) Rint = 2.0 (2.245) mm (scale 0.1 W/m2); (d) Rint = 3.0 (3.165) mm (scale 0.06 W/m2); (e) Rint = 4.0 (4.123) mm (scale 0.06 W/m2); (f) Rint = 4.9 (5.0) mm (scale 0.05 W/m2). 13.6.3. The Truncated CPC (T-CPC) 13.6.3.1. Optical Efficiency by ILM Fig. 13.64a shows T-CPC during measurements at inverse light with the ILM method. Distance T-CPC-screen: 360 cm. Fig. 13.64b shows the T-CPC in the dark. The image on the screen is viewed through the mirror (mi), used because the angle view of the CCD objective was too small. Fig. 13.65a and Fig. 13.65b show the use of the HiPic software (Hamamatsu) to trace the horizontal and vertical profiles of the irradiance map, respectively. The relative, horizontal and vertical, inverse radiance profiles are calculated by using Eq. (13.65) and are shown in Fig. 13.66a and Fig. 13.66b, respectively. We find 90 50 = 0.75°, acc = 1.47°. the following, average acceptance angles: acc 383 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 13.64. (a) Photo of the T-CPC concentrator during measurements with inverse light. It is visible the lamp (lp), the integrating sphere (is) and the planar screen (ps). (b) The T-CPC in the dark, with the CCD oriented towards the mirror (mi), reflecting the image on the screen and used to double the screen-CCD distance (see Fig. 13.34). The CCD is shown in the box. (a) (b) Fig. 13.65. Horizontal (a) and vertical (b) profiles of the irradiance map, taken by HiPic. (a) (b) Fig. 13.66. Horizontal (a) and vertical (b) profiles of the relative radiance. 384 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators 13.6.4. The Rondine Concentrators 13.6.4.1. Optical Efficiency by DCM and ILM Fig. 13.67a shows the Rondine-Gen1 concentrator [59-66] during measurements at inverse light with the ILM method. Distance T-CPC-screen: 360 cm. Here we show the economical configuration of ILM experimental setup, where it is used a single sphere, a webcam to record the inverse image and the “ImageJ” software (free of charge from the web) for its elaboration. Later we used a CCD camera and the HiPic software of Hamamatsu (see Fig. 13.67b). The x-axis and y-axis irradiance profiles of the image of inverse light on the screen, taken by HiPic, can be found in Fig. 13.18a. Fig. 13.68a shows the irradiance ILM map and Fig. 13.68b shows the horizontal (y) and vertical (x) profiles of the relative inverse radiance of Rondine Gen1, calculated by using Eq. (13.65). The two profiles have a different trend due to the quasi-rectangular shape of the exit aperture (see Fig. 13.18a). We find the following average acceptance angles: x-axis profile, 90 50 90 50 = 4.3°, acc = 9.5°; y-axis profile, acc = 6.3°, acc = 9.5°. These values can be acc compared to that obtained by DCM (see Fig. 13.69). The corresponding average 90 50 90 = 4.4°, acc = 9.1°, y-axis profile acc = 6.5°, acceptance angles are: x-axis profile  acc 50 = 9.6°, practically the same found with ILM. For the orientation of the Rondine, refer acc to Fig. 13.18a. The horizontal or vertical orientations are not related to the x/y axes, but on the way the Rondine is mounted on its holder. We observe that the direction of longer side of exit aperture (x-axis) gives rise to a shorter acceptance angle at 90 % of maximum efficiency. Fig. 13.70a shows the irradiance ILM map and Fig. 13.70b shows the horizontal x/y profiles of the relative inverse radiance of Rondine-Gen2, calculated by using Eq. (13.65). The square symmetry of the exit aperture of Gen2 gives rise to the same 90 = 5.0°, profile for the x-axis and y-axis, with the following average acceptance angles: acc 50 = 8.0°. acc Fig. 13.67. (a) A single integrating sphere and a webcam were used in the first, economical, experimental setup of ILM applied to the Rondine-Gen1 concentrator. (b) Two integrating spheres and a CCD were used later in the more advanced experimental setup of ILM. 385 Advances in Optics: Reviews. Book Series, Vol. 3 1,1 rel rel LC () , opt () 1,0 0,9 acc=6.3° acc=-6.3° acc=-4.4° acc=+4.2° 0,8 0,7 acc=9.5° 0,6 acc=-9.5° 0,5 horizontal profile acc=6.3° 0,4 vertical profile 0,3 -10 (a) -5 0 acc=4.3° 5 Emission / incidence angle,  (°) 10 (b) Fig. 13.68. (a) Irradiance ILM map of the Rondine-Gen1; (b) Horizontal (y-axis) (black) and vertical (x-axis) (red) profiles of the relative inverse radiance. (a) (b) Fig. 13.69. (a) Relative efficiency of the Rondine Gen1 concentrator traced along the x-axis. (b) Relative efficiency of the Rondine Gen1 concentrator traced along the y-axis. A small distortion of the curves is observed, due to a not perfect uniformity of the parallel beam. We have already explained (see Section 13.4) that the application of DCM to the Rondine concentrators requires the adding of a box at the entrance side to recover the removed four planar walls (see Fig. 13.19), whereas the application of ILM can be made on the Rondine as it is. To demonstrate that the presence of the box do not alter the optical properties of the Rondine at inverse light, we have simulated by ILM also the Rondine-Gen2+box concentrator, with ideal mirror box walls. The measured relative inverse radiance is shown in Fig. 13.71a for the two configurations. The perfect overlap of the two curves demonstrates the non-influence of the box on the optical properties of the Rondine, so the idea of Antonini to remove the four walls proved to be very correct [59-66], because, also in the best real case, the four walls would never be equivalent to two ideal reflectors. 386 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators 1,1 rel rel LC () , opt () 1,0 0,9 acc =+4.9° 0,7 0,6 0,5 0,4 50 0,2 acc =8.2° 50 acc =-7.8° horizontal profile vertical profile 0,3 (a) 90 90 acc =-5.1° 0,8 -8 -6 -4 -2 0 2 4 Emission / incidence angle,  (°) 6 8 (b) Fig. 13.70. (a) Irradiance ILM map of the Rondine-Gen-2. (b) Horizontal (black) and vertical (red) profiles of the relative inverse radiance. (a) (b) Fig. 13.71. (a) Simulated inverse radiances vs. emission angle of Rondine-Gen2, with and without box, averaged over the x and y directions, with the ILM method at d = 229 cm. Angular resolution =1.0°; (b) Average optical efficiency profile measured along x/y axes, compared to that measured along the diagonal of input aperture. Being the Rondine-Gen2 concentrator of non-cylindrical symmetry, we have proven to look at its transmission efficiency curve when the azimuthal angle is of  = 45°. Whereas this requires new simulations measurements when applying the DCM, the ILM immediately gives this information; it is sufficient in fact to take the map of irradiance (see Fig. 13.70a), to extract the data occupying the diagonal row of the map, to calculate the correct values for the polar angle, and finally to transform them in terms of relative radiance (efficiency) by applying Eq. (13.65). The result is shown in Fig. 13.71b and compared with the curve taken at  = 0°. 387 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 13.71b leads us to another consideration. The Rondine-Gen2 module is actually oriented in such a way to track the Sun disk along the x or y-axis (they are equivalent in Gen2). Angular misalignments between Gen2 and Sun are then expected along that direction. The curves of Fig. 13.71b, however, seem to favor the diagonal direction, instead of the x/y one. If we calculate the inverse radiance curve along the diagonal direction, indeed, we find that it is wider than that calculated along the x/y axis. We have 90 50 = 5.0±0.2°; acc = 8.4±0.2°; and in the diagonal direction: in fact in the x/y direction: acc 90 50 = 4.9±0.2°; acc = 9.9±0.2°. Then, to increase the collection capability of the acc Rondine-Gen2 along the direction of tracking, it could be convenient to have the tracking direction parallel to the diagonal of the opening entrance. This is valid for misalignments 90 50 90 , because, as it can be seen in Fig. 13.71b, acc is improved of ≈ 20 %, but acc > acc remains almost constant. 13.6.4.2. Optical Efficiency by Parretta-Method and Parretta-Herrero Method Here we compare the results of optical simulations carried out on the Rondine concentrators by applying two different inverse methods, the ILM, or Parretta Method, and the Parretta-Herrero Method. I remember that the first one is applied by projecting the inverse light on a far screen, as an ideal absorber, whereas the second is applied by interposing a parabolic mirror between the Rondine and the screen, which focus the inverse light on the ideal absorber screen, where, differently from ILM, the polar angle of a point is displayed linearly with the distance from the center. Here we demonstrate that the P-Method (ILM) and the Parretta-Herrero Method give the same results if both are applied correctly [37-45]. The flux distributions (irradiances in W/m2) recorded for the P-Method and the PHMethod applied to Rondine-Gen1 are shown in Figs. 13.72a and b, respectively. The map of Fig. 13.72b directly reproduces the radiance of the inversely irradiated concentrator, because of the reflection of the inverse light on the parabolic mirror surface. From the two flux distributions we derive the corresponding x-axis and y-axis profiles of the normalized transmission efficiency as follows: for the P-Method the normalized x/y-axes profiles of the intensity (irradiance) on the screen were multiplied by the (cosθ)−4 factor (see Eq. (13.65)); for the PH-Method, on the contrary, they were directly obtained from the normalized x/y-axes profiles of the intensity (irradiance) on the screen. The x-axis and y-axis profiles of the transmission efficiency, ηdir(θ), derived by the two methods, are shown in Figs. 13.73a and b, respectively. The maps of flux distribution recorded for the P-Method and the PH-Method applied to the Rondine-Gen2 are shown in Figs. 13.74a and b, respectively. The corresponding x-axis and y-axis profiles of the normalized transmission efficiency, ηdir(θ), are shown in Fig. 13.75a and b, respectively. Comparing the efficiency profiles of both concentrators (Figs. 13.73 and 13.75), the excellent overlap of the profiles obtained with the two methods appears evident, and hence the equivalence between P-Method and PH-Method. 388 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators (a) (b) Fig. 13.72. Rondine-Gen1: (a) map of flux distribution recorded on the screen (ps) for the P-Method; (b) map of flux distribution recorded on the screen (ps) for the PH-Method. (a) (b) Fig. 13.73. Rondine-Gen1: x-axis (a) and y-axis (b) of the normalized optical efficiency profiles derived by the P-Method and the PH-Method. We are finally able to give a value to the on-axis absolute efficiency,  dir (0) , by experimental measurements of the radiance emitted frontally by the Rondine concentrators. Applying the Eq. (13.68) in Section 13.5.2: dir (0)  Linv (0) / LREC , where Linv (0) is the average radiance of the on-axis front image of input aperture, and LREC is the average radiance of the lambertian source (the integrating sphere). Fig. 13.76a shows the front image of the Rondine-Gen1 concentrator. It is visible the central, blue frame, region of the receiver (the window of the integrating sphere). The radiance LREC is slightly underestimated due to the presence of the baffle at the center (see Fig. 13.39b). By correcting this effect [A. Parretta, to be published], we obtain a better estimate of LREC , 389 Advances in Optics: Reviews. Book Series, Vol. 3 which, together with the average radiance of the red frame region, Linv (0) , gives: dir (0) = 84.0 ± 1.0 %. For the Rondine-Gen2 we obtain: dir (0) = 86.0 ± 1.0 %. The experimental values of dir (0) , together with those of  dir , rel ( ,  ) found so far by the DCM and ILM methods, give the final absolute transmission efficiency  dir ( ,  ) of the concentrators. The more advanced Rondine-Gen2 version proves therefore more efficient than the old Rondine-Gen1 version. (a) (b) Fig. 13.74. Rondine-Gen2: (a) map of flux distribution recorded on the screen (ps) for the P-Method; (b) map of flux distribution recorded on the screen (ps) for the PH-Method. (a) (b) Fig. 13.75. Rondine-Gen2: x-axis (a) and y-axis (b) of the normalized optical efficiency profiles derived by the P-Method and the PH-Method. 390 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators (a) (b) (c) Fig. 13.76. Front image of the Rondine-Gen1 concentrator. The radiance LREC is obtained averaging the intensity of the blue frame region; the radiance Linv (0) is obtained averaging the intensity of the red frame region (a). On-axis image of the front aperture of the Rondine-Gen1 (b). Off-axis (right side) image of the front aperture of the Rondine-Gen1. The appearance of a darker zone on the left side means that the optical efficiency is reducing (c). 13.6.5. The PhoCUS Concentrator 13.6.5.1. Optical Simulations The focalization properties of the two lenses (prismatic and hybrid), without SOE, were simulated by TracePro with 100k rays. Fig. 13.77 shows the images produced at the focal distance of the two lenses [57]. Only the prismatic lens gives a uniform distribution of flux on the focal plane. To obtain a uniform flux on the receiver with both lenses, the distance lens-receiver was increased to 23 cm, greater than the focal length. 391 Advances in Optics: Reviews. Book Series, Vol. 3 (a) (b) Fig. 13.77. Simulated maps of irradiance at the focal plane of the prismatic (a) and hybrid (b) PhoCUS lenses, without SOE. The chromatic aberration introduced by the prismatic lens has been studied by simulating, with the direct method, the image produced on the receiver by blue light ( = 450 nm) and red light ( = 650 nm). The results are visible in Fig. 13.78. Being the receiver behind the focal plane, the short wavelengths are more expanded on the receiver surface. (a) (b) Fig. 13.78. Images of the flux on the receiver plane, at 23 cm from the lens, due to  = 450 nm light (a) and  = 650 nm light (b). The optical efficiency curves, simulated by DCM and ILM, of the two lenses (without SOE) are shown in Fig. 13.79a. First of all, we observe that the curves are typical of an “imaging” concentrator, with a monotonic decrease of efficiency, in contrast to “nonimaging” concentrators, characterized by a plateau followed by a rapid decrease at 50 90  1.5÷2.0° and acc  0.5°. We observe also that: the acceptance angle. We measure acc i) the two lenses have similar curves for the same method; ii) each method gives similar 392 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators curves for the two lenses. The inverse method ILM has been also applied to simulate the PhoCUS mock-up, that is the lens + SOE, and the relative efficiency curves for the two types of lens are shown in Fig. 13.79b. The effect of SOE is positive, because it increases 50 90  2.4° and acc  1.0° for both lenses and equalizes the two the acceptance angles: acc curves. We now observe a behavior more similar to a nonimaging solar concentrator. (a) (b) Fig. 13.79. (a) Optical efficiency of the prismatic and hybrid lenses (no SOE), simulated by DCM and ILM. (b) Optical efficiency of the prismatic and hybrid lenses + SOE, simulated by ILM. 13.6.5.2. Experimental Measurements Some experimental results are shown in the following. (a) (b) Fig. 13.80. (a) Comparison between the experimental and simulated direct efficiency curves (DCM) of the prismatic lens without SOE; (b) Experimental direct efficiency curves of the hybrid lens without SOE, obtained by using a sphere or a solar cell as receiver. 393 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 13.80a shows that coupling the PhoCUS mock-up to the integrating sphere (see Fig. 13.30c), gives the optical properties of the concentrator alone, as obtained by simulations, confirming that the integrating sphere behaves as an ideal receiver. The use of the solar cell, on the contrary, gives a thinner curve, showing that the solar cell is not able to collect with the same efficiency all the incident light, particularly that incident at input at higher angles. The knowledge of difference between the sphere and the cell curve is important to establish how much the cell could be optimized to improve its light collection. Similar results were obtained with the hybrid lens, as shown in Fig. 13.80b. 13.7. Conclusions Purpose of this work has been to provide the experimentalist with all the theoretical and practical tools useful to investigate the main optical properties of individual units of a concentrated PV system, in order to confirm the overall characteristics of the original project. The subject of this work has been the description of methods for the optical characterization of solar concentrator prototypes, which involves the study of many properties, the most important of which are the optical transmission efficiency and the spatial and angular distribution of the flux on the receiver, the solar cell. However, to give this work a wider breath, the practical treatment of the characterization methods was preceded by a general theoretical discussion, which examines all the possible irradiation modes of a solar concentrator, schematized as a box with an entrance and an exit aperture, and a completely unknown interior. This discussion has allowed me to deepen theoretically the traditional "direct irradiation” model, that reflecting the solar concentrator operating mode in the field, and to consolidate the theory of innovative models, recently introduced, those based on the "inverse" (or "reverse") irradiation of the concentrator. I have proposed also new ones, though a bit bizarre, accompanied by the definition of new optical quantities, which could be useful in the future. The second part of this work has been devoted to the operating modes to follow in order to correctly apply two main characterization methods, namely the "direct" and the "inverse" ones. The optical quantity on which I have been most concerned was the optical transmission efficiency, which is represented by a curve that defines how efficient the light transmission is from the input to the output opening of the concentrator, as function on the orientation of the concentrator respect to a parallel light beam, simulating the direct sunlight component. In this part I have highlighted all the advantages of using the "reverse" method to derive the optical efficiency curve. The advantages are innumerable and they are both in optical simulation and in laboratory testing. During optical simulation, the method setting is simple and requires few operations. This is to set up a Lambertian source on the concentrator output opening, to place an absorbing screen on the side of the input opening, and sufficiently far from the concentrator, and to start raytracing, which can also last for hours, depending on the needed angular resolution on the curve. At the end of raytracing, the file of irradiance on the screen is converted to the reverse "radiance" file, which corresponds (this is the heart of the inverse method) to the "direct" optical efficiency file, containing all information for all possible polar and 394 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators azimuth orientations. The simulated "direct" method, on the other hand, requires a great deal of operativeness at the computer, because an absorber must be positioned at the concentrator output, a parallel beam must be arranged at its input, oriented in a certain way, and the output flux measured. This operation must be repeated for a number (dozens) of polar incidence angles when the concentrator has cylindrical symmetry. If symmetry is not such, the previous sequence must be repeated for as many times as the azimuth angles of interest. Even in laboratory experimentation, the "inverse" method is very simple to accomplish. It is to illuminate the back of the concentrator with a lamp and one or two integrating spheres; the screen absorber is replaced by a white screen for projections and there is no need to wait. The image on the screen is recorded by a CCD and processed at the computer as before. The experimental "direct" method, on the other hand, requires the preparation, by means of a parabolic mirror, of a uniform parallel beam to be orientated towards the concentrator, and many flux measurements on the receiver by changing the orientation of the CPC from time to time. In addition to time and simplicity, the advantage of the "inverse" method is in its costeffectiveness. In fact, the CCD can be replaced by a webcam, while the "direct" method requires a very expensive parabolic mirror to obtain a uniform parallel beam of at least 1 %. The last part of this work has been devoted to the optical characterization of some solar concentrator prototypes, all nonimaging type. In particular, reflecting concentrators of the light cone type, or CPC, have been characterized, all obtained by transforming ideal CPC originals through subsequent modifications. The two characterization methods were applied and their respective results compared, highlighting the advantages and disadvantages of the two methods. The "reverse" method, however, remains the most profitable one. Among the concentrator prototypes, of great interest has been the Rondine® one, a very innovative concentrator of the light cones class, created by a spinoff company of the University of Ferrara (CPower), resulting from a sophisticated transformation of a classic CPC, which has shown remarkable results on the field. Acknowledgements This chapter of optics is the result of some years of work that I have been doing at the Physics Department of the University of Ferrara. The contribution I received from the students, the young researchers and the teaching staff of the Physics Department was decisive to get the described results. My thanks go to the late Prof. G. Martinelli, who trusted me and welcomed me to his lab, and to Ing. G. Palazzi, who promoted the fruitful collaboration between ENEA and Ferrara University. I am very grateful to Prof. F. Frontera for starting me on the didactic activity for PhD students, which was followed by the course of Applied Optics. The didactic commitment allowed me to deepen the theoretical issues encountered in the experimental work. I thank my young colleague 395 Advances in Optics: Reviews. Book Series, Vol. 3 A. Antonini for the long and fruitful collaboration, and the young researchers of the formidable CPower group, which brought to a high level the technology of nonimaging optics concentrators. I was introduced to this science by Prof. L. Vicari of Napoli University, who put the Welford and Winston's text "High collection of nonimaging optics" in my hand, obliging me to read it, and I was supported from then on by the uncontainable energy of Prof. C. Rubbia. I am finally very grateful to Prof. R. Calabrese, F. Petrucci and R. Tripiccione for the sharing of an exciting season of scientific and educational collaboration. References [1]. A. Luque, S. Hegedus, Handbook of Photovoltaic Science and Engineering, John Wiley & Sons Ltd., Chichester, UK, 2011. [2]. A. Luque, V. M. Andreev, Concentrator Photovoltaics, Springer-Verlag, Berlin, D, 2007. [3]. A. Martí, A. Luque, Next Generation Photovoltaics, Institute of Physics, Bristol, UK, 2004. [4]. V. M. Andreev, V. A. Grilikhes, V. D. Rumyantsev, Photovoltaic Conversion of Concentrated Sunlight, John Wiley & Sons, Chichester, UK, 1997. [5]. A. Luque, Solar Cells and Optics for Photovoltaic Concentrators, Institute of Physics, Bristol, UK, 1989. [6]. D. Barlev, R. Vidu, P. Stroeve, Innovations in concentrated solar power, Solar Energy Materials and Solar Cells, Vol. 95, 2011, pp. 2703-2725. [7]. T. M. Razikov, C. S. Ferekides, D. Morel, E. Stefanakos, H. S. Hullal, H. M. Upadhyaya, Solar photovoltaic electricity: Current status and future prospects, Solar Energy, Vol. 85, 2011, pp. 1580-1608. [8]. J. S. Coventry, A solar concentrating photovoltaic/thermal collector, PhD Thesis, Australian National University, Canberra, AUS, 2004. [9]. D. Mills, Advances in solar thermal electricity technology, Solar Energy, Vol. 76, 2004, pp. 19-31. [10]. S. A. Kalogirou, Solar thermal collectors and applications, Progresses in Energy and Combustion Science, Vol. 30, 2004, pp. 231-295. [11]. G. Smestad, H. Ries, R. Winston, E. Yablonovich, The thermodynamic limits of light concentrators, Solar Energy Materials, Vol. 21, 1990, pp. 99-111. [12]. A. Antonini, M. Stefancich, D. Vincenzi, C. Malagù, F. Bizzi, A. Ronzoni, G. Martinelli, Contact grid optimization methodology for front contact concentration solar cells, Solar Energy Materials and Solar Cells, Vol. 80, 2003, pp. 155-166. [13]. D. Vincenzi, M. Stefancich, S. Baricordi, M. Gualdi, G. Martinelli, A. Parretta, A. Antonini, Effects of irradiance distribution uneveness on the ohmic losses of CPV receivers, in Proceedings of the 24th European Photovoltaic Solar Energy Conference (EU PVSEC’09), 2009, pp. 725-728. [14]. E. Franklin, J. Coventry, Effects of highly non-uniform illumination distribution on electrical performance of solar cells, in Proceedings of the ANZSES Solar Annual Conference, Newcastle, Australia, 2002. [15]. R. Leutz, A. Suzuki, A. Akisawa, T. Kashiwagi, Flux uniformity and spectral reproduction in solar concentrators using secondary optics, in Proceedings of the ISES Solar World Congress (SWC’01), Adelaide, 2001, pp. 775-784. [16]. J. Chaves, Introduction to Nonimaging Optics, CRC Press Taylor & Francis Group, Boca Raton, USA, 2008. [17]. R. Winston, J. C. Miñano, P. Benítez, Nonimaging Optics, Elsevier Academic Press, Amsterdam, NL, 2005. 396 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators [18]. W. T. Welford, R. Winston, High Collection Nonimaging Optics, Academic Press Inc., San Diego, USA, 1989. [19]. J. J. O’Gallagher, Nonimaging Optics in Solar Energy, Morgan & Claypool Publishers, Lexington, KY, USA, 2008. [20]. W. T. Welford, R. Winston, The Optics of Nonimaging Concentrators, Light and Solar Energy, Academic Press, 1978. [21]. P. Sansoni, F. Francini, D. Fontani, Optical characterization of solar concentrator, Optics and Lasers in Engineering, Vol. 45, 2007, pp. 351-359. [22]. A. Parretta, Optics of solar concentrators. Part I: Theoretical models of light collection, International Journal of Optics and Applications, Vol. 3, Issue 4, 2013, pp. 27-39. [23]. A. Parretta, A. Antonini, Optics of solar concentrators. Part II: Models of light collection of 3D-CPCs under direct and collimated beams, International Journal of Optics and Applications, Vol. 3, Issue 5, 2013, pp. 72-102. [24]. A. Parretta, Optics of solar concentrators. Part III: Models of light collection of 3D-CPCs under direct and Lambertian irradiation, International Journal of Optics and Applications, Vol. 5, Issue 3, 2015, pp. 82-102. [25]. A. Parretta, A. Antonini, Optics of solar concentrators. Part IV: Maps of transmitted and reflected rays in 3D-CPCs, International Journal of Optics and Applications, Vol. 7, Issue 1, 2017, pp. 13-30. [26]. A. Parretta, A. Antonini, M. Butturi, E. Milan, P. Di Benedetto, D. Uderzo, P. Zurru, Optical methods for indoor characterization of small-size solar concentrators prototypes, Advances in Science and Technology, Vol. 74, 2010, pp. 196-204. [27]. A. Parretta, F. Aldegheri, D. Roncati, C. Cancro, R. Fucci, Optical efficiency of "PhoCUS" C-Module concentrators, in Proceedings of the 26th European Photovoltaic Solar Energy Conference (EU PVSEC’11), Hamburg, Germany, 5-9 Sept. 2011, pp. 658-663. [28]. A. Parretta, L. Zampierolo, G. Martinelli, A. Antonini, E. Milan, C. Privato, Methods of characterization of solar concentrators, in Proceedings of the 25th European Photovoltaic Solar Energy Conference (EU PVSEC’10), Feria Valencia, Valencia, Spain, 6-10 Sept. 2010. [29]. A. Parretta, L. Zampierolo, D. Roncati, Theoretical aspects of light collection in solar concentrators, in Proceedings of the Conference on Optics for Solar Energy (SOLAR), Advancing the Science and Technology of Light, The Westin La Paloma, Tucson, AZ, USA, 7-10 June 2010, p. STuE1. [30]. A. Parretta, A. Antonini, M. A. Butturi, P. Di Benedetto, E. Milan, M. Stefancich, D. Uderzo, P. Zurru, D. Roncati, G. Martinelli, M. Armani, How to "display" the angle-resolved transmission efficiency of a solar concentrator reversing the light path, in Proceedings of the 23rd European Photovoltaic Solar Energy Conference (EU PVSEC’08), Feria Valencia, Valencia, Spain, 1-5 Sept. 2008, pp. 95-98. [31]. A. Parretta, A. Antonini, M. Stefancich, V. Franceschini, G. Martinelli, M. Armani, Laser characterization of 3D-CPC solar concentrators, in Proceedings of the 22nd European Photovoltaic Solar Energy Conference (EU PVSEC’07), Fiera Milano, Milan, Italy, 3-7 September 2007. [32]. A. Parretta, A. Antonini, M. Stefancich, V. Franceschini, G. Martinelli, M. Armani, Characterization of CPC solar concentrators by a laser method, Proceedings of SPIE, Vol. 6652, 2007, 665207. [33]. E. Bonfiglioli, Sviluppo di metodi per la caratterizzazione ottica indoor di concentratori solari, Graduate Thesis in Physics and Astrophysics, Università degli Studi di Ferrara, Ferrara, Italy, 2007-2008. [34]. A. Antonini, M. Stefancich, J. Coventry, Modelling of compound parabolic concentrators for photovoltaic applications, International Journal of Optics and Applications, Vol. 3, Issue 4, 2013, pp. 40-52. 397 Advances in Optics: Reviews. Book Series, Vol. 3 [35]. A. Parretta, G. Martinelli, M. Stefancich, D. Vincenzi, R. Winston, Modelling of CPC-based photovoltaic concentrator, Proceedings of SPIE, Vol. 4829, 2003, pp. 1045-1047. [36]. A. Parretta, E. Bonfiglioli, L. Zampierolo, M. Butturi, A. Antonini, Optical characterization of Rondine® PV solar concentrators, International Journal of Optics and Applications, Vol. 4, Issue 4a, 2014, pp. 25-52. [37]. A. Parretta, A. Antonini, M. Butturi, P. Zurru, Optical simulation of Rondine® solar concentrators by two inverse characterization methods, Journal of Optics, Vol. 14, 2012, 125704. [38]. A. Parretta, A. Antonini, E. Bonfiglioli, M. Campa, D. Vincenzi, G. Martinelli, Il metodo inverso svela le proprietà dei concentratori solari, PV Technology, Vol. 3, Agosto/Settembre 2009, pp. 58-64 (in Italian). [39]. A. Parretta, A. Antonini, E. Milan, M. Stefancich, G. Martinelli, M. Armani, Optical efficiency of solar concentrators by a reverse optical path method, Optics Letters, Vol. 33, 2008, pp. 2044-2046. [40]. A. Parretta, L. Zampierolo, A. Antonini, E. Milan, D. Roncati, Theory of "inverse method" applied to characterization of solar concentrators, in Proceedings of the 25th European Photovoltaic Solar Energy Conference (EU PVSEC’10), Feria Valencia, Valencia, Spain, 6-10 September 2010. [41]. A. Parretta, D. Roncati, Theory of the "inverse method" for characterization of solar concentrators, in Proceedings of the Conference Optics for Solar Energy (SOLAR), Advancing the Science and Technology of Light, The Westin La Paloma, Tucson, AZ, USA, 7-10 June 2010, p. STuE2 [42]. A. Parretta, G. Martinelli, A. Antonini, D. Vincenzi, Direct and inverse methods of characterization of solar concentrators, in Proceedings of the Conference Optics for Solar Energy (SOLAR), Advancing the Science and Technology of Light, The Westin La Paloma, Tucson, AZ, USA, 7-10 June 2010, p. STuA1. [43]. A. Parretta, A. Antonini, M. Stefancich, G. Martinelli, M. Armani, Optical characterization of CPC concentrator by an inverse illumination method, in Proceedings of the 22nd European Photovoltaic Solar Energy Conference (EU PVSEC’07), Fiera Milano, Milan, Italy, 3–7 Sept. 2007. [44]. A. Parretta, A. Antonini, M. Stefancich, G. Martinelli, M. Armani, Inverse illumination method for characterization of CPC concentrators, Proceedings of SPIE, Vol. 6652, 2007, 665205. [45]. A. Parretta, F. Aldegheri, A. Antonini, M. Butturi, P. Zurru, Optical simulation of PV solar concentrators by two inverse characterization methods, International Journal of Optics and Applications, Vol. 2, Issue 5, 2012, pp. 62-71. [46]. R. Winston, Principles of solar concentrators of a novel design, Solar Energy, Vol. 16, 1974, pp. 89-95. [47]. A. Parretta, A. Sarno, P. Tortora, H. Yakubu, P. Maddalena, J. Zhao, A. Wang, Angledependent reflectance and transmittance measurements on photovoltaic materials and solar cells, Optics Communications, Vol. 172, 1999, pp. 139-151. [48]. C. Cancro, R. Fucci, G. Graditi, A. Romano, A. Antonini, M. Armani, Comparison between two different S.O.E. geometrical shape to increase the PhoCUS C-Module energy performances, in Proceedings of the 21st European Photovoltaic Solar Energy Conference (EU PVSEC’06), Dresden, Germany, 4-8 Sept. 2006. [49]. R. Leutz, L. Fu, H. P. Annen, Material choices and tolerance budget in the optical design of solar photovoltaic concentrators, in Proceedings of Optics for Solar Energy (SOLAR), Advancing the Science and Technology of Light, The Westin La Paloma, Tucson, AZ, USA, 7-10 June 2010, p. STuD4. [50]. R. C. Jones, On reversibility and irreversibility in optics, Journal of Optical Society of America, Vol. 43, 1953, pp. 138-144. 398 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators [51]. M. Born, E. Wolf, Principles of Optics, Cambridge University Press, Cambridge, UK, 1999. [52]. R. Herrero, C. Domínguez, S. Askins, I. Antón, G. Sala, Two-dimensional angular transmission characterization of CPV modules, Optics Express, Vol. 18, 2010, pp. A499-A505. [53]. R. Herrero, C. Domínguez, I. Antón, S. Askins, G. Sala, Tools development for CPV characterization based on CCD camera measurements, in Proceedings of Optics for Solar Energy (SOLAR), Advancing the Science and Technology of Light, The Westin La Paloma, Tucson, AZ, USA, 7-10 June 2010, p. STuA4. [54]. P. Würfel, The chemical potential of radiation, Journal of Physics C: Solid State Physics, Vol. 15, 1982, pp. 3967-3985. [55]. L. Ferraioli, P. Maddalena, E. Massera, A. Parretta, M. A. Green, A. Wang, J. Zhao, Evidence for generalized Kirchhoff's law from angle-resolved electroluminescence of high efficiency silicon solar cells, Applied Physics Letters, Vol. 85, 2014, pp. 2484-2486. [56]. K. L. Coulson, Polarization and Intensity of Light in the Atmosphere, A. Deepak Publishing, Hampton, Virginia, USA, 1988. [57]. A. Parretta, F. Aldegheri, C. Cancro, R. Fucci, G. Graditi, R. Schioppo, Optical characterization of “PhoCUS” refractive solar concentrators, International Journal of Optics and Applications, Vol. 4, Issue 4A, 2014, pp. 12-24. [58]. Lambda Research Corporation, http://lambdares.com/ [59]. A. Antonini, M. A. Butturi, P. Di Benedetto, D. Uderzo, P. Zurru, E. Milan, M. Stefancich, M. Armani, A. Parretta, N. Baggio, Rondine® PV concentrators: Field results and developments, Progress in Photovoltaics, Vol. 17, 2009, pp. 451-459. [60]. A. Antonini, M. A. Butturi, P. Di Benedetto, D. Uderzo, P. Zurru, E. Milan, M. Stefancich, A. Parretta, N. Baggio, Rondine PV concentrators: Field results and innovations, in Proceedings of the 24th European Photovoltaic Solar Energy Conference (EU PVSEC’09), Hamburg, Germany, 21-25 Sept. 2009. [61]. A. Antonini, M. A. Butturi, P. Di Benedetto, D. Uderzo, P. Zurru, E. Milan, A. Parretta, N. Baggio, Rondine PV concentrators: Field results and progresses, in Proceedings of the 34th IEEE Photovoltaic Solar Energy Conference (PVSC’09), Philadelphia, US, 7-12 June 2009, pp. 001043-001047. [62]. A. Antonini, M. A. Butturi, P. Di Benedetto, E. Milan, D. Uderzo, P. Zurru, M. Stefancich, A. Parretta, M. Armani, N. Baggio, Field testing per la tecnologia di concentratori a disco e per moduli “Rondine”, in Proceedings of the “ZeroEmission Rome 2008” Conference, Rome, Italy, 1-4 October 2008. [63]. A. Antonini, M. A. Butturi, P. Di Benedetto, D. Uderzo, P. Zurru, E. Milan, M. Stefancich, M. Armani, A. Parretta, N. Baggio, Development of "Rondine" concentrating PV module field results and progresses, in Proceedings of the 23rd European Photovoltaic Solar Energy Conference (EU PVSEC’08), Valencia, Spain, 1-5 Sept. 2008. [64]. A. Antonini, M. Stefancich, P. Zurru, M. Butturi, P. Di Benedetto, D. Uderzo, E. Milan, G. Martinelli, M. Armani, A. Parretta, N. Baggio, L. Secco, Experimental results of a new PV concentrator system for silicon solar cells, in Proceedings of the 22nd European Photovoltaic Solar Energy Conference (EU PVSEC’07), Milan, Italy, 3-7 September 2007. [65]. A. Antonini, M. Butturi, P. Di Benedetto, P. Zurru, D. Sartore, In-Field performances of first commercial Rondine® PV concentrators, in Proceedings of the 26th European Photovoltaic Solar Energy Conference (EU PVSEC’11), Hamburg, Germany, 5-9 September 2011, pp. 640-643. [66]. A. Antonini, M. Butturi, P. Di Benedetto, E. Milan, D. Uderzo, P. Zurru, D. Sartore, GEN2 Rondine® PV concentrators, in Proceedings of the 25th European Photovoltaic Solar Energy Conference (EU PVSEC’10), Feria Valencia, Valencia, Spain, 6-10 September 2010, pp. 990-993. 399 Advances in Optics: Reviews. Book Series, Vol. 3 [67]. A. Parretta, M. L. Addonizio, Optical and structural characterization of diffuse reflectance standards, International Journal of Optics and Applications, Vol. 5, Issue 2, 2015, pp. 33-49. [68]. A. Parretta, G. Calabrese, About the definition of "multiplier" of an integrating sphere, International Journal of Optics and Applications, Vol. 3, Issue 6, 2013, pp. 119-124. [69]. A. Parretta, H. Yakubu, F. Ferrazza, P. P. Altermatt, M.A. Green, J. Zhao, Optical loss of photovoltaic modules under diffuse light, Solar Energy Materials and Solar Cells, Vol. 75, 2003, pp. 497-505. [70]. P. Maddalena, A. Parretta, P. Tortora, P. Altermatt, J. Zhao, Simultaneous optical losses and current measurements in photovoltaic devices at variable angle of the incident light, Solar Energy Materials and Solar Cells, Vol. 75, 2003, pp. 397-404. [71]. A. Parretta, P. P. Altermatt, J. Zhao, Transmittance from photovoltaic materials under diffuse light, Solar Energy Materials and Solar Cells, Vol. 75, 2003, pp. 387-395. [72]. P. Maddalena, A. Parretta, A. Sarno, P. Tortora, Novel techniques for the optical characterization of photovoltaic materials and devices, Optics and Lasers in Engineering, Vol. 39, 2003, pp. 165-177. [73]. A. Parretta, H. Yakubu, F. Ferrazza, Method for measurement of the hemispherical/hemispherical reflectance of photovoltaic devices, Optics Communications, Vol. 194, 2001, pp. 17-32. [74]. A. Parretta, P. Grillo, P. Tortora, P. Maddalena, Method for measurement of the directional/hemispherical reflectance of photovoltaic devices, Optics Communications, Vol. 186, 2000, pp. 1-14. [75]. A. Parretta, A. Sarno, H. Yakubu, Non-destructive optical characterization of photovoltaic modules by an integrating sphere. Part I: Mono-Si modules, Optics Communications, Vol. 161, 1999, pp. 297-309. [76]. A. Parretta, C. Privato, G. Nenna, A. Antonini, M. Stefancich, Monitoring of concentrated radiation beam for photovoltaic and thermal solar energy conversion applications, Applied Optics, Vol. 45, 2006, pp. 7885-7897. [77]. A. Parretta, M. Tucci, M. Izzi, Optical properties of light diffusers as targets for concentrated solar beams characterization, in Proceedings of the 30th European Photovoltaic Solar Energy Conference (EU PVSEC’15), CCH Congress Centre, Hamburg, Germany, 14-18 September 2015, pp. 1481-1486. [78]. A. Parretta, G. Nenna, A. Antonini, Monitoring of concentrated solar radiation beams by scattering from lambertian surface and CCD recording, in Proceedings of the Conference on Optical Diagnostics and Monitoring: from Advanced Components to Novel Devices (OpDiMon’04), Capo Miseno, Bacoli (Napoli), Italy, 22-26 March 2004. [79]. A. Parretta, A. Antonini, M. Stefancich, Codice di calcolo "LS-CCD" per l'analisi dell'immagine prodotta intercettando un fascio di luce con un diffusore, registrato presso la SIAE, Registro Pubblico Speciale per i Programmi per Elaboratore, Num. 007144, Ord. D006364, 15 Aprile 2009 (in Italian). [80]. Paolo Colombani Design, Ferrara, Italy, www.paolocolombani.com [81]. G. Kortum, Reflectance Spectroscopy, Principles, Methods, Applications, Springer, Berlin, Germany, 1969, p. 28. Appendix 13.A Note on the Optical Concentration Ratio Let us consider again Eq. (13.17) of Section 13.2.1: 400 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators 3D  Copt n'2  sin 2  out . n 2  sin 2 in (13.A1) The meaning of in in Eq. (13.A1) must be correctly interpreted. In the case of an ideal SC, the angle below which all incident rays are transmitted and above which all incident 50 rays are back reflected is the acceptance angle acc . These SCs therefore have an optical response, i.e. a transmission efficiency curve, which has the shape of a descending step. The SCs that show the same response to the ideal are the 2D-CPCs [17]. Concentrators of the 3D-CPC type, on the other hand, show an almost ideal response, such as that of Fig. 13.1, if their form (canonical) is not altered by the effect of a truncation. The 50 , should not be confused with the angular acceptance angle of the concentrator,  acc 50 divergence S of the input light source. Let's imagine then to have a 3D-SC with acc acceptance angle. The actual and maximum optical concentration ratios will be given respectively by the following expressions: 3D  Copt n'2  sin 2  out , 50 n 2  sin 2  acc 3D Copt , max  n '2 . 50 n 2  sin 2  acc (13.A2) (13.A3) It is now clear that the optical concentration of such a concentrator will be determined by 50 , and not by the angular divergence of the incoming rays, in . its acceptance angle, acc 50 , its optical Although we illuminate the concentrator with a light beam with in < acc 3D 50 . If, however, we are concentration ratio Copt will always be determined by acc 3D considering the geometry of the incident beam and we want to know what values of Copt 3D and Copt , max we can realistically achieve with that beam by appropriately designing the SC, then the results will be: 3D  Copt n'2  sin 2  out , n 2  sin 2 in 3D Copt , max  n '2 . n 2  sin 2  in (13.A4) (13.A5) About out , the SC can be designed for different values of angular divergence, and in that 3D case Copt will be given by Eq. (13.A2) or (13.A4). The maximum out value is 90 °, and 3D then, at this value, the maximum value for Copt will be reached as indicated by Eq. 401 Advances in Optics: Reviews. Book Series, Vol. 3 2D (13.A3) or (13.A5) for a 3D-SC. Formulas for the optical concentration ratios Copt of a 2D concentrator can be obtained by Eqs. (13.A2)-(13.A5) after making the square root. Let us now consider a 3D-SC in the air, such as the 3D-CPC, and let us calculate the values of the optical concentration ratios for out  90 ° and out = 90: 3D Copt , max (out  90)  3D Copt , max ( out  n'2 sin2 out  n'2 sin2 out  46000, sin2 S (13.A6) n'2  90)  2  n'2 46000, sin S (13.A7) where  S is the angular divergence of solar radiation. For a SC with the receiver in air, like a 3D-CPC, at best we could focus the light 46,000 3D times. By immersing the receiver in a dielectric with n'  1.5, we could bring Copt to 3D values 100,000. Eqs. (13.A6) and (13.A7) establish that the limit values of Copt depend on the source-concentrator system geometry. By approaching the Sun, these limit values would decrease, while, moving away (for example on Mars), we would be able to reach higher optical concentration ratios. Appendix 13.B An Introduction to the 3D-CPC Concentrators Our interest on the 3D-CPCs (Three-Dimensional Compound Parabolic Concentrators) is that they allow to reach very high concentration levels, comparable to the theoretical ones, and that their optical transmission efficiency is quite constant within a defined angle of incidence of the collimated beam. A further advantage of these SC is that they operate with reflective surfaces, that do not induce the spectral dispersion of light. The 3D-CPC is a nonimaging concentrator developed by R. Winston to efficiently collect Cherenkov radiation in high energy experiments [46]. Since then, the nonimaging concentrators have been widely used to concentrate sunlight [16-20]. The CPC is a reflective concentrator with parabolic profile and is characterized by a quasi-step-like transmission efficiency (see Fig. 13.1) allowing the efficient collection of light from 0° to a maximum angle, called the acceptance angle acc . A 3D-CPC is characterized by the following parameters: a = radius of entrance aperture; a’ = radius of exit aperture; L = length; acc = acceptance (or tilt) angle; f = focal length of the parabolic profile. An ideal (canonical) 3D-CPC is completely determined by two of the above five parameters, related by the following basic relationships: 402 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators f  a ' (1  sin  acc ), (13.B1) a'  a  sinacc, (13.B2) L  (a  a')  ctgacc, (13.B3) z p'1 acc p1 a' acc acc F1 O x F2 f Fig. 13.B1. Basic scheme of construction of an ideal 3D-CPC. The project of a 3D-CPC is very simple. We can start fixing, for example, the dimension of exit aperture, with radius a’, and the focal length, f; the acceptance angle acc is derived directly from Eq. (13.B1), but can be also obtained by the following geometric construction. We start drawing the longitudinal cross section of the 3D-CPC on the x/z plane (see Fig. 13.B1). On a Cartesian plane x/z the projected exit aperture of the CPC, with diameter F1F2, is aligned along the x-axis and centered on the origin O. Now we draw the parabola p1 with upward concavity, focal length f and focus on F1(-a’, 0) (see Fig. 13.B1). The parabola p1 is then rotated counter clockwise (CCW) around the axis perpendicular to the x/z plane and passing through F1, until it touches the point F2. The corresponding angle of rotation is the acceptance angle acc . The positive segment of the rotated parabola p'1 is the right profile of the CPC, intersected by the x/z plane. The left profile of the CPC is obtained starting from a second parabola with focus on F2 and rotating it clockwise (CW), and is the specular image of p'1 respect to the z axis. The analytical expression for the z’=z’(x’) coordinate of a profile of the CPC becomes: z '  d  w( x' )  e  w( x' )  g , (13.B4) where d, e and g are constant quantities: 403 Advances in Optics: Reviews. Book Series, Vol. 3 d cos acc , 16 f  sin 2  acc (13.B5) e a' cos acc 2b  cos acc 1   , 2 2 16 f  sin  acc 4 f  sin  acc (13.B6) g b 2  cos acc a'2  cos acc a' b  cos acc b  ...    2 16 f  sin 2  acc 4f 4 f  sin  acc (13.B7) ...  a' sin  acc  f  cos acc , where also b and c are constant quantities: b  4 f  cosacc  2a' sinacc, c  ( a ' ) 2  sin  acc  4 f 2  sin  acc  4 f a '(1  cos  acc ), (13.B8) (13.B9) and the function w (x’) is given by: w( x ' )  b 2  4  ( 4 f  x ' c )  sin  acc . (13.B10) Eqs. (13.B4)-(13.B10) allow to calculate the slope of the CPC curve at any point. In particular, we look for the point where the tangent is parallel to the z’ axis. This point defines the upper limit of the CPC profile, that is the maximum length L of the CPC, then it also defines the maximum value of the input opening radius, a (see Fig. 13.B2). By deriving Eq. (13.B4) we have: d z' d z' d w    ... d x ' d w d x'   e ...    d   (16 f  sin  acc ).  2 b 2  4  (4 f x'c)  sin  acc    (13.B11) The condition for a tangent to the curve parallel to z’ axis is: d z'  d x'  b 2  4  (4 f x'c)  sin  acc  0 f  a' sin  acc a'  x'  x'(L)    a. sin  acc sin  acc (13.B12) The profile of the CPC ends at the z’ = L, corresponding to the point A of Fig. 13.B2, where the tangent to the profile is parallel to the z’ axis. The surface of the 3D-CPC is finally constructed by turning the left and right profiles of the angle  around the z’ axis. Because of this construction, two extreme rays (1 and 2 in Fig. 13.B2) incident at acc and crossing the z’ axis (meridian rays), will be both collected at F1. 404 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators acc acc z' a A 2 1 L acc acc x' F1 a' F2 Fig. 13.B2. Longitudinal cross section profile of the 3D-CPC. For the optical simulations (see Section 13.3), we have used an ideal (not truncated) 3D-CPCs with the following parameters: acc = 5°, L = 712.9 mm, a = 57.4 mm, a’ = 5.0 mm, f = 5.4 mm. All the optical simulations were carried out by using the TracePro ray-tracing software for opto-mechanical modeling of Lambda Research [58]. A brief note should be made here to clarify what happens when a 3D-CPC is halved. This discussion is not easy, because in a 3D-CPC the behavior is not the same for all rays; for example, meridians rays, i.e. those whose incidence plane contains the optical axis of the CPC, behave differently from the others, and in the same way as all rays behave in an ideal 2D-CPC concentrator. It is not easy to predict the overall behavior of rays in a 3D-CPC without a raytracing simulation, but we know that the response of optical efficiency in a 3D-CPC is much like that of a 2D-CPC, apart from the fact that the steplike optical efficiency curve is slightly rounded. So, our discussion with 3D-CPCs will take account of only meridians rays, knowing that this is an approximation that approaches enough the real behavior of the 3D-CPC. 50 Let's consider then an ideal 3D-CPC, which we call CPC1 with acceptance angle  acc 1, 50 having an ideal internal wall ( Rw = 1.0), designed to have a out = 90° at  in =  acc 1. Fig. 13.B3 shows the longitudinal section of the CPC1. An extreme ray (1), tilted of 50  in =  acc 1 , is transmitted and exits at 90° (edge ray principle), touching the point F1, 50 which is the focus of parabola p1. Also, the ray (2), inclined to  in =  acc 1 , is transmitted after touching point F1. This is because the axis of parabola p1 is parallel to the direction of rays (1) and (2), and the parabolic profile of the CPC1 has been designed to be inclined 50 at an angle equal to acc 1 . It is clear from Fig. 13.B3 that the rays entering the CPC1 at 50 50  in < acc 1 are all transmitted, by the definition of acc . 405 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 13.B3. Longitudinal cross section profile of the CPC1. Let’s try to slightly increase  in of the rays (1) and (2); it is easy to demonstrate, observing Fig. 13.B3, that in both cases, the rays (1') and (2') are reflected backwards. It 50 is also easy to convince that all rays with  in > acc 1 are back reflected. The CPC 50 transmission efficiency curve is then step-like shaped. For  in < acc 1 , we have therefore: Φout = Φin ; Eout  Aout  Ein  Ain ; A E out  C opt  in  C geo  Copt  Cgeo. Aout E in (13.B13) We now halve the CPC1 by removing the portion containing the input opening, building the CPC2 (see Fig. 13.B4). It is clear from Fig. 13.B4 that all rays incident on CPC2 with 50 50 50  in < acc 1 are transmitted, so we can already state that  acc 2 is at least equal to acc1 . 50 From Fig. 13.B4 it is also clear that rays entering the CPC2 with  in > acc 1 and hitting 50 the internal wall, such as radius (3), are rejected. However, rays with  in > acc 1 , entering the CPC2 through the annulus with OC inner radius and OB outer radius, without being impacted on the wall, are transmitted. Thus, the effect of halving the CPC1 is to extend a 50 , only little the optical efficiency curve, but it does not involve a significant increase in  acc 50 the appearance of a tail for  in > acc 1 , as it can be seen clearly from Fig. 13.14b. 406 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators Although the CPC2 acceptance angle is almost equal to that of CPC1, it is a fact that its optical efficiency curve is extended to higher angles, and this must have a consequence on optical concentration ratio. It should be borne in mind, in fact, that the definition of angle of acceptance at 50 % of maximum efficiency is just a convention and works well when the curve is symmetrical around this angle, which is not the case for the halved CPC2. Fig. 13.B4. Longitudinal cross section profile of the CPC1 and its half CPC2. Fig. 13.B5. Longitudinal cross section profile of the CPC3 and its half CPC4. Let's see now what happens to the other CPC2 parameters after the halving operation. We have that Cgeo is decreased due to the decrease of Ain , but not too much, because the profile of parabola p1 is almost parallel to the optical axis in the truncated portion of the CPC1. Decreasing Cgeo , Copt must decrease, because we know that C geo is the insurmountable limit for Copt , being the optical efficiency always  1 (see Eq. (13.18a)). The decrease in Copt can also be evoked by Eq. (13.17a), as consequence of the increase of the  in denominator, while out remains unchanged and equal to 90°. It is interesting to continue this analysis by repeating it for a 3D-CPC concentrator that has been designed to have a maximum divergence at output out < 90°, CPC3 (see Fig. 13.B5). It’s clear that, for this 3D-CPC, the halving operation, which brings to CPC4, 407 Advances in Optics: Reviews. Book Series, Vol. 3 50 has different consequences. For example, rays incident at in >  acc 1 will all be transmitted until extreme rays such that of type (3) in Fig. 13.B5 will not reach 50 out = 90°. This condition defines the acceptance angle  acc 2 , which will be greater than 50  acc 1 and will result in an optical efficiency curve like that of CPC1, that is, step-like, but wider. Appendix 13.C How to Calculate the Inverse Radiance When the inverse method is simulated, the planar screen is configured as an ideal absorber and the profile of the measured incident irradiance E ( ,  ) (see Fig. 13.C1a) is converted into the profile of the radiance distribution function of the concentrator, Linv ( , ) , by the (cos)-4 factor. Indeed, if P ( ,  ) is a point on the screen, E ( ,  ) the corresponding incident irradiance and dS an elementary area around P ( ,  ) , the flux through area dS is d  E ( ,  )  dS and it is confined within the solid angle d  given by: d  dS  cos dS  cos dS   2  cos3  . 2 2 r ( ) d (d / cos ) (13.C1) The inverse radiance produced by the concentrator towards ( ,  ) direction will be therefore expressed by: Linv ( ,  )  d  ... d  Ain  cos  E ( ,  )  dS d 2 E ( ,  ) ...   , ( dS  cos 3  / d 2 )  Ain  cos  Ain cos 4  (13.C2) where Ain is the input aperture area of concentrator. The radiance can be normalized to the value at  = 0° giving: Lrel inv ( ,  )  E rel ( ,  ) Linv ( ,  ) E ( ,  )   . Linv (0) E (0)  cos 4  cos 4  (13.C3) The inverse radiance is related to the optical efficiency of the concentrator in the following way: 408 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators Lrel inv ( ,  )   ( , ) Linv ( , ) rel ( ,  )  dir   dir . Linv (0)  dir (0) (a) (13.C4) (b) (c) Fig. 13.C1. (a) Schematic of the irradiation of the planar screen (ps) by the inverse light produced by the solar concentrator (sc); (b) Process of recording by the CCD of the image produced by the irradiance map produced on the planar screen. P( , ) is a point on the screen and E( ,) is the corresponding incident irradiance; (c) The CCD is schematized as a lens and a plane representing the image sensor of the CCD; Pccd ( ,  ) is a point and I ccd ( ,  ) is the intensity (irradiance) on the CCD sensor. We conclude that, when we simulate the inverse method, the normalized profile of the direct transmission efficiency of the concentrator is directly derived by the normalized irradiance incident on the ideal absorbing screen, by the expression: rel  dir ( ,  )  E rel ( ,  )  cos 4  . (13.C5) When the “inverse” method is applied experimentally, the screen is used to send back the diffuse, inverse light towards the CCD and must have a Lambertian character (reflectivity independent on the incidence angle, and constant radiance of the reflected light, as function of observation angle) in order to allow the reconstruction of the irradiance map on the screen from the intensity map produced on the CCD. 409 Advances in Optics: Reviews. Book Series, Vol. 3 If the CCD is aligned with the optical z axis and close to the concentrator (see Fig. 13.C1b), the intensity profile of CCD image must be corrected by a further (cos )-4 factor, as we demonstrate in the following (see also Fig. 13.C1c). The total flux reflected by the unitary area of (ps) centered in P ( ,  ) is: E R ( , )    LR ( , ), (13.C6) where R is the reflectance of (ps), ER ( , )  R  E ( , ) is the reflected irradiance, and L R ( ,  ) is the radiance of the screen. The flux reflected by the unitary area of (ps) and flowing inside the solid angle by which the unitary area is seen by point O ( ,  ) is:  R ( , )  LR ( , )  cos4  . d2 (13.C7) This flux is the same reaching the CCD sensor area (c / d)2 centered on point Pccd ( ,  ) . The intensity of the CCD image at point Pccd ( ,  ) , proportional to the irradiance incident at that point, is therefore: I ccd ( , )  k   R ( , )  ... (c / d ) 2 cos 4  cos 4  ( , )    k  E   ... R   c2 c2 cos 4  . k  R  E ( , )    c2 k  LR ( , )  (13.C8) By using Eq. (13.C2), we obtain: I ccd ( , )  k  R  Ain  Linv ( , )  cos8  . 2 2  c d (13.C9) From Eq. (13.C9) we finally obtain the inverse radiance of the concentrator from the intensity on the CCD: Linv ( ,  )    c2  d 2 k  R  Ain  I ccd ( ,  )  cos 8  . (13.C10) The normalized radiance becomes: rel 8 Lrel inv ( ,  )  I ccd ( ,  )  cos  . (13.C11) Finally, from Eq. (13.C4) we obtain the normalized transmission efficiency of the concentrator: 410 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators rel rel inv ( , )  I ccd ( , )  cos8  . (13.C12) Appendix 13.D Realization of Lab-Made Integrating Spheres We have seen how it is important the use of integrating spheres in experimentation on the optical characterization of solar concentrators. We have made widespread use of integrating spheres both as Lambertian light sources, once they have been coupled to a lamp, either as light detectors or radiometers, once they have been coupled to photodetectors (or solar cells) inside it, in turn coupled to a voltmeter or to a lock-in amplifier. The problem of integrating spheres is that they must be built "ad hoc" for the specific application to which they are addressed. They are also made of aluminum and must be equipped with a series of windows for the entrance and exit of light, as well as for the measurement of the internal irradiation or light spectrum, depending on how sophisticated is desired the system [47, 68-75]. The inside of the spheres must be made with a highly reflective and highly diffusing white coating, so as to behave like a Lambertian diffuser [81]. I remember that a Lambertian diffuser is a diffuser that, when irradiated by a collimated light beam, emits reflected light of constant radiance in all directions, which depends on the cosine of the angle of incidence of light. Generally, the cost of these spheres is very high, in the order of thousands of euros, depending on the dimensions. Our main goal, before starting the work on SCs’ characterization, was therefore to organize ourselves to build in-the-lab low-cost integrating spheres. We took into consideration the use of plastic. Garden lamps with plastic transparent globes, of various sizes, are available in supermarkets at very low prices. They are made of polyethylene, robust and already equipped with a circular opening for the light bulb. Of course, the use of aluminum spheres has a very precise reason, i.e. to ensure mechanical stability to them, to be easily machined without any risk of fractures of the material, which can happen with plastic, but above all to ensure a good thermal stability to the sphere once illuminated by sources that can reach hundreds of watts of power. Of course, we were aware of this, but we considered that the power of the light sources we used was low enough to not produce any problems of thermal stability, deformation of the sphere, etc. To build a good integrating sphere [47, 68-75] it is necessary that the total surface of the openings present on it does not exceed 4-5 % of the total sphere area, limit most likely to be exceeded in small spheres. If the original sphere has already an opening exceeding these limits, then it is necessary to reduce it with a suitable flange. Before providing this, however, it is useful to work with a big opening at first, for example that of the original garden lamp (see Fig. 13.D1a), since it is easier to manually access it when going through all the surface treatments that we will discuss later. It should also be taken into account that these spheres do not always work alone, but sometimes they are paired with each other to better disperse the light inside them, so it is necessary to adapt the windows to one another. Finally, the spheres are not always empty; they can contain a lamp, when this is not simply faced to a window; for example, it can be used the original lamp holder to put inside the sphere an appropriate lamp. Inside the sphere, we can also insert some 411 Advances in Optics: Reviews. Book Series, Vol. 3 baffles, which serve to disperse the light, especially those particularly directional coming from a source. As for the size of the spheres, it must be big enough compared to the overall dimensions of the apertures, but not too much because the increase of dimensions reduces their sensitivity. So, a compromise must be found between these two requirements. Fig. 13.D1 shows some plastic globes used in the laboratory. (a) (b) Fig. 13.D1. Plastic globes made from garden lamps (a). Two spheres were optically coupled (b). In order to achieve a thick and well-adhering coating inside the sphere, it is necessary to properly prepare its surface. It must be abraded evenly using abrasive paper, or rather a sandblasting machine. Fig. 13.D2 shows an abraded sphere, during a temperature control in the interior. To make the openings on the sphere, we can engrave the surface using a common drawing compass after replacing the pencil lead with the blade of a cutter (see Fig. 13.D3a). After having sanded the inner and outer surfaces of the sphere, and before coating it with the diffusing coating, it is necessary to make the surface opaque, being originally transparent or semi-transparent. For this purpose, the outer and inner surfaces have been coated with a sprayed white paint (see Fig. 13.D3b), then the outer surface was coated with a chromium-plating spray paint (see Fig. 13.D3c). This ensures a perfect optical insulation between the inside and outside of the ball. Finally, to prevent spurious reflections from the outer surface of the sphere during the measurements, the outer surface was coated with an opaque black spray paint (see Fig. 13.D3d). The Lambertian inner coating can be prepared in different ways. We have chosen to use Barium Sulphate (BaSO4), which, besides being simple to apply, also shows the best optical properties. However, the preparation of the optimal suspension of Barium Sulphate requires a long work of study in which it is necessary to choose the correct components and their proportion, besides how to apply the solution on the inner surface of the sphere. Fig. 13.D4 shows some photos taken during the final coating phase. In Fig. 13.D5a it is shown a sphere coupled with the Rondine-Gen1 concentrator, and in Fig. 13.D5b it is shown a baffle for the optical decoupling of input and output windows. In order to decide how to prepare the BaSO4 suspension, many specimens were prepared by varying all possible parameters, including the substrate material (see Fig. 13.D6). Subsequently, spectral reflectance measurements were performed at the ENEA-Portici laboratories. 412 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators Fig. 13.D7 shows the best results. It can be seen how the reflectance is over 90 % in the silicon range, whose technology is the base of all the devices we use, both photodetectors and solar cells. (a) (b) Fig. 13.D2. Example of plastic globe with abraded surface, with a lamp in the interior (a). In (b) a thermocouple, connected to a multimeter, has been inserted inside to reveal the temperature reached on the sphere surface. (a) (b) (c) (d) Fig. 13.D3. E. Bonfiglioli realizes the opening in a sphere (a). Preparation of the surfaces with different types of paints (b)-(d) before the application of the final BaSO4 coating. 413 Advances in Optics: Reviews. Book Series, Vol. 3 (a) (b) Fig. 13.D4. E. Bonfiglioli applies the BaSO4 suspension inside a sphere (a) and dries it with the hair-dryer (b). (a) (b) Fig. 13.D5. Photo of a sphere coupled to the Rondine-Gen1 concentrator (a). A baffle is visible inside the sphere (b). Fig. 13.D6. E. Bonfiglioli during the preparation of the Barium Sulphate specimens. 414 Chapter 13. Optical Methods for the Characterization of PV Solar Concentrators Fig. 13.D7. Results of the spectral analysis carried out on some samples in the spectral range 200-2500 nm (Silicon work range is between 300 and 1200 nm). 415 Chapter 14. Determination of the Heights of the Smoke-Plume Layering in the Vicinity of Wildfires and Prescribed Burns Chapter 14 Determination of the Heights of the SmokePlume Layering in the Vicinity of Wildfires and Prescribed Burns V. Kovalev, C. Wold, A. Petkov, S. Urbanski and W. M. Hao1 14.1. Introduction Recent decades have witnessed an increase in the frequency, duration, and severity of wildfires around the world. Fires are a major source of fine particulates (PM2.5), ozone (O3), and other pollutants that are detrimental to human health and degrade visibility. Heightened concerns about the impact of poor air quality on public health and the more strenuous regulatory established by the Environmental Protection Agency elevate the need for air regulatory and land management agencies to address the contribution of fires to air pollution. To comply with regulatory rules, land management agencies and air regulators need modeling tools to accurately predict the contribution of fire emissions to visibility impairment and PM2.5 and O3 pollution. Unfortunately, the ability of existing models to simulate smoke production and dispersion has not been thoroughly tested. The uncertainties and biases of these models and the limits of their applications are mostly unknown or poorly characterized, which is due mostly to the lack of adequate data for evaluation. The few smoke dispersion data sets available for model validation were from prescribed fires [1-3], which often differ significantly from wildfires in fuel conditions and meteorology, etc. To validate plume rise and high-resolution smoke dispersion models for a wide range of meteorological, fire behavior, fuel, and terrain conditions, smoke plume rise, dynamics, and transport in the near and far vicinities of wildfires and prescribed burns were investigated. This allowed for some evaluation of plume rise and high-resolution smoke dispersion models. Quantifying the plume heights under different meteorological conditions can potentially be achieved by remote sensing. Lidar profiling of the atmosphere is one of the most V. Kovalev Missoula Fire Sciences Laboratory, Missoula, Montana, USA 417 Advances in Optics: Reviews. Book Series, Vol. 3 suitable methods for this purpose. It allows extracting vertical profiles of the smoke plumes and probably is the best instrument for the investigation of smoke plume heights and downdrafts that transport the smoke particulates downward, worsening air quality at ground level. During the last decade, a number studies were published, in which the characterization of smoke plume behavior, including estimates of the plume injection height, were made using the information derived from satellite data, in particular, from the CALIPSO data [4-8]. However, the information derived from space data is limited to smoke plumes having discernable features. Satellite measurements are most effective when wildfires produce plumes that penetrate through the boundary layer and are transported downwind over great distances. However, they are much less effective for the investigation of smoke that remains within the boundary layer. Ground-based remote sensing instrumentation, in particular scanning lidar, is the most appropriate tool for monitoring of wildfire smoke-plume dynamics and heights within the boundary layer. The ground-based lidar is the only instrument capable of obtaining highly detailed, three-dimensional range and height-resolved information for smoke distributions and optical properties over ranges as long as 10+ km. It can operate from a position far outside the burning area with complete safety for the personnel involved. Lidar allows continuous monitoring of smoke-polluted atmospheres adjacent to both wildfires and prescribed burns, and for investigating temporal and spatial variation of aerosol properties, plume heights and dynamics, and direction and rate of smoke plume movement in near real-time. The lidar temporal data series can reveal strong downdrafts that transport smoke particulates downward, worsening air quality at ground level. An example of such a field experiment designed to obtain the data necessary to improve the air quality models used by agricultural smoke managers in the northwestern United States was performed in August of 2013. In that experiment, the ground-based mobile lidar, developed at the US Forest Service Missoula Fire Science Laboratory, was used to monitor plume rise heights for nine agricultural fires in the northwestern United States. The lidar measurements were compared with plume rise heights calculated with the Briggs equations, which are used in several smoke management tools. The preliminary evaluation results and recommendations regarding the application of the models to agricultural burning based on lidar measurements made in the vicinity of Walla Walla, Washington, on August 24, 2013 are published in [9]. Till recently, the basic issue with smoke measurements was the absence of a practical scanning lidar methodology for determining parameters of interest, that is, the methodology adapted to the specifics of smoke polluted atmosphere. In general, accumulated experience from investigations of the boundary layer is helpful. However, the structure and the temporal and spatial changes of smoke plumes in the vicinity of wildfires and prescribed burns differ dramatically from the features typically present in the entrainment zone of non-disturbed atmosphere. Temporal changes of the backscatter coefficient values in smoke plumes may be much larger or much less than those in the boundary layer. Spatial gradients in the backscattering at the smoke plume edge, which 418 Chapter 14. Determination of the Heights of the Smoke-Plume Layering in the Vicinity of Wildfires and Prescribed Burns must be determined to locate the smoke boundary, may have an extremely wide range of values. When determining the boundaries of the atmospheric aerosol formations, generally some relative rather than absolute characteristics of the lidar signal are used. For example, when using the most common gradient method, one can determine the aerosol boundary location as the range where the examined parameter (e.g., the derivative of the square-rangecorrected lidar signal) is a maximum; or where it decreases from the maximum value down to a fixed, user-defined level [10]. However, there is no unique way to establish standard values for this level which would be acceptable, at least for the most likely atmospheric situations. This issue is common for all such techniques; even the use of the modern wavelet technique requires an a priori selection of concrete parameters [11]. Recently in the studies [12-13], a new data processing method for determining the upper heights of the regions of intense backscatter was considered, which has significant advantages over the conventional technique of analyzing the derivative of the square range corrected signal. First, the method does not require separation of the square-range corrected backscatter signal from the recorded total signal, and accordingly, it does not require the estimation of the constant offset in the total signal. The second advantage of that method is that specific functions used in the data processing technique are generally significantly less noisy than the derivative of the range corrected signal, especially at distant ranges. Finally, the third advantage of this method is the possibility to use all information, which is present in the set of the multiangle signals of a scanning lidar. The differentiation technique, commonly used for the investigation of the atmospheric boundary layer, is generally applied to signals of one-directional, usually, zenith directed lidar; its application for multiangle profiling of the atmosphere, which would allow the investigation of the atmospheric layers close to ground level, is problematic. Below an advanced data processing methodology, based on principles proposed in [12] and [13] is considered. Its essence and specifics are cleared up by the analysis of real signals of a scanning lidar, obtained in August 2016 in the vicinity of wildfires near Missoula, Montana. 14.2. Determination of the Heights of the Smoke-Plume Layering in the Vicinity of Wildfires: Theory and Data Processing Methodology Let us consider the basic algorithms used for processing the lidar data. The lidar signal, P(r,), measured under elevation angle  and recorded at the range r, is the sum of the variable backscatter signal, P(r,) and the range-independent offset, B, P (r , )  P(r , )  B , (14.1) where the backscatter signal is defined as, 419 Advances in Optics: Reviews. Book Series, Vol. 3 P ( r , )    1 C   , m ( r )    , p ( r ) [Tm ( 0 , r )] 2 [T p ( 0 , r )] 2 , r2 (14.2) here C is the lidar constant, which includes also the transmitted light pulse energy; ,m(r) and , p(r) are the molecular and particulate backscatter coefficients; [Tm(0, r)]2 and [Tp(0, r)]2 are the molecular and particulate two-way transmission from the lidar to the slope range r, respectively. The total signal P(r,) is transformed in the auxiliary function Y(x,), defined as [13], Y ( x ,  )  [ P ( x ,  )  B ] x ( ), (14.3) where x() = r2 is the new independent variable. Using the method of finite differences, the sliding numerical derivative of this function, Y/x, is calculated, and the intercept point of the extrapolated slope fit with the vertical axis is found. The intercept function versus x() can be found as the difference of two terms, Y0* ( x)  Y ( x)  Y x, x (14.4) here for simplicity the function x() is written as single x. The retrieval technique is based on determining the range-weighted intercept function defined as, Y0 ( x)  Y0* ( x) Y ( x) dY   . x x dx (14.5) The absolute values of the function Y0(x), are transformed into the functions of height, Y0(h, ), where Y0(h, ) ≥ 0. Then the envelope of these positive functions, env[Y0(h)] for the all N elevation angles used during scanning is determined, envY0 (h)  maxY0 (h,1 )  Y0 (h, 2 )  ...  Y0 (h, N ) . (14.6) To reduce the influence of the high frequency spikes, the envelope is smoothed. The maximum value of the smoothed envelope, Y0,max  maxenvY0 (h) , (14.7) is found, and the function env[Y0(h)] is normalized to 1, that is, envY0 (h)norm  envY0 (h) . Y0. max (14.8) The data points of interest, Y0,(h) within the total height interval from hmin to hmax are found as, 420 Chapter 14. Determination of the Heights of the Smoke-Plume Layering in the Vicinity of Wildfires and Prescribed Burns Y0,  (h)  envY0 (h)norm . (14.9) The function, Y0,(h) is calculated using a selected set of fixed independent variables χ, which values may be chosen within the range 0 ≤ χ < 1. This function is then used for the determining the heights of interest, hmax(χ) for each selected χ. These heights are determined as the maximum heights, where Y0,(h) still differs from zero. An example of the functions Y0(h, ), obtained from real lidar signals, measured in a smoke polluted atmosphere, and their non-smoothed envelope, env[Y0(h)], are shown in Fig. 14.1. The lidar was scanning the atmosphere vertically within the angular sector from 10o to 80o. The functions Y0(h, ), measured under different elevation angles, , are shown as the colored curves; their non-smoothed envelope is shown as the thick black solid curve. Fig. 14.1. Set of functions Y0(h, ), measured by a scanning lidar in smoke polluted atmosphere (colored curves), and the initial non-smoothed envelope function, env[Y0(h)] (thick black curve). The dependence of the normalized function, env[Y0(h)]norm, on height is shown in Fig. 14.2. As can be seen in the figure, different  yield different hmax(). In our case, the selection  from 0.1 to 0.4 yields the values hmax() which are close to the actual maximum of the smoke plume, hmax, located close to height 1000 m. Meanwhile, the selection of the larger  = 0.5 yields hmax() ≈ 470 m, which significantly differ from the actual height hmax. Thus, the proper selection of , which provides hmax() close to the actual smokeplume height, hmax, may be a significant issue. The only sensible way to solve this issue is the examination of the behavior of hmax() in some range of the independent . The same as in [13], here the consecutive values of  with the fixed step  = 0.05 are used; that is, min = 0, 1 = 0.05, 2 = 0.1, etc. For each discrete , the corresponding height, hmax() is found. As shown in Fig. 14.2, the latter is determined as the maximum height where Y0,χ(h) still exceeds zero. 421 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 14.2. The solid curve is the smoothed and normalized envelope function env[Y0(h)]norm. The arrows illustrate the dependence of hmax() on selected . The dependence of the heights, hmax() on  for the case under consideration is shown in Fig. 14.3, where the discrete heights, hmax() versus χ are shown as black filled circles. Following the methodology in the study [13], the selection of the upper heights of smoke polluted layers is based on finding areas, where the differences between the adjacent heights hmax(k) and hmax(k+1) at k and k+1 are minimum. The second requirement in [13], used here, is that the differences between the previous adjacent heights hmax(k-1) and hmax(k) are maximum. Fig. 14.3. The black filled circles show the dependence of the heights hmax() on discrete  and the red horizontal dashes show the differences between the adjacent heights hmax(i) and hmax(j). Red arrows show the points where the differences between the adjacent heights are maximum. 422 Chapter 14. Determination of the Heights of the Smoke-Plume Layering in the Vicinity of Wildfires and Prescribed Burns In the figure, the differences, Δhi,j = hmax(i) - hmax(j), are shown as horizontal dashes. The corresponding upper heights, hup, of the layers of increased backscattering, determined in accordance with the above recommendations, are located at hup ≈ 1100 m and hup ≈ 480 m. 14.3. Some Results of Lidar Profiling of the Smoke-Polluted Atmosphere in the Vicinity of Spotted Wildfires The elements of the mobile Missoula Fire Sciences Laboratory (FSL) scanning lidar used in below investigation are given in [14]. As the light source, a short-pulsed Nd:YAG laser, attached to the top of a receiving telescope is used. The lidar receiver measures the backscatter signal at two wavelengths, 355 and 1064 nm, simultaneously. The laser beam is emitted parallel to the telescope after going through a periscope, so that the effective exit aperture is offset 0.41 m from the center of the telescope. The periscope increases the distance at which the laser beam overlaps the telescope field of view up to ~ 1000 m, simultaneously decreasing the dynamic range of the signals and increasing the total measurement range. The telescope-laser system is able to turn through 180 horizontally and 90 vertically. The schematic of determining the smoke plume heights in the vicinity of wildfire is shown in Fig. 14.4. The lidar, located generally at a distant range from a wildfire (or scattered wildfires), scans the plume under a number of elevation angles. The different colors in the figure show different intensity of the backscattering in relative units. The slant lines show the discrete directions of lidar scanning and the black filled dots mark the maximum heights of the smoke plume fixed under different elevation angles. Fig. 14.4. Schematic of determining the height of the smoke-plume layering with scanning lidar in a distant vicinity of wildfire. The schematic of the scanning lidar setup during lidar profiling of the smoky atmosphere polluted by three closely located wildfires is shown in Fig. 14.5. The Roaring Lion Fire [15], Observation Fire [16], and Cedar Fire [17] were located 28 miles, 33 miles, and 423 Advances in Optics: Reviews. Book Series, Vol. 3 35 miles from the lidar location, respectively. The lidar was oriented to scan east-west sectors to the south of its location; it operated making vertical scans in a wide azimuthal sector with the fixed azimuthal step 10o. In the morning of August 4, 2016 the Roaring Lion and Observation fires produced a well-defined column of smoke that flowed north along the east side of the Bitterroot Mountains, located on the west side of the Bitterroot Valley in Montana. In the afternoon, this column became diffused and spread eastward across the valley. The Cedar fire in Idaho contributed higher elevation smoke as smoke reaching the valley had to traverse the Bitterroot Range. Fig. 14.5. The relative locations of the three wildfires producing the majority of the smoke in the Bitterroot Valley during lidar scans made on 4 August 2016. Let us consider the spatial and temporal distribution of the smoke-plume upper boundaries, hup, and their dynamics during the day by analyzing simultaneously the temporal and spatial behavior of the discrete data points h(). Note that instead applying the whole set of , as was shown in Fig. 14.3, the data points shown in Figs. 14.6 and 14.7 are analyzed within the restricted range of , within the range, 0.05 ≤  ≤ 0.5. Our analysis revealed that the exclusion of the data points with  > 0.5 allowed better results in determining the height of interest, hup. 424 Chapter 14. Determination of the Heights of the Smoke-Plume Layering in the Vicinity of Wildfires and Prescribed Burns Fig. 14.6. Data points hmax() as a function of local time during the period from 11:40 to 14:00. The dashed arrows show the change of the upper heights of the areas of increased backscattering during the measurement interruptions. Fig. 14.7. The same as in Fig. 14.6 but for the period from 14:06 to 16:40. In Fig. 14.6, the set of the data points h() is shown obtained during the time period from 11:40 to 14:00 local time at the wavelength 1064 nm. The lidar operated in the multiangle mode, scanning the atmosphere vertically within the angular sector from 10o to 80o with the angular separation ~3o. The clusters of data points within the altitude range 400-1100 m, show the real layering dynamics. Initially at 11:40, two layers with increased backscattering were recorded, the upper at the height ~1000 m, and the bottom at and below 600 m. However, then the upper smoke-plume layering moved down, and after ~12:30 this dispersed layer can be seen at heights hup ~ 750 m and below, that is, two previously separated layers of increased backscattering merge. After ~13:00, an additional 425 Advances in Optics: Reviews. Book Series, Vol. 3 clusters with dispersed boundaries appears within the height interval 1300-1600 m, which have tendency of dispersing up and down. In Fig. 14.7, the next set of the data points h() is shown. These results were obtained during the period from 14:06 to 16:40. Note that after ~16:10, in addition to the areas of increased backscatter within the lower heights, below 1500 m, a small cluster of data points appears close to the height ~2500 m. Presumably it originated from the appearance of another smoke layering at this increased height. However, the absence of lidar data after 16:45 did not allow making trustworthy conclusions about the origin of these data points. 14.4. Determination of the Heights of the Smoke-Plume Layering in the Vicinity of Prescribed Burns The general schematic for determining the smoke plume heights in the vicinity of prescribed burns is shown in Fig. 14.8. The lidar, located close to the vicinity of a local scale fire, scans the plume under a number of elevation angles. The slant lines in the figure show the directions of lidar scanning and the filled dots mark the maximum height of the smoke plume fixed under different elevation angles. Fig. 14.8. Schematic of determining the height of the smoke plume column with scanning lidar. Let us consider experimental data obtained on August 25, 2013 in the area of the prescribed fires near Walla Walla, WA, USA [18]. These data were obtained with the same FSL scanning lidar which was used in the measurements discussed in Section 14.3. The vertical scans were performed in a fixed azimuthal direction. The vertical scans were made within the angular sector from 5o to 60o with the angle resolution 1o. Each recorded signal was the average of 30 shots, and each total scan was recorded for ~75 sec. Same as above, only the backscatter signals at 1064 nm were applied for the analysis. The typical optical situations met in August 25, 2014 are illustrated in Figs. 14.9 (a)-(d). Initially, when the prescribed fire is starting, the vertically stratified smoke-plume column without well-defined horizontal layering in its vicinity is observed (Fig. 14.9(a)). Then the smoke plume aerosols start accumulating in the vicinity of the maximum height of the 426 Chapter 14. Determination of the Heights of the Smoke-Plume Layering in the Vicinity of Wildfires and Prescribed Burns vertically stratified plume (Fig. 14.9(b)). The strong dispersing processes create the specific combination, in which the vertically stratified smoke plume and horizontal smoke-plume layering take place (Fig. 14.9(c)). Finally, when the fuel for the fire becomes exhausted, the vertical smoke-plume structure degrades. At that stage, the smoke plume particles create a horizontal (or close to horizontal) layer elevated over ground level (Fig. 14.9(d)). Fig. 14.9. (a) The white contour shows the two-dimensional image of the smoke plume boundaries determined on August 25, 2013, at 10:57 of local time; (b) The same as in (a) but at 11:03; (c) The same as in (a) but at 11:08; (d) The same as in (a) but at 11:23. Following the definition of plume injection height given in [4], we will define it as the maximum height of the vertical zone in which a buoyant plume begins to transport horizontally away from its origin source. In the case under consideration, the plume in in Fig. 14.9(a) is fixed at the initial moment of reaching the plume injection height, whereas Figs. 14.9(b)-(d) show the smoke plumes with well-developed horizontal layering. There is some inconsistency between the term “injection height” and the possibility of its accurate determination. In many cases, the lidar can reliably determine only top of such a zone. The bottom part of this zone is often extremely dispersed, moreover, it may have multilayering structure, or extend down to ground level, so that the definition of the lowest 427 Advances in Optics: Reviews. Book Series, Vol. 3 boundary of the injection height may often be an issue. However, just the top of this height determines the maximum distance over which the smoke plume may travel and impact its destination, for example, such as the Arctic ice. Keeping this observation in mind, we will focus here on the determination the maximum plume height when the smoke plume reveals well-developed horizontal layering, such as in Figs. 14.9(b)-(d). The general data processing methodology for determining the maximum plume heights in the case of prescribed burns are in principle the same as that used for the analysis lidar data obtained in the vicinity of wildfires. The only difference was some transformation of the intercept formula made to avoid increased values Y0(h, φ) over the heights, h < 500 m (see Fig. 14.1). The essence of this transformation is clarified in [12, 13]. In Figs. 14.10(a)-(d) the dependence of the heights h() on different  are shown for the same cases as shown in Figs. 14.9(a)-(d). One can see that the well-defined boundary of the smoke plume can be observed only in Fig. 14.10(d), whereas in other cases the smoke plumes are significantly dispersed, having multilayer structures (Figs. 14.10(b) and (c)). Figs. 14.10. (a)-(d). Dependence of the height, h() on  extracted from the scans given in Figs. 14.9 (a)-(d). An alternative way of determining the upper boundary of the smoke plume may be based on the straightforward usage of the signals of scanning lidar. In this variant, the simplest 428 Chapter 14. Determination of the Heights of the Smoke-Plume Layering in the Vicinity of Wildfires and Prescribed Burns function, the square-range-corrected signal, may be used as a basic function. The signal as the function of the slope  and the height, h, that is, 2 S ( , h)  P ( h)h sin   , (14.10) is calculated with the straightforward formula,   2 S ( , h )  P ( h,  )   B  h sin   , (14.11) where  B  is an estimate of the constant component, B, along the slope direction ; it may be found using the total lidar signals, P(h, ,), recorded over distant ranges, at which the backscatter signal presumably vanishes, that is, where P(h)  0. The maximum plume heights of interest, the envelope function, env[S(h)], should be calculated from the set of the functions, S(, h), as   env[ S ( h )]  max  S ( min , h), ... S ( j , h ), S ( j 1 , h), ... , S ( max , h)  ,   (14.12) and then normalized to unit, that is, Snorm (h)  env[ S (h)] . Smax,max (14.13) Here Smax,max is the maximum value of env[S(h)] within the altitude range from hmin = rmin sin  min to hmax = rmaxsinmax, where rmin and rmax are minimum and maximum points of the lidar operative range. The range of the normalized function, Snorm(h), may vary from zero to one, same as the range of the function env[Y0(h)]norm. In practice, the functions S(, h) are extremely scattered, it is sensible to average the function env[S(h)] before determining Smax,max and Snorm(h). The normalization of the square-range-corrected signal allows for determining the smoke plume maximum height using the selected levels,  < 1, same as was done above for the function, Y0,(h). In Fig. 14.11, typical normalized functions, Snorm(h), are shown. Here the filled circles show the smoke-plume maximum heights determined at the level,  = 0.1, whereas the filled squares show the heights where Snorm(h) reaches its maximum value equal to unit. As stated in [4], the surface smoke-particulate concentration, predicted by models, is sensitive to the amount of plume mass injected at various heights. The knowledge of the vertical structures of the smoke plumes may allow producing better smoke dispersion predictions. Assuming that the vertical structure of plume concentration and the shape of the lidar backscatter signal are uniquely related, the utilization of the normalized squarerange-corrected signals versus height allows obtaining some information about the 429 Advances in Optics: Reviews. Book Series, Vol. 3 investigated smoke plumes. Such a method can provide experimental data that allow some estimation of the temporal variations of the smoke-plume concentration at different heights and the height, at which the smoke-plume concentration is presumably at maximum. As shown in Fig. 14.6, these heights, symbolized as h(Smax,max), are different for different times, increasing from 385 m to 1143 m during the period from 10:57 to 11:23. Fig. 14.11. The normalized functions, Snorm(h), for lidar scans under consideration. The heights, h(Smax,max) for the different times shown in the legend are shown as filled squares. The filled dots show the maximum smoke-plume heights determined for the level,  = 0.1. The rate of heat release, which can be monitored by the behavior of the parameter Smax,max, is directly related to the rate of biomass consumption. In Fig. 14.12, the variations of the heights h(Smax,max), at which the maximum backscatter signals are located, are shown during the whole period of smoke-plume profiling. One can see that initially, at 10:57, the most intensive smoke particulates were located in the vicinity of the height h(Smax,max)  400 m; then the height increases, reaching its maximum, 1244 m at 11:20. After that it decreases down to the heights 900-1000 m. The maximum backscattering intensity, Smax,max, equal to 54.9 a. u., was fixed at 11:14, then it monotonically decreased, vanishing at the end of the prescribed burn period, at 11:47. Performing such analysis, one should keep in mind that there is no simple relationship between the smoke-plume concentration and the backscatter signal obtained in the process of profiling the smoke plume. Nevertheless, two assumptions used in the above analysis look sensible enough. First, in not too dense smoke, the heights of the maximum smoke430 Chapter 14. Determination of the Heights of the Smoke-Plume Layering in the Vicinity of Wildfires and Prescribed Burns plume concentration and the maximum backscatter signal are the same, or at least, are close to each other. Second, the temporal variations of the maximum smoke-plume concentration and the maximum backscatter signal are similar. However, one should keep in mind that in some cases these assumptions may be not satisfied and the vertical profiles of these parameters may be significantly different. Fig. 14.12. Variations of the maximum values of the square-range-corrected signals, Smax,max in arbitrary units, their heights h(Smax,max), and the heights, hmax, determined at the level,  = 0.1 on August 25, 2014 during the time period from 10:57 to 11:47. 14.5. Summary The reliability of determining with lidar the smoke-plume boundaries in the vicinity of wild or prescribed fires depends on the concentration of smoke particulates injected in the atmosphere, on the presence of the multi-layering aerosol structures in it, and on the specifics of the lidar system used for atmospheric profiling. Because the lidar signal dramatically decreases with range, the level of random noise in the derived backscatter signals becomes a main factor which may significantly restrict the operative range of the lidar. The situation becomes much worse when the areas of increased backscattering has dispersed boundaries with decreased spatial gradients. This is why no commonly accepted methodology of such lidar measurements exists. The data-processing technique developed in the Fire Science Laboratory in Missoula, Montana, USA, is based on specific methodology of exposing and separating the data points of interest from noise, and on determination of the spatial and time locations of the data-point clusters. The thorough analysis of real experimental data shows that such 431 Advances in Optics: Reviews. Book Series, Vol. 3 technique can often yield significantly more reliable results. To illustrate the specifics of the inversion technique in the vicinity of wildfires, typical results of monitoring the temporal variations of the heights of the layers of increased backscattering in the smoke polluted atmosphere are presented. These measurements were performed with the scanning lidar on 4 August 2016 in the vicinity of three closely located wildfires, the Roaring Lion Fire, Observation Fire, and Cedar Fire near Missoula. This data processing methodology can be applied not only for examination of the smoke plumes originated by wildfires, but also for monitoring dust clouds, aerosol clouds created by volcano eruptions, etc. When measuring smoke plume propagation and dispersion, different situations can be encountered. Lidar monitoring of a smoke plume in vicinity of prescribed burns requires methodology significantly different from the one used for monitoring smoke plumes from wildfires. First, the prescribed burns are always spatially restricted, and the burning area is known beforehand. This allows selecting lidar measurement site quite close to the burning area. Second, stratification of the smoke plume, significantly changes during the burn. Initially the vertically stratified smoke-plume column without horizontal layering is observed. After that, dispersing processes create the specific combination of vertically stratified smoke plume and horizontal smoke-plume layering. On the last stage vertical smoke-plume structure degrades and the smoke plume particles create a horizontal (or close to horizontal) layer elevated over ground level. Accordingly, the lidar measurement methodology should follow these changes. References [1]. L. F. Radke, J. H. Lyons, P. V. Hobbs, D. A. Hegg, D. V. Sandberg, D. E. Ward, Airborne monitoring and smoke characterization of prescribed fires on forest lands in Western Washington and Oregon. Final Report, General Technical Report PNW-GTR-251, USDA Forest Service, 1990. [2]. P. V. Hobbs, J. S. Reid, J. A. Herring, J. D. Nance, R. E. Weiss, J. L. Ross, D. A. Hegg, R. D. Ottmar, C. Liousse, Particle and trace-gas measurements in the smoke from prescribed burns of forest products in the Pacific Northwest, in Biomass Burning and Global Change (J. S. Levine, Ed.), MIT Press, Cambridge, Mass., 1996, pp. 697-715. [3]. P. V. Hobbs, P. Sinha, R. J. Yokelson, T. J. Christian, D. R. Blake, S. Gao, T. W. Kirchstetter, T. Novakov, P. Pilewskie, Evolution of gases and particles from a savanna fire in South Africa, J. Geophys. Res., Vol. 108, 2003, 8485. [4]. S. M. Raffuse, K. J. Craig, N. K. Larkin, T. T. Strand, D. C. Sullivan, N. J. M. Wheeler, R. Solomon, An evaluation of modeled plume injection height with satellite-derived observed plume height, Atmosphere, Vol. 3, 2012, pp. 103-123. [5]. R. J. Dirksen, K. F. Boersma, J. de Laat, P. Stammes, G. R. van der Werf, M. V. Martin, H. M. Kelder, An aerosol boomerang: Rapid around-the-world transport of smoke from the December 2006 Australian forest fires observed from space, J. Geophys. Res., Vol. 114, 2009, D21201. [6]. V. Amiridis, E. Giannakaki, D. S. Balis, E. Gerasopoulos, I. Pytharoulis, P. Zanis, S. Kazadzis, D. Melas, C. Zerefos, Smoke injection heights from agricultural burning in Eastern Europe as seen by CALIPSO, Atmos. Chem. Phys., Vol. 10, 2010, pp. 11567-11576. 432 Chapter 14. Determination of the Heights of the Smoke-Plume Layering in the Vicinity of Wildfires and Prescribed Burns [7]. R. A. Kahn, Y. Chen, D. L. Nelson, F.-Y. Leung, Q. Li, D. J. Diner, J. A. Logan, Wildfire smoke injection heights: Two perspectives from space, Geophys. Res. Lett., vol. 35, 2008, L04809. [8]. A. J. Soja, T. D. Fairlie, D. J. Westberg, G. Pouliot, Biomass Burning Plume Injection Height Using CALIOP, MODIS and the NASA Langley Trajectory Model, http://www.epa.gov/ttnchie1/conference/ei20/session7/asoja.pdf [9]. S. Urbanski, V. Kovalev, A. Petkov, A. Scalise, C. Wold, W.M. Hao, Validation of smoke plume rise models using ground-based lidar, Proceedings of SPIE, Vol. 9239, 2014, 92391S. [10]. L. Menut, C. Flamant, J. Pelon, P. H. Flamant, Urban boundary-layer height determination from lidar measurements over the Paris area, Appl. Opt., Vol. 38, 1999, pp. 945-954. [11]. I. M. Brooks, Finding boundary layer top: Application of a wavelet covariance transform to lidar backscatter profiles, J. Atmos. and Oceanic Technol., Vol. 20, 2003, pp. 1092-1105. [12]. V. Kovalev, A. Petkov, C. Wold, S. Urbanski, W. M. Hao, Determination of smoke plume and layer heights using scanning lidar data, Appl. Opt., Vol. 48, 2009, pp. 5287-5294. [13]. V. Kovalev, A. Petkov, C. Wold, W. M. Hao, Lidar monitoring of the regions of intense backscatter with poorly defined boundaries, Appl. Opt., Vol. 50, 2011, pp.103-109. [14]. Fire Science Brief, Issue 103, http://www.firescience.gov/projects/briefs/04-1-104_FSBrief103.pdf [15]. InciWeb, Incident Information System, http://inciweb.nwcg.gov/incident/4913/ [16]. InciWeb, Incident Information System, http://inciweb.nwcg.gov/incident/4819/ [17]. InciWeb, Incident Information System, http://inciweb.nwcg.gov/incident/4874/ [18]. V. Kovalev, A. Petkov, C. Wold, S. Urbanski, W. M. Hao, Determination of the smoke-plume heights and their dynamics with ground-based scanning lidar, Appl. Opt., Vol. 54, 2015, pp. 2011-2017. 433 Chapter 15. Precision Glass Molding Chapter 15 Precision Glass Molding Weidong Liu and Liangchi Zhang1 15.1. Introduction Modern optical systems rely on advanced optical components [1-6]. The following are some examples. Precision aspherical optical lenses are critical to many consumer electronic systems such as digital cameras, mobile phones to assure image quality and reduce system weight [1, 7-9]. Micro-lens arrays are vital in the coupling of optical fibers in high-speed information transfer [10], wave-front sensors [11], high-resolution liquid crystal display panels [12], artificial compound eye structures [13], advanced laser systems [14-18] and 3D displaying systems [19]. Large-scale glass lenses are essential to space telescopes to explore the universe [20-23]. Diffractive optics with complicated surface microstructures are widely applied in diffraction gratings and advanced imaging systems [24]. Traditionally, macro glass optical components are manufactured by mechanical machining methods involving many steps such as grinding, polishing and lapping [25, 26]. In the last few decades, these machining methods have been significantly advanced owing to the development of computer numerical control (CNC) machines, single point diamond turning techniques, and multi-axial precision polishing centres [27-28]. It was these CNC machines that enable the production of aspherical and freeform glass optical components. However, the optics manufacturing using these methods is very time consuming and expensive [30]. As such, an aspheric optical lens can be some thousands of dollars [29, 30]. Moreover, Si-based optical glass can be easily damaged by severe cleavages and micro-chipping, and a diamond tool surface can quickly worn out [31-33]. Presently, micro optical components are mostly fabricated by physical and/or chemical methods such as photolithography [34, 35], wet/dry etching [36-38], and high energy beams [39-42]. However, these techniques are not cost-effective for large scale components and high-volume production. Mechanical machining methods have also been Weidong Liu Laboratory for Precision and Nano Processing Technologies, School of Mechanical and Manufacturing Engineering, University of New South Wales, Australia 435 Advances in Optics: Reviews. Book Series, Vol. 3 tried to make microstructures on optical glass surfaces using, e.g., cutting [43], grinding [44], micro-end milling [45] and sand blasting [46, 47]. However, it is very difficult to obtain damage-free optical surfaces due to the brittleness of glass materials. As a process to remove the barriers associated with mechanical machining processes outlined above, the technique of precision glass molding (PGM) has been developed [48-50]. The PGM makes use of the softening behavior of optical glass in its super-cooled liquid region (above glass transition temperature Tg) [29, 51], which enables the production of an optics in a single step. Once the surface of a mold cavity is made to have the designed features, a precision glass optical component can be thermally deformed to copy the features [29]. Hence, the PGM can significantly reduce the production time and cost. Fig. 15.1 compares schematically the PGM and machining processes, showing the convenience and efficiency of the PGM. Fig. 15.1. A comparison of PGM and machining processes [29]. The concept of PGM is straightforward and promising; but the molding of a precision macro/micro optical component is still challenging. This is because a successful PGM requires a suitable glass preform, a high quality mold with properly designed features, a well-controlled molding process, and a reliable method for quality inspection. In the last few decades, many studies have been carried out to improve the quality of optical components made by PGM [29, 52, 53]. A number of manufacturers have established technologies for making some specific optical components such as aspherical lenses, lens arrays, toroidal and ultra-micro lenses, for digital projectors, digital cameras and microscopes etc. [52, 54]. However, the production quality is still below that of the top optics fabricated by the precision machining methods [29]. This chapter will introduce the latest advancement of the PGM technology. Following the PGM production chain, the basic elements required in a typical PGM process will be introduce. Then the state-of-the-art progress and challenges of each element will be outlined, including glass material selection, preform preparation, mold fabrication, molding and quality characterization. Finally, the multi-objective optimization and digitalization of PGMs will be discussed for improving the quality of molded components. 436 Chapter 15. Precision Glass Molding 15.2. An Introduction to PGM PGM requires a suitable preform of glass, a precision mold and a controlled temperature and pressure environment, as illustrated in Fig. 15.2. A PGM process involves heating, soaking, molding, cooling, demolding and final cooling, as shown in Fig. 15.3. In the heating stage, the glass preform and mold are heated from room temperature to molding temperature, at which the glass viscosity should be low enough (normally in the range of 107 to 108 Pa·s) for copying the features of the mold cavity [29]. To avoid oxidationinduced deterioration of mold, the chamber of the PGM machine is normally filled with insert gas (e.g., nitrogen) before heating. The soaking stage provides additional time for the preform to achieve a uniform temperature distribution. Then the molding stage can be conducted either via a force-controlled process or a displacement-controlled process. Some PGM machines allow multi-step pressing in this stage to achieve a better molding quality. After molding, the optical component will usually be cooled down to glass transition temperature at a small cooling rate and pressure. The pressure will then be removed in the demolding stage, followed by a fast cooling stage for the sake of production efficiency [51, 55]. The last but most important step is the quality inspection of the molded component, to see if the PGM is successful or needs to be improved. The final quality of the molded optics is affected by many factors at different stages in the PGM process. These will be discussed in detail below. Fig. 15.2. A typical PGM process [29]. Fig. 15.3. A typical PGM processing cycle [29]. 437 Advances in Optics: Reviews. Book Series, Vol. 3 15.3. Selection of Glass and Preparation of Its Preform The index of refraction nd and Abbe number vd of optical glass are its two important properties used in the design of an optical component, although some other factors should also be considered for production efficiency and cost. The principle of PGM is that most optical glass materials become soft and moldable at high temperature above their Tg [30, 55]. Thus in general, optical glass with a lower Tg has a higher moldability. This is because a lower molding temperature generally means a smaller shape distortion at cooling, a less material property change during PGM, and a longer mold service life. Fig. 15.4 provides a summary of some moldable glass materials with their key optical properties [30]. Of those with similar optical properties, the one with a lower Tg is normally recommended for PGM. This figure also presents optical plastics with good transmission and very low Tg. As many of them are mainly used in the production by injection molding, they will not be discussed in this chapter. Fig. 15.4. Some moldable optical glass materials [1]. With the glass selected, the next step is to prepare a suitable preform for molding, and the preform quality would directly influence the final product quality. Generally, the surface quality of the final molded product cannot be better than that of the preform or the mold surface. For molding macro convex lenses, the spherical preform is usually used due to its following advantages: (1) A sphere can be easily deformed in PGM into many commonly used lens geometries; and (2) the cost-effective quality manufacture of 438 Chapter 15. Precision Glass Molding spherical preforms is already mature [29]. The diameter of a spherical preform is normally in the range of 1 mm to 8 mm [55]. Disk-shaped preforms are more suitable for molding macro concave or convex-concave lenses, micro/nano micro-lens arrays, V-groove arrays and other thin components [55]. To mold an optical component with a complex geometry or a large dimension, a near-net shape preform is often required to minimize the geometrical change in PGM, although producing such preforms is expensive [29]. 15.4. Precision Mold Fabrication The harsh environment in a PGM process calls for high performance mold materials. A mold should have excellent mechanical, physical and chemical properties at high temperature, and does not adhere with glass. Moreover, it should have acceptable machinability that allows the manufacture of precision cavity profiles and fine surface features of the mold. 15.4.1. Mold Materials in PGM According to the Tg of optical glass materials, a PGM process could be divided into three categories, i.e., ultra-low Tg PGM (Tg < 400 ˚C), low-Tg PGM (400 ˚C < Tg < 620 ˚C), and high Tg PGM (Tg > 620 ˚C) [55]. Table 15.1 lists some commonly used mold materials for each type of the PGM processes. Table 15.1. Selection of mold materials for different PGM processes [55]. Process Ultra-low Tg Low Tg Manufacturing process Cost < 400 ˚C Electroless nickelphosphor Single point diamond turning Low 400 ˚C <Tg < 620 ˚C Carbides or ceramics High Tg > 620 ˚C Carbides or ceramics Microgrinding Very high Tooling life Low Medium Very low Tg of glass Molds Micro-grinding High Tg Electroless nickel-phosphor (Ni-P) is a commonly used mold material for ultra-low Tg PGM, because of its good hardness, excellent corrosion resistance, anti-wear property, and excellent machinability using single-point diamond turning [29, 56-60]. This material was first produced by Brenner and Riddell in 1946 using electroless plating technique [61], and since then emerged as an outstanding hard coating material in industry applications. The mechanical behavior and machinability of Ni-P strongly depend on the phosphorus content. A recent work [62] reported that with the increase of phosphorus content, the structure of the electroless Ni-P coating changes from nanocrystalline to a mixture of nanocrystalline and amorphous phases, and finally to a pure amorphous phase. A record high hardness of 910 HV0.1 of as-deposited Ni-P coating was obtained at 439 Advances in Optics: Reviews. Book Series, Vol. 3 7.97 at. % phosphorus content [62]. However, the structure and properties of Ni-P is not stable at a temperature above 400 ˚C; after which its properties are deteriorated [63]. Therefore, the as-deposited Ni-P is suitable only to ultra-low Tg PGM processes. Some studies have tried to anneal it to raise its working temperature [53, 64]. Super-hard materials, such as tungsten carbide (WC) [65] and silicon carbide (SiC) [66, 67] are desirable mold materials in low-Tg PGM [29, 55]. Due to their strong covalent bonds, these materials have very high hardness, excellent thermal and corrosion resistance, and low thermal expansion. However, the machinability of these materials is normally poor because of the intrinsic brittleness. Significant efforts have been devoted to developing advanced precision machining techniques for these hard-brittle materials [68, 69]. To reduce significant workload in achieving the optical surface finish by precision grinding, a pre-shaped mold is normally needed [29, 55]. Glassy carbon and other hard carbon materials can work at a high temperature up to 1,500 °C, with good chemical stability, high hardness, wear resistance and gas impermeability [70, 71]. Therefore, they are suitable for molding high-Tg glass such as quartz glass (Tg = 1200 °C) [72]. Similar to carbide materials, glassy carbon is very brittle and thus is difficult to be machined. Some studies reported that complicated microstructures can be fabricated on the surface of glassy carbon by using high-energy beams [71, 72]. It should be noted that an inert environment is required to avoid the oxidization of glassy carbon. A recent investigation found that oxidation-induced property deterioration and surface cracking can occur at a low temperature of 500 °C [73]. To extend mold life and thus reduce cost, a mold surface is often coated. Three types of coatings have widely been used, which are noble metal coating (Re/Ir [74, 75] and Pt/Ir [76, 77]), ceramic coating (TiAlN, TiBCN, TiBC, and CrN,) [78], and hard carbon coating (diamond-like coating, amorphous carbon coating) [79]. Noble metal coatings, especially the Re/Ir coating, demonstrates very stable properties [74] and low wetting angle [75-77]. Ceramic coatings have widely been used in machining tools, and therefore are easily applicable to PGM molds [29]. However, some ceramic coatings are prone to adhesion with glass melts at high temperature [29, 76]. Hard carbons are promising coating materials for PGM; but they require an inert environment to avoid oxidation at high temperature [29, 79, 80]. The major factors that deteriorated the mold in a PGM process are: (1) the moldworkpiece mechanical interactions (pressure and friction), (2) significant thermal stress variations and adhesion during repeated heating-cooling cycles, and (3) the interface damages between the mold and coating. Therefore, in selecting mold/coating materials it is important to have an appropriate assessment or test involving all the above factors. A quick testing facility [81, 82] was proposed recently to assess the service life of mold coatings [29]. Three coatings (TiAlN, CrAlN, and Pt-Ir) on flat WC mold were studied for the molding of B270 glass. The images of the mold (pins) and glass imprints after 20 pressing steps are shown in Fig. 15.5 [82]. It is clear that the performance of the WC pin with the Pt-Ir coating is the best. 440 Chapter 15. Precision Glass Molding Fig. 15.5. Performance comparison of TiAlN, CrAlN and PtIr coatings after 20 pressing steps [82]. 15.4.2. Mold Fabrications Most mold materials are difficult to be machined. Various manufacturing techniques have been developed to reach the strict requirements of optical applications. It should be noted that the requirements for machining macro mold and micro mold (mold with micro features) are different. The former normally requires a high surface finish and form accuracy while the later has many complicated micro-scale features to be produced. 15.4.2.1. Macro Mold Fabrication Macro mold in PGM is normally fabricated by ultra-precision grinding and polishing [83]. Ultra-precision grinding is characterized by a low material removal rate with the depth of cut in the sub-micrometres range [83-85] and low feed rate [86]. Cross axis grinding is regarded as the most common grinding technique, and has been widely used in the grinding of large aspheric optical molds [88]. To allow the grinding wheel shaft to clear the mold, tilted grinding technique was developed through tilting the grinding wheel axis by an angle with respect to the normal of the workpiece spindle axis [87]. However, this grinding method requires extensive compensation to reach the required final form. Wheel grinding was developed to overcome the disadvantages of cross axis grinding and tilted grinding. Fig. 15.6 presents a typical set-up of wheel grinding, which can fabricate a variety of different tool geometries including aspheric concave cavities, aspheric convex surfaces, and microstructured surfaces [86, 87]. Freeform wheel normal grinding was also developed by functionally controlling the position of the work spindle. This technique can grind cylindrical, toric and lens array molds. However, it should be noted that the success of these grinding techniques relies on the ability to determine and compensate workpiece form errors caused by wheel wear and machine tool repeatable errors [86, 87]. Fig. 15.7 441 Advances in Optics: Reviews. Book Series, Vol. 3 presents a step-by-step procedure for achieving the required tolerance of the mold by ultraprecision grinding [86]. The performance of ultra-precision grinding is strongly influenced by the condition of grinding tools. Therefore, in-process tool dressing is required for stable, controllable, and optimal grinding processes [86]. Electrolytic inprocess dressing (ELID) grinding is widely used in ultra-precision mirror surface for hard brittle materials [52]. Fig. 15.6. Wheel normal grinding of precision molds [86, 87]. Fig. 15.7. Flow chart for grinding of optical glass molds [86]. 442 Chapter 15. Precision Glass Molding To make a mold for ultra-precision optical components, polishing/finishing of the mold surface is normally needed after grinding to remove the surface and sub-surface defects, to smooth the surface to ~1 nm rms, and to achieve a more accurate shape profile [88]. Commonly used polishing techniques include magnetorheological finishing (MRF) [89], fluid jet and bonnet polishing [90], and vibration-assisted polishing [91]. 15.4.2.2. Micro Mold Fabrication Ultra-precision milling and single point diamond turning can fabricate highly accurate surface features/textures on mold surfaces of hard materials [92]. The most successful case is machining complicated micro-features on the Ni-P surfaces by using the singlepoint diamond turning technique [63, 93, 94], which can achieve submicron form accuracy and <10 nm roughness [53]. Fig. 15.8 presents two typical miniaturised surfaces fabricated by V-shaped diamond tool and R-shaped diamond tool, respectively. In the machining process, burs, chips and some other defects could be generated. By properly making use of the material removal mechanism, such Ni-P mold surfaces with high form accuracy and low surface roughness can be fabricated through optimizing the cutting tool geometry and machining parameters, as shown in Fig. 15.9 [53]. Fig. 15.8. Single-point diamond cutting process: (a) Microgroove arrays; (b) Microlens arrays [53]. Electroless Ni-P can experience crystallization at 400 °C, which will cause mold surface deformation and thus will reduce the accuracy of PGM [53]. To avoid this problem, a heat treatment method prior to micro surface feature fabrication was proposed to eliminate the irregular concave deformation due to crystallization [53, 64]. That is, before the singlepoint machining, the amorphous Ni-P plating was heated to its annealing temperature (approximately 600 °C) to ensure the complete transformation into its crystalline state [53, 64], such that the crystallization effect in PGM above 400 °C could be avoided. Ultra-precision milling and turning can also be used for making micro-features on the surfaces of hard brittle materials such as WC and SiC. However, the cost for high-volume production is extremely high [53, 95]. As an alternative, micro-electrochemical machining and micro-electric discharge machining have been proposed for producing complex 3Dshapes on these material surfaces [53, 96]. High energy beams, such as laser and focused ion beams, have also been applied in the rapid fabrication of mold surface features [53]. 443 Advances in Optics: Reviews. Book Series, Vol. 3 However, these techniques are limited by poor surface finish. Indentation has also been used in producing microstructures on molds [53], but its accuracy and efficiency are low. Fig. 15.9. (a) Microgrooves, and (b) Micropyramids fabricated by single-point diamond turning [53]. 15.5. PGM Process After having the preform and mold manufactured, the PGM process can be done on a precision glass molding machine capable of a precise control of mold positioning, load application and temperature variation. The capability and flexibility of a molding machine in the design and implementation of customized load and temperature profiles are essential for the optimization of a PGM process. 15.5.1. Stages in a PGM Process The heating and soaking stages in a PGM process is to make the temperature distribute uniformly in the glass preform and mold. Thus a precision temperature control is central. The proportional-integral-derivative (PID) algorithm is the commonly used, which continuously calculates the difference between a desired temperature and a measured temperature and applies corrections based on the proportional, integral and derivative terms [97]. The determination of soaking time is important, because a too small soaking time cannot guarantee a uniform temperature distribution, but a too large soaking time increases the processing time and reduces the production efficiency. Finite element simulations can be used to provide a good estimation of soaking time. The molding stage can be conducted via two control modes: the displacement control mode and the force control mode. The former can provide an accurate final molding shape, while the later can achieve a stable force applied during the molding process. Molding temperature is vital in the molding stage, which should be carefully chosen for a good replication quality of optical features. If the molding temperature is too high, more profile derivation and residual stresses will be generated during the cooling process [29]. Moreover, glass is prone to adhere to a mold at a high temperature and in turn damage it. However, if the molding temperature is too low, a high pressing load is required, which would also increase the risk of mold damage. Thus molding temperature selection is a 444 Chapter 15. Precision Glass Molding particularly critical step in making complicated micro-optics such as Fresnel lenses and diffractive optical elements. Cooling is one of the most important steps in PGM. This is because (1) profile derivation, residual stresses and many defects initiate and grow in this stage, and (2) the cooling stage in a PGM process is the longest period of time which determines productivity [29]. Normally, cooling is in two-steps. In its first step, molded glass optics is cooled down to Tg at a small rate to minimise defects and residual stresses. The second step of cooling down to room temperature is faster to improve production rate. Therefore, optimizing the cooling rates is important to both the product quality and production efficiency. Demolding is normally conducted between the two cooling stages. Recently, a design of experiments (DOE) approach was used to characterize the effect of process parameters in PGM on the repeatability of the final thickness of molded N-BK7 and L-BAL35 glasses. The parameters include heating and cooling rates, soaking time, molding temperature and molding force [98]. It was found that the cooling rate has the largest impact on the repeatability of the final thickness of the molded components. However, it is still difficult to optimize a whole PGM process for a real optical product by the DOE approach because of the large combinations of processing parameters and quality inspection factors. 15.5.2. Finite Element Analysis In PGM, shape derivation and residual stresses are the major problems that influence the quality of molded optical components [29, 30, 51, 99-102]. Experimentally, however, one can only use a trail-and-error method to reduce their effects without knowing their exact details. Finite element (FE) simulations have been recognized as a useful tool for revealing the deformation mechanisms of workpiece materials and for minimizing the trial-anderror design effort [29, 51, 103-106]. Since a glass material in PGM experiences complicated thermo-mechanical deformation, a reliable constitutive description of optical glass is the basis of a reliable FE analysis. This is therefore discussed below. 15.5.2.1. Constitutive Modeling of Optical Glass A complete constitutive model of glass suitable for PGM should be able to reflect accurately the relationships of (1) the thermo-viscoelastic relationship of stress, strain, strain rate and temperature, and (2) the nonlinear temperature dependence of the material properties such as modulus, viscosity and coefficient of thermal expansion [29, 51]. Most constitutive models developed for glass were based on the classical viscoelastic models such as Maxwell model, Kelvin-Voigt model, standard linear solid model, and generalized Maxwell model. The temperature-dependent rheology was often modeled by the classical phenomenological Vogel-Fulcher-Tammann equation [51] or the thermosrheological simple assumption [107, 109]. A method was also proposed for identifying the shear relaxation modulus and the structural relaxation function by measuring the time variation of glass plate thickness [109]. The CTE variation was often described by the 445 Advances in Optics: Reviews. Book Series, Vol. 3 Tool-Narayanaswamy-Moynihan (TNM) model [109], with witch the parameterization needs structure relaxation tests and thermal expansion tests. Some studies obtained the viscoelastic properties of glass by using the relaxation data from a cylinder compression test, assuming that the material is incompressible [104]. Elasto-viscoplastic models have also been developed for glass to account for permanent plastic deformation [107]. Recently, a modulus-based constitutive model was developed for analyzing PGM processes numerically [51]. As summarized in Table 15.2, all the temperature-dependent material properties in this model are determined by the relationship between the elastic moduli and microstructure of a material. To do this, the strain and stress tensors are divided into their volumetric and deviatoric parts. The relationship between deviatoric stress and strain is regarded as a standard linear solid (SLS) [108], and the volumetric part is viewed as a thermal elastic relationship because the bulk viscosity of super-cooled glass can be considered to be infinite [51]. The material’s temperature-dependent moduli are measured by an impulse excitation method [109]. The temperature-dependent viscosity is directly linked to the shear modulus by using the shoving model [109, 110]. The CTE of glass is predicted through the Young’s modulus based on a phenomenological TNM model [105, 106], in which the parameters in TNM model can be determined by the temperature-dependent modulus changes in the impulse excitation method. The above constitutive model with the measured/derived parameters has been verified and numerically programmed [51], and has been used to reveal the formation mechanisms of profile derivation and residual stress in molding a macro glass lens. 15.5.2.2. Mechanisms of Profile Distortion Profile accuracy, including geometrical accuracy and surface finish quality, is critical to an optical component [29]. The shape derivation of a molded lens can be as high as 20 μm, about 20 times higher than the deviation allowed by the optical design specifications [29, 111]. Thus the quality of a molded optical component depends largely on the profile distortion during its PGM process [29]. The FE analysis is particularly useful to understand the mechanisms. Fig. 15.10a shows the evolution of a lens profile during a typical PGM process. In the molding stage, the preform was compressed to comply with the mold cavity. The subsequent demolding did not lead to a significant shape deviation [29]. In the cooling stage, however, a large shape deviation occurred near the center of the lens. The derivation kept increasing until the internal temperature reduced to below Tg [29]. Fig. 15.10b presents the evolution of profile derivation with respect to the mold in the radial direction. It was found that the large deviations near the center and the edge of the lens are due to the cooling-induced shrinkage and edge effect, respectively [51]. It should be noted that for a precision lens, the allowed center thickness change is about 25 μm [55], and the maximum deviation of overall surface shape should be within several micrometers or smaller [105, 106]. This analysis indicated that an ultra-precision mold is not enough for molding a high quality glass optical component. Compensation should be made in the model design stage [29]. 446 Chapter 15. Precision Glass Molding Table 15.2. Modulus-based constitutive model for optical glass [51]. Relationship Stress and strain Volumetric relationship Deviatoric relationship Equation  ij  eij  tr( ) ij 3 ,  ij  Sij  tr ( ) ij 3 tr ( ) / 3    T  tr ( ) / 9 K (1  Sij S ij G Gr )eij  r eij   G s 2G 2 s Viscosity variation s  0 exp(VcG (T ) / kBT ) Thermal expansion    G   L   G T f T T f  T  ((TT )) M p (   ' ) 0 Structure relaxation description dT d ' d ' t    1  p dt  0 M p ( )  exp[  (  pr )  ]  p   0 expxH / RT  (1  x)H / RT f  Variable definition εij - strain tensor, σij – stress tensor, eij - deviatoric strain, Sij - deviatoric stress, tr(ε) - the trace of the strain tensor, tr(σ) - the trace of the stress tensor, δij - Kronecker delta, K - bulk modulus, α - the coefficient of thermal expansion, T - temperature, Gr - the modulus in the elastic branch of the SLS model, G - the shear modulus in the Maxwell branch of the SLS model, ηs - shear viscosity, η0 - reference viscosity, kB - Boltzmann constant, Vc - characteristic temperature-independent microscopic volume, G∞(T) - instantaneous shear modulus, G - the reference CTE at low temperature glassy state, L - the reference CTE at high temperature liquid state, Tf - effective temperature, T0 - the reference temperature, Mp(ξ) - the structural relaxation function, ξ - reduced time, τp - structural relaxation time, ∆H - the active energy, R - the ideal gas constant, τ0, x, β – constants. 447 Advances in Optics: Reviews. Book Series, Vol. 3 (a) (b) Fig. 15.10. (a) The evolution of a lens shape in PGM, (b) the deviation with respect to the mold cavity [51]. Based on a parametric study [105, 106], it was found that the structural relaxation of glass in the supercooled liquid region was the primary cause for profile distortion in PGM. As aforementioned, the structural relaxation of glass in an FE analysis is normally described by the TNM model, in which the activation energy and relaxation time are the key parameters. A novel method has been developed recently [29, 109] to identify these parameters based on an impulse excitation technique. Some studies [105, 106] also suggested that the most critical stage to reduce lens distortion is the beginning of demolding, in which thermal expansion coefficients of the mold material and internal stresses of the lens play an important role. Other important factors include molding temperature, loading-unloading paths and cooling rates [105, 106]. 15.5.2.3. Residual Stresses Most glass preforms do not have significant internal residual stress because they have been well annealed by manufacturers. In a PGM process, however, a high cooling rate is often used to increase production efficiency, which can easily lead to internal residual stresses [29, 112-114]. Such residual stresses can severely alter the local density of optical glass, and lead to inhomogeneous refractive index in an optical lens [29, 112]. For example, a residual stress of 3 MPa in P-BK7 glass lens can bring about a variation of refractive index of 4×10-4, and thus produce unwanted changes in the light path, intensity, and deterioration of image quality [29, 115, 116]. Therefore, it is important to understand the formation mechanism of residual stresses in PGM and its effect on optical properties. A recent comprehensive investigation [51] revealed the formation mechanisms of residual stresses in molding a convex-convex optical lens. Fig. 15.11a shows a typical distribution of residual stress (von Mises stress) in a molded lens. Two minima of the von Mises stresses locate symmetrically close to the top and bottom subsurfaces [29, 51]. To reveal the formation mechanism of the residual stress, the stress evolution at the top, middle and bottom points of the lens during the PGM process were studied as shown in Fig. 15.11b 448 Chapter 15. Precision Glass Molding [29, 51]. It is clear that the internal stresses before 270 s were very small except in the initial pressing stage [29, 51]. At around 270 s (in cooling stage), however, the internal stresses increased to a plateau till the end of the PGM process to form residual stresses [29, 51]. To further reveal the stress distribution around 270 s, the internal stress distributions along the central line through the lens thickness were shown in Fig. 15.11d. It was found that both the magnitude and gradient of the internal stresses increase significantly from 279.5 s to 299.5 s. The stress increase was related to the heterogeneous evolution of CTE of the optical glass during PGM [51], as revealed in Fig. 15.11d. Due to the inhomogeneous temperature distribution in the lens, the changes of the CTE at different positions are asynchronous [29, 51]. The difference of CTEs reached the maximum at 280 s (see the insert of Fig. 15.11d), leading to the significant increase of the magnitude and gradient of the internal stresses. (a) (c) (b) (d) Fig. 15.11. (a) The distributions of residual von Mises stress; (b) Variations of the von Mises stresses with time at different points in the lens; (c) the stress distributions along the central line through the lens thickness, and (d) the variations of CTEs at different points with time [51]. Since residual stresses arise due to the sharp increase of internal stresses during cooling, controlling the cooling rate should be able to reduce residual stresses. Some studies proved that the duration of cooling from the molding temperature to Tg is important in minimizing residual stresses [103] and that the residual stresses in a molded lens can be controlled to a very small value if a proper cooling is applied [112]. A recent study [51] has explicitly 449 Advances in Optics: Reviews. Book Series, Vol. 3 shown that a good strategy of minimizing residual stress would be to use a small cooling rate in the first stage, and then a larger cooling rate in the second stage for the sake of production efficiency [51]. It is noted that residual stress could also be affected by changing the rheology behavior of glass at molding temperature, the friction at the glass/mold interface, and the time/temperature at which the demolding is applied [105, 106]. 15.6. Quality Inspection Techniques Quality inspection is the last but most important step in the production chain of PGM. Without a reliable and systematic quality characterization method, one cannot assess and improve a PGM process. Standard optical inspection methods have been established for machined macro optical components, and some can be utilized for assessing the quality of molded optical components. However, similar standard inspection methods for micro optical components are still under development. In addition, residual stress formed during the PGM process can significantly alter the optical properties and should be inspected as well. In the following sections, three quality inspection aspects for molded optical components will be introduced, i.e., surface characterization, internal residual stress characterization and the latest inspection methods for micro optical components. 15.6.1. Surface Characterization Surface quality is the key to the performance of optical components because the functional light refraction normally occurs in the surface. Surface roughness and profile derivation are two major factors that influence the surface quality and optical functions of a molded optical component. Therefore, they need to be accurately characterized and controlled within tolerance. Surface quality characterization methods can be divided into two categories, i.e. contact and non-contact types. At the microscopic scale, a contact type stylus profiler using electronic amplification is the most common tool [117]. However, its measurement accuracy can be significantly affected by the stylus size, scan area size, scan speed, etc. At a nanoscale, atomic force microscopy has been used [117]; but its scan area is limited to tens of micrometers. Non-contact measurements are normally achieved by optical methods, such as interferometric techniques and confocal microscopy [118]. The Fizeau configuration is the most commonly used interferometer because of its succinct set-up design [118, 119]. As shown in Fig. 15.12, the Fizeau interferometer consists of two plano-parallel glass plates aligned at an angle. Interference occurs between the beams reflected from the surfaces of the plates facing each other. Interferometric techniques can quickly characterize the entire surface in a single measurement with sub-nanometric resolution. In measuring aspherical or freeform optical components, however, a correction element such as a null lens or computer generated hologram are needed to generated a suitable reference wave front [119]. It is noted that the measurement of interferometer is very sensitive to 450 Chapter 15. Precision Glass Molding the environmental influences [119]. Moreover, certain surface structures such as high curvature area or steps in micro optics would lead to optical edge artefacts [120]. Fig. 15.12. A schematic setup of Fizeau interferometer [119]. 15.6.2. Residual Stress Characterization Internal residual stress after PGM can significantly affect the optical performance of molded optics [94, 121, 122]. It is noted that most transparent materials possess birefringence property, in which the difference of the principal refractive indexes is proportional to the difference of the principal stresses. Therefore, the measurement of internal residual stress can be converted to an optical problem [94, 122]. Some researchers [123] measured the optical retardation using a plane polariscope, and then compared the experimental results with numerical simulation as a validation of the modeling approach [123]. Some others [124] measured the whole field residual birefringence distribution of a molded P-SK57TM lens by a six step phase shifting technique. Fig. 15.13 presents the schematic of the experimental setup. To increase the measurement accuracy, the molded lens was placed in a tank with a glass bottom containing a liquid of matching refractive index. Fig. 15.14 shows the distribution of retardation in the P-SK57TM glasses lens molded with two different flow rates of N2 gas [124]. This technique can directly measure the retardation and the direction of the maximum principal residual stress. However, it is very difficult to work out the individual components of the residual stress, which needs an effective stress separation algorithm [125]. 15.6.3. Micro-Optics Characterization and Standardization All the techniques aforementioned can be used in characterizing micro-optics if the resolution is high enough. However, due to the small sizes and complicated structures of micro optics, the terminologies and definitions widely used for macro optics are no longer suitable for micro optics [126]. For example, in a microlens it is very difficult to find the principal plane and the optimum image plane to define the focal length. Therefore, effective focal length has been used instead, which is defined as the distance from the vertex of the microlens to the position of the focus given by locating the maximum of the power density distribution. Therefore, specific international standards need to be established to unify the terminology and characterization methods [126]. 451 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 15.13. Experimental setup for measuring the residual birefringence [124]. International standardization activities for microlenses were initialed in Japan in 1991. The first part of ISO standard of microlens arrays, ISO 14880-1 Microlens array Part 1 (Vocabulary), was published in 2001. After that, the test method standards were summarized in three different parts: ‘Part 2 Test methods for wavefront aberrations’, ‘Part 3 Test methods for optical properties other than wavefront aberrations’ and ‘Part 4 Test methods for geometrical Properties’. In 2006, a new part of the standard series was initiated to explain the different methods for testing microlenses and microlens arrays [126]. Fig. 15.15 shows the development road map of the microlens array standards. However, similar work has not been done for other micro optics. 452 Chapter 15. Precision Glass Molding Fig. 15.14. Experimentally measured retardation in a molded aspherical lens with two different flow rates of N2 gas [124]. Fig. 15.15. Development road map of microlens array standards [126]. 15.7. Optimization of PGM Process As have been discussed in previous sections, the quality of molded optical components is influenced by a series of factors [29], and their relationship is complex and highly nonlinear. Therefore, it is difficult to improve the PGM process by using a trial-and-error method. A process optimization with the aid of a reliable numerical simulation is a cost-effective way to minimize the problems in the manufacturing chain of lens production [29]. 15.7.1. Optimization Strategy To optimize a process, objective functions, optimization algorithms and criteria are required. The objective function for the lens PGM is not a simple equation but a FE numerical simulation (see Fig. 15.16) [99, 127, 128]. The FE analysis should be able to 453 Advances in Optics: Reviews. Book Series, Vol. 3 generate the parameters to be optimized from the optimization algorithms, and produce results to be assessed by the criteria. Some studies have carried out a single objective optimization process, such as reducing the shape deviation or minimizing the residual stresses. On the contrary, multi-objective optimization of PGM process is much more difficult. Fig. 15.16. A typical optimization process [131]. 15.7.2. Mold Shape Optimization Considering that the shape distortion of molded optical components is inevitable due to the thermal shrinkage in PGM process, the mold geometry and dimension must be optimized to compensate such effects in the initial design stage [99, 127, 128]. Different algorithms have been used for optimizing the mold shape, such as iterative algorithms [129, 130], sequential quadratic programming methods [128] and iterative deviation methods [99]. Based on the simplex method, a numerical platform was recently established for compensating aspherical molds [131]. A formulated aspherical lens surface was defined by Eq. (15.1), where X is the distance from the lens axis, Y is the Y-component of the distance from the vertex, R is the radius of curvature, k is the conic constant and a is the correction coefficient of high order terms [131]. X2 Y(X )  R (1  1  (1  k ) 2 X ) R2  aX 4 . (15.1) The profile-mean-square-deviation (PMSD) was selected as the optimization objective. PMSD can be calculated by Eq. (15.2), where N is the node number on the lens surface, yi  yˆ i represents the shape derivation at the ith node. Considering that a high-quality optical lens requires that the PMSD < 1 μm [132], the following optimization criterion in this optimization was applied [131]. PMSD  454 2  iN1 ( yi  yˆ i )  1 m . N (15.2) Chapter 15. Precision Glass Molding Fig. 15.17 shows the changes of the PMSD by optimizing R and k simultaneously. The value of a was set to zero during the optimization process [131]. The criterion was satisfied after 21 optimization cycles, and the optimized parameters were R = 11.819 mm and k = 2.0365 [131]. Fig. 15.17. (a) Variation of the PMSD during the optimization of R and k, and (b) the corresponding parameters [131]. To further reveal the compensation mechanism [131], the shape derivations of the molded lens along its radial direction with and without the mold optimization were compared as shown in Fig. 15.18a. It is clear that using the optimized mold could effectively reduce the large shape deviation near the edge. The evolutions of the PMSD during the molding process with and without the mold shape optimization were also compared as shown in Fig. 15.18b. It is clear that the deviation can be effectively reduced during cooling by using the optimized mold shape [131]. Fig. 15.18. (a) Comparison of the profile deviations of the molded lens along the radial direction with and without a die optimization, and (b) the evolution of the PMSD during the molding processes [131]. 455 Advances in Optics: Reviews. Book Series, Vol. 3 15.7.3. Residual Stress Optimization As revealed above, residual stresses in a molded optical component form in the cooling stage [51]. Therefore, it is reasonable to optimize the cooling curve for reducing the residual stress. An optimization trial [133] has been on the whole cooling stage in PGM. In this study, the cooling curve in the temperature region Tg -50 oC to Tg +100 oC was divided by 7 key points. Then the position of these points was optimized to reach the residual stress threshold. Fig. 15.19 shows the cooling curves before and after optimization for different stress threshold. Fig. 15.19. Initial and optimized cooling curves [133]. As have been clarified previously, lens cooling in PGM is in two stages. The first cooling stage influences the formation of residual stresses but the second stage does not. Thus the residual stress optimization can focus on the first cooling stage [131]. The division point between the first and the second cooling stages was selected as the optimization parameters (see Fig. 15.20a). The von Mises stress at the lens center (point N in Fig. 15.20b) was used as the optimization target with the threshold of 2.5 MPa [131]. Simplex method was used to optimize the position (t2, T2) to minimize the residual von Mises stress at point N. Figs. 15.21a and 15.21b present the evolution of the parameters during optimization and the corresponding residual stress changes. It can be seen that the optimization enables the residual stress reduction until reaching the criteria (<2.5 MPa). 15.7.4. Multi-Objective Optimization Multi-objective optimization is much more challenging than the single-objective optimization mentioned above [29]. With the progress of computational capability and 456 Chapter 15. Precision Glass Molding algorithm in recent years, multi-objective optimization approaches has been used in optimizing machining parameters. Non-dominated sorting genetic algorithm (NSGA-II) is the most popular multi-objective optimization tool [134], which has been applied in the design of injection molding processes [135], in getting good laser brazing parameters [136], in improving the hard turning performance of bearing steel [137], and in optimizing the process parameters of wire electrical discharge machine [138]. However, similar work on PGM is not available. (a) (b) Fig. 15.20. (a) Loading and temperature history of PGM, and (b) the distribution of residual von Mises stress in a molded glass [131]. (a) (b) Fig. 15.21. (a) Changes of parameters during optimization, and (b) the corresponding residual stresses [131]. 457 Advances in Optics: Reviews. Book Series, Vol. 3 To optimize a PGM production process, an effort has been place to digitize the whole manufacturing chain of PGM [94]. A web-based software for PGM was developed for all users to share the massive production data, including process developers and quality control inspectors. Based on these data, one can analyses the correlations between “input” preform, mold, processing parameters and “output” quality of molded optical components [29]. In this way, the optimization of the whole manufacturing chain could be achieved. However, how to effectively use this “Big Data” to make added value is still challenging. 15.8. Summary This chapter has introduced the specifications, challenges and latest progress of PGM, as a promising technique to manufacture better and lower-cost optical components. A brief summary and perspectives are given below: (1) PGM can produce glass optical components in a single-step process, and thus can significantly reduce the time and cost compared with traditional machining methods. (2) The quality of molded optical components strongly depends on the preform, mold and process control parameters. Defects in these preparatory elements will be transferred to the final products. Fabricating ultraprecision mold is one of the most important and difficult steps in the manufacturing chain of PMG. (3) FE-based numerical methods can simulate the PGM process and reveal the formation mechanism of shape distortion and residual stresses in molded optical components. However, a suitable constitutive model, which can describe the complicated thermomechanical deformation of glass over a wide range of temperature, should be developed and used. (4) Standard quality characterization methodologies should be developed, especially for micro optical components. They are the foundation to improve the PGM process. (5) Multi-objective optimization could be an effective way to improve the quality of molded optical components. Digitization of a real PGM process can provide massive insitu data to analyse the correlations between “input” preform, mold, processing parameters and “output” quality of molded optical components. Acknowledgements The work presented in this chapter was financially supported by the Australian Research Council. References [1]. R. E. Fischer, B. Tadic-Galeb, P. R. Yoder, Optical System Design, McGraw-Hill, New York, 2008. 458 Chapter 15. Precision Glass Molding [2]. M. U. Erdenebat, K. C. Kwon, E. Dashdavaa, Y. L. Piao, K. H. Yoo, G. Baasantseren, et al., Advanced 360-degree integral-floating display using a hidden point removal operator and a hexagonal lens array, Journal of the Optical Society of Korea, Vol. 18, Issue 6, 2014, pp. 706-713. [3]. J. L. Li, S. Hu, L. X. Zhao, Study of the influence of the shape of projection lens focal plane on the focus control of advanced lithography, Optik, Vol. 125, Issue 22, 2014, pp. 6775-6777. [4]. C. M. Tsai, Y. C. Fang, C. Z. Hsiang, A study of optical design and optimization applied to lens module of laser beam shaping of advanced modern optical device, Proceedings of SPIE, Vol. 8130, 2011, 81300W. [5]. Y. Zhang, T. D. Milster, J. S. Kim, S. K. Park, Advanced lens design for bit-wise volumetric optical data storage, Japanese Journal of Applied Physics. Part 1 – Regular Papers Short Notes & Review Papers, Vol. 43, Issue 7B, 2004, pp. 4929-4936. [6]. R. Hentschel, B. Braunecker, H. J. Tiziani, Advanced Optics Using Aspherical Elements, SPIE Press, Bellingham, 2008. [7]. K. Matsusaka, S. Ozawa, R. Yoshida, T. Yuasa, Y. Souma, Ultracompact optical zoom lens for mobile phone, Digital Photography III, Vol. 6502, 2007, pp. 50203-50203. [8]. H. P. D. Shieh, Y. P. Huang, K. W. Chien, Micro-optics for liquid crystal displays applications, Journal of Display Technology, Vol. 1, Issue 1, 2005, pp. 62-76. [9]. Y. C. Fang, H. L. Tsai, Y. H. Chien, W. Y. Wu, C. C. Hu, H. W. Su, 2X optical digital zoom lens with short total length and extremely small front aperture for two million pixel CMOS on mobile phones, Proceedings of SPIE, Vol. 6342, 2006, 63421I. [10]. S. Tong, H. Bian, Q. Yang, F. Chen, Z. Deng, J. Si, et al., Large-scale high quality glass microlens arrays fabricated by laser enhanced wet etching, Optics Express, Vol. 22, Issue 23, 2014, pp. 29283-29291. [11]. B. C. Platt, R. Shack, History and principles of Shack-Hartmann wavefront sensing, Journal of Refractive Surgery, Vol. 17, Issue 5, 2001, pp. S573-S577. [12]. M. Sieler, P. Schreiber, A. Brauer, Microlens array based LCD projection display with software-only focal distance control, Proceedings of SPIE, Vol. 8643, 2013, 86430B. [13]. K. H. Jeong, J. Kim, L. P. Lee, Biologically inspired artificial compound eyes, Science, Vol. 312, Issue 5773, 2006, pp. 557-561. [14]. T. R. M. Sales, Structured microlens arrays for beam shaping, Optical Engineering, Vol. 42, Issue 11, 2003, pp. 3084-3085. [15]. H. Yumoto, H. Mimura, T. Koyama, S. Matsuyama, K. Tono, T. Togashi, et al., Focusing of X-ray free-electron laser pulses with reflective optics, Nature Photonics, Vol. 7, Issue 1, 2013, pp. 43-47. [16]. V. V. Strelkov, High-order optical processes in intense laser field: Towards nonperturbative nonlinear optics, Physical Review A, Vol. 93, Issue 5, 2016, 053812. [17]. B. Dong, R. Wang, Laser beam cleanup using improved model-based wavefront sensorless adaptive optics, Chinese Optics Letters, Vol. 14, Issue 3, 2016, 031406. [18]. C. Holly, M. Traub, D. Hoffmann, C. Widmann, D. Brink, C. Nebel, et al., Monocrystalline CVD-diamond optics for high-power laser applications, Proceedings of SPIE, Vol. 9741, 2016, 974104. [19]. H. Arimoto, B. Javidi, Integral three-dimensional imaging with digital reconstruction, Optics Letters, Vol. 26, Issue 3, 2001, pp. 157-159. [20]. M. Clampin, Recent progress with the JWST Observatory, Proceedings of SPIE, Vol. 9143, 2014, 914302. [21]. C. H. Rao, L. Zhu, X. J. Rao, L. Q. Zhang, H. Bao, L. Kong, et al., Second generation solar adaptive optics for 1-m New Vacuum Solar Telescope at the Fuxian Solar Observatory, Chinese Optics Letters, Vol. 13, Issue 12, 2015, 120101. [22]. D. H. Hashiguchi, J. M. Murder, R. A. Paquin, The evolution of beryllium space telescope optics, Sampe Journal, Vol. 51, Issue 6, 2015, pp. 40-48. 459 Advances in Optics: Reviews. Book Series, Vol. 3 [23]. G. P. Lousberg, V. Moreau, J. M. Schumacher, M. Pierard, A. Somja, P. Gloesener, et al., Design and analysis of an active optics system for a 4-m telescope mirror combining hydraulic and pneumatic supports, Proceedings of SPIE, Vol. 9626, 2015, 962624. [24]. W. Moreshead, J. L. Nogues, L. Howell, B. F. Zhu, Replication of diffractive optics in silica glass, Proceedings of SPIE, Vol. 2689, 1996, pp. 142-152. [25]. R. N. Doetsch, History of microscope, Science, Vol. 133, Issue 345, 1961, pp. 946-947. [26]. M. Herzberg, History of image of microscope, Optik, Vol. 39, Issue 2, 1973, pp. 93-98. [27]. W. Konig, N. Koch, Aspherical glass surfaces and options for their production, F&MFeinwerktechnik & Messtechnik, Vol. 98, Issue 6, 1990, pp. 254-258. [28]. D. J. Nicholas, J. E. Boon, The generation of high-precision aspherical surfaces in glass by Cnc-Machining, Journal of Physics D-Applied Physics, Vol. 14, Issue 4, 1981, pp. 593-600. [29]. L. C. Zhang, W. D. Liu, Precision glass molding: Toward an optimal fabrication of optical lenses, Frontiers of Mechanical Engineering, Vol. 12, Issue 1, 2017, pp. 3-17. [30]. J. Deegan, Precision Glass Molding Technical Brief, Rochester Precision Optics, W. Henrietta, NY, USA, 2007, pp. 1-17. [31]. N. lkawa, R. R. Donaldson, R. Komanduri, W. Konig, T. H. Aachen, P. A. McKeown, et al., Ultraprecision metal cutting – The past, the present and the future, Annals of the CIRP, Vol. 40, Issue 2, 1991, pp. 587-594. [32]. T. Nakasuji, S. Kodera, S. Hara, H. Matsunaga, N. lkawa, S. Shirnada, Diamond turning of Brittle materials for optical components, Annals of the CIRP, Vol. 39, Issue 1, 1990, pp. 89-92. [33]. F. Z. Fang, X. D. Liu, L. C. Lee, Micro-machining of optical glasses – A review of diamondcutting glasses, Sadhana, Vol. 28, 2003, pp. 945-955. [34]. A. Revzin, R. J. Russell, V. K. Yadavalli, W. G. Koh, C. Deister, D. D. Hile, et al., Fabrication of poly(ethylene glycol) hydrogel microstructures using photolithography, Langmuir, Vol. 17, Issue 18, 2001, pp. 5440-5447. [35]. K. Totsu, K. Fujishiro, S. Tanaka, M. Esashi, Fabrication of three-dimensional microstructure using maskless gray-scale lithography, Sensors and Actuators A: Physical, Vol. 130, 2006, pp. 387-392. [36]. S. E. Jee, P. S. Lee, B. J. Yoon, S. H. Jeong, K. H. Lee, Fabrication of microstructures by wet etching of anodic aluminum oxide substrates, Chemistry of Materials, Vol. 17, Issue 16, 2005, pp. 4049-4052. [37]. K. Murakami, Y. Wakabayashi, K. Minami, M. Esashi, Cryogenic dry etching for high aspect ratio microstructures, in Proceedings of the Conference: Micro Electro Mechanical Systems (MEMS'93), 1993, pp. 65-70. [38]. Ü. Sökmen, A. Stranz, S. Fündling, H. H. Wehmann, V. Bandalo, A. Bora, et al., Capabilities of ICP-RIE cryogenic dry etching of silicon: review of exemplary microstructures, Journal of Micromechanics and Microengineering, Vol. 19, Issue 10, 2009, 105005. [39]. C. H. Chao, S. C. Shen, J. R. Wu, Fabrication of 3-D submicron glass structures by FIB, Journal of Materials Engineering and Performance, Vol. 18, Issue 7, 2009, pp. 878-885. [40]. S. Reyntjens, R. Puers, A review of focused ion beam applications in microsystem technology, Journal of Micromechanics and Microengineering, Vol. 11, Issue 4, 2001, pp. 287-300. [41]. R. Wirth, Focused Ion Beam (FIB) combined with SEM and TEM: Advanced analytical tools for studies of chemical composition, microstructure and crystal structure in geomaterials on a nanometre scale, Chemical Geology, Vol. 261, Issue 3-4, 2009, pp. 217-229. [42]. S. Mailis, I. Zergioti, G. Koundourakis, A. Ikiades, A. Patentalaki, P. Papakonstantinou, et al., Etching and printing of diffractive optical microstructures by a femtosecond excimer laser, Applied Optics, Vol. 38, Issue 11, 1999, pp. 2301-2308. [43]. F. Z. Fang, L. J. Chen, Ultra-precision cutting for ZKN7 glass, CIRP Annals, 2000, pp. 17-20. 460 Chapter 15. Precision Glass Molding [44]. E. S. Lee, S. Y. Baek, A study on optimum grinding factors for aspheric convex surface micro-lens using design of experiments, International Journal of Machine Tools & Manufacture, Vol. 47, Issue 3-4, 2007, pp. 509-520. [45]. T. Ono, T. Matsumura, Influence of tool inclination on brittle fracture in glass cutting with ball end mills, Journal of Materials Processing Technology, Vol. 202, Issue 1-3, 2008, pp. 61-69. [46]. S. Bouzid, N. Bouaouadja, Effect of impact angle on glass surfaces eroded by sand blasting, Journal of the European Ceramic Society, Vol. 20, Issue 4, 2000, pp. 481-488. [47]. S. M. Chen, H. S. Kwok, Light extraction from organic light-emitting diodes for lighting applications by sand-blasting substrates, Optics Express, Vol. 18, Issue 1, 2010, pp. 37-42. [48]. A. Y. Yi, A. Jain, Compression molding of aspherical glass lenses – A combined experimental and numerical analysis, Journal of the American Ceramic Society, Vol. 88, Issue 3, 2005, pp. 579-586. [49]. Y. Aono, M. Negishi, J. Takano, Development of large aperture aspherical lens with glass molding, Proceedings of SPIE, Vol. 4231, 2000, pp. 16-23. [50]. M. Muranaka, Y. Noro, M. Takagi, K. Wada, Precise aspherical plastic lens for electronic appliances - Molding technology and applications, in Integration of Fundamental Polymer Science and Technology – 5, Elsevier Science Publishers Ltd., 1991, pp. 385-389. [51]. W. D. Liu, L. C. Zhang, Thermoforming mechanism of precision glass molding, Applied Optics, Vol. 54, Issue 22, 2015, pp. 6841-6849. [52]. S. H. Yin, H. P. Jia, G. H. Zhang, F. J. Chen, K. J. Zhu, Review of small aspheric glass lens molding technologies, Frontiers of Mechanical Engineering, Vol. 12, Issue 1, 2017, pp. 66-76. [53]. T. F. Zhou, X. H. Liu, Z. Q. Liang, Y. Liu, J. Q. Xie, X. B. Wang, Recent advancements in optical microstructure fabrication through glass molding process, Frontiers of Mechanical Engineering, Vol. 12, Issue 1, 2017, pp. 46-65. [54]. U. Fotheringham, A. Baltes, P. Fischer, P. Hohn, R. Jedamzik, C. Schenk, et al., Refractive index drop observed after precision molding of optical elements: A quantitative understanding based on the Tool-Narayanaswamy-Moynihan model, Journal of the American Ceramic Society, Vol. 91, Issue 3, 2008, pp. 780-783. [55]. M. Schaub, J. Schwiegerling, E. C. Fest, A. Symmons, R. H. Shepard, Molded Optics Design and Manufacture, CRC Press, New York, USA, 2011. [56]. J. Yan, T. Oowada, T. Zhou, T. Kuriyagawa, Precision machining of microstructures on electroless-plated NiP surface for molding glass components, Journal of Materials Processing Technology, Vol. 209, Issue 10, 2009, pp. 4802-4808. [57]. J. Y. Hou, C. S. Wang, Y. Bai, Research on crafts of Ni-P-SiC electroless composite plating on surface performances of the mould, Advanced Polymer Science and Engineering, Vol. 221, 2011, pp. 389-393. [58]. Y. Sawa, K. Yamashita, T. Kitadani, D. Noda, T. Hattori, Fabrication of high hardness Ni mold with electroless nickel-boron thin layer, Microsystem Technologies, Vol. 16, Issue 8-9, 2010, pp. 1369-1375. [59]. K. S. Chon, Y. Namba, Single-point diamond turning of electroless nickel for flat X-ray mirror, Journal of Mechanical Science and Technology, Vol. 24, Issue 8, 2010, pp. 1603-1609. [60]. T. Yamaguchi, M. Higuchi, S. Shimada, T. Kaneeda, Tool life monitoring during the diamond turning of electroless Ni-P, Precision Engineering, Vol. 31, Issue 3, 2007, pp. 196-201. [61]. A. Brenner, G. E. Riddell, Nickel plating on steel by chemical reduction, Journal of Research of the National Bureau of Standards, Vol. 37, Issue 1, 1946, pp. 31-34. [62]. M. Yan, H. G. Ying, T. Y. Ma, Improved microhardness and wear resistance of the asdeposited electroless Ni-P coating, Surface & Coatings Technology, Vol. 202, Issue 24, 2008, pp. 5909-5913. 461 Advances in Optics: Reviews. Book Series, Vol. 3 [63]. R. Parkinson, Status of the nickel plating and electroforming industry, Transactions of the Metal Finishers Association of India, Vol. 5, Issue 4, 1996, pp. 231-236. [64]. T. F. Zhou, J. W. Yan, Z. Q. Liang, X. B. Wang, R. Kobayashi, T. Kuriyagawa, Development of polycrystalline Ni-P mold by heat treatment for glass microgroove forming, Precision Engineering, Vol. 39, 2015, pp. 25-30. [65]. C. L. Chao, C. J. Chang, C. C. Chen, W. C. Chou, K. J. Ma, Precision grinding of tungsten carbide mold insert for molding of sub-millimeter glass aspheric lenses, Proceedings of SPIE, Vol. 8769, 2013, 87691U. [66]. J. Shin, S. I. Tanaka, A. Esashi, Nanostructured silicon carbide molds for glass press molding, in Proceedings of the 2nd IEEE International Conference on Nano/Micro Engineered and Molecular Systems (NEMS’07), Vols. 1-3, 2007, pp. 1152-1155. [67]. K. O. Min, S. Tanaka, M. Esashi, Micro/nano glass press molding using silicon carbide molds fabricated by silicon lost molding, in Proceedings of the 18th IEEE International Conference on Micro Electro Mechanical Systems (MEMS’05), 2005, pp. 475-478. [68]. Z. H. Wu, W. D. Liu, L. C. Zhang, Revealing the deformation mechanisms of 6H-silicon carbide under nano-cutting, Computational Materials Science, Vol. 137, 2017, pp. 282-288. [69]. T. Bletek, F. Klocke, M. Hunten, O. Dambon, Dressing of fine grained diamond grinding wheels for ultra precision grinding of structured molds in brittle hard materials, Proceedings of SPIE, Vol. 8884, 2013, 888408. [70]. H. Maehara, H. Murakoshi, Quartz glass molding by precision glass molding method, IEEJ Transactions on Sensors and Micromachines, Vol. 122, Issue 10, 2002, pp. 494-497. [71]. S. W. Youn, M. Takahashi, H. Goto, R. Maeda, Microstructuring of glassy carbon mold for glass embossing – Comparison of focused ion beam, nano/femtosecond-pulsed laser and mechanical machining, Microelectronic Engineering, Vol. 83, Issue 11-12, 2006, pp. 2482-2492. [72]. K. Prater, J. Dukwen, T. Scharf, H. P. Herzig, S. Ploger, A. Hermerschmidt, Micro-structuring of glassy carbon for precision glass molding of binary diffractive optical elements, Optical Materials Express, Vol. 6, Issue 11, 2016, pp. 3407-3416. [73]. W. D. Liu, M. Liu, L. C. Zhang, Oxidation-induced mechanical deterioration and hierarchical cracks in glassy carbon, Carbon, Vol. 100, 2016, pp. 178-186. [74]. J. Masuda, J. W. Yan, T. F. Zhou, T. Kuriyagawa, Y. Fukase, Thermally induced atomic diffusion at the interface between release agent coating and mould substrate in a glass moulding press, Journal of Physics D-Applied Physics, Vol. 44, Issue 21, 2011. [75]. X. Y. Zhu, J. J. Wei, L. X. Chen, J. L. Liu, L. F. Hei, C. M. Li, et al., Anti-sticking Re-Ir coating for glass molding process, Thin Solid Films, Vol. 584, 2015, pp. 305-309. [76]. J. Hagen, F. Burmeister, A. Fromm, P. Manns, G. Kleer, Iridium coatings with titanium sublayer deposited by RF magnetron sputtering: Mechanical properties and contact behavior with RoHS-compliant glass melt, Plasma Processes and Polymers, Vol. 6, 2009, pp. S678-S683. [77]. K. D. Fischbach, K. Georgiadis, F. Wang, O. Dambon, F. Klocke, Y. Chen, et al., Investigation of the effects of process parameters on the glass-to-mold sticking force during precision glass molding, Surface & Coatings Technology, Vol. 205, Issue 2, 2010, pp. 312-319. [78]. Y. Y. Chang, C. J. Wu, Mechanical properties and impact resistance of multilayered TiAlN/ZrN coatings, Surface & Coatings Technology, Vol. 231, 2013, pp. 62-66. [79]. J. Vetter, 60 years of DLC coatings: Historical highlights and technical review of cathodic arc processes to synthesize various DLC types, and their evolution for industrial applications, Surface & Coatings Technology, Vol. 257, 2014, pp. 213-240. [80]. F. Bernhardt, K. Georgiadis, L. Dolle, O. Dambon, F. Klocke, Development of a ta-C diamond-like carbon (DLC) coating by magnetron sputtering for use in precision glass molding, Materialwissenschaft und Werkstofftechnik, Vol. 44, Issue 8, 2013, pp. 661-666. 462 Chapter 15. Precision Glass Molding [81]. F. Klocke, K. Georgiadis, O. Dambon, K. D. Bouzakis, S. Gerardis, G. Skordaris, A complete qualification methodology for coatings of precision glass molding tools, Proceedings of SPIE, Vol. 8126, 2011, 81260T. [82]. F. Bernhardt, K. Georgiadis, O. Dambon, F. Klocke, Novel testing facility for investigating wear on PGM sample tools, Proceedings of SPIE, Vol. 8884, 2013, 88841V. [83]. I. Zarudi, L. C. Zhang, On the limit of surface integrity of alumina by ductile-mode grinding, Journal of Engineering Materials and Technology, Vol. 122, Issue 1, 2000, pp. 129-134. [84]. T. S. Liu, B. A. Latella, L. C. Zhang, Grinding of ceramics: Strength, surface features and grinding conditions, Abrasive Technology: Current Development and Applications I, 1999, pp. 3-10. [85]. T. S. Liu, B. A. Latella, L. C. Zhang, Grinding of ceramics: Strength, surface features and grinding conditions, Precision Machining of Advanced Materials, Vol. 196, 2001, pp. 53-60. [86]. E. Brinksmeier, Y. Mutlugunes, F. Klocke, J. C. Aurich, P. Shore, H. Ohmori, Ultra-precision grinding, CIRP Annals, Vol. 59, Issue 2, 2010, pp. 652-671. [87]. Y. E. Tohme, Grinding aspheric and freeform micro-optical molds, Proceedings of SPIE, Vol. 6462, 2007, 64620K. [88]. D. Golini, W. I. Kordonski, P. Dumas, S. Hogan, Magnetorheological finishing (MRF) in commercial precision optics manufacturing, Proceedings of SPIE, Vol. 3782, 1999, pp. 80-91. [89]. I. V. Prokhorov, W. I. Kordonsky, L. K. Gleb, G. R. Gorodkin, M. L. Levin, New highprecision magnetorheological instrument based method of polishing optics, OSA OF & T Workshop Digest, Vol. 24, 1992, pp. 134-136. [90]. A. T. H. Beaucamp, R. R. Freeman, A. Matsumoto, Y. Namba, Fluid jet and bonnet polishing of optical moulds for application from visible to X-ray, Proceedings of SPIE, Vol. 8126, 2011, 81260U. [91]. J. Guo, Corrective finishing of a micro-aspheric mold made of tungsten carbide to 50 nm accuracy, Applied Optics, Vol. 54, Issue 18, 2015, pp. 5764-5770. [92]. S. Goel, The current understanding on the diamond machining of silicon carbide, Journal of Physics D-Applied Physics, Vol. 47, Issue 24, 2014, 243001. [93]. J. S. Taylor, C. K. Syn, T. T. Saito, R. R. Donaldson, Surface finish measurements of diamond-turned electroless-nickel-plated mirrors, Optical Engineering, Vol. 25, Issue 9, 1986, pp. 1013-1020. [94]. A. Pramanik, K. S. Neo, A. Rahman, X. P. Li, M. Sawa, Y. Maeda, Cutting performance of diamond tools during ultra-precision turning of electroless-nickel plated die materials, Journal of Materials Processing Technology, Vol. 140, 2003, pp. 308-313. [95]. A. Barbacki, M. Kawalec, A. Hamrol, Turning and grinding as a source of microstructural changes in the surface layer of hardened steel, Journal of Materials Processing Technology, Vol. 133, Issue 1-2, 2003, pp. 21-25. [96]. M. S. Huang, Y. C. Chiang, S. C. Lin, H. C. Cheng, C. F. Huang, Y. K. Shen, et al., Fabrication of microfluidic chip using micro-hot embossing with micro electrical discharge machining mold, Polymers for Advanced Technologies, Vol. 23, Issue 1, 2012, pp. 57-64. [97]. A. R. Agrawal, I. O. Pandelidis, M. Pecht, Injection-molding process-control – A review, Polymer Engineering and Science, Vol. 27, Issue 18, 1987, pp. 1345-1357. [98]. B. Gleason, P. Wachtel, J. D. Musgraves, Using design of experiments to improve precision glass moulding, International Journal of Experimental Design and Process Optimisation, Vol. 3, Issue 3, 2013, pp. 263-275. [99]. P. Mahajan, T. P. Dora, T. S. Sandeep, V. M. Trigune, Optimized design of optical surface of the mold in precision glass molding using the deviation approach, International Journal for Computational Methods in Engineering Science and Mechanics, Vol. 16, 2015, pp. 53-64. 463 Advances in Optics: Reviews. Book Series, Vol. 3 [100]. L. J. Su, Y. Chen, A. Y. Yi, F. Klocke, G. Pongs, Refractive index variation in compression molding of precision glass optical components, Applied Optics, Vol. 47, Issue 10, 2008, pp. 1662-1667. [101]. L. J. Su, F. Wang, P. He, O. Dambon, F. Klocke, A. Y. Yi, An integrated solution for mold shape modification in precision glass molding to compensate refractive index change and geometric deviation, Optics and Lasers in Engineering, Vol. 53, 2014, pp. 98-103. [102]. S. F. Tseng, C. T. Lee, K. C. Huang, D. Y. Chiang, C. Y. Huang, C. P. Chou, Mechanical Properties of Pt-Ir and Ni-Ir Binary Alloys for Glass-Molding Dies Coating, Journal of Nanoscience and Nanotechnology, Vol. 11, Issue 10, 2011, pp. 8682-8688. [103]. M. Arai, Y. Kato, T. Kodera, Characterization of the thermo-viscoelastic property of glass and numerical simulation of the press molding of glass lens, Journal of Thermal Stresses, Vol. 32, Issue 12, 2009, pp. 1235-1255. [104]. A. Jain, A. Y. Yi, Numerical modeling of viscoelastic stress relaxation during glass lens forming process, Journal of the American Ceramic Society, Vol. 88, Issue 3, 2005, pp. 530-535. [105]. B. Ananthasayanam, P. F. Joseph, D. Joshi, S. Gaylord, L. Petit, V. Y. Blouin, et al., Final shape of precision molded optics: Part I – Computational approach, material definitions and the effect of lens shape, Journal of Thermal Stresses, Vol. 35, Issue 6, 2012, pp. 550-578. [106]. B. Ananthasayanam, P. F. Joseph, D. Joshi, S. Gaylord, L. Petit, V. Y. Blouin, et al., Final shape of precision molded optics: Part II – Validation and sensitivity to material properties and process parameters, Journal of Thermal Stresses, Vol. 35, Issue 7, 2012, pp. 614-636. [107]. Y. C. Tsai, C. H. Hung, J. C. Hung, Glass material model for the forming stage of the glass molding process, Journal of Materials Processing Technology, Vol. 201, Issue 1-3, 2008, pp. 751-754. [108]. F. Richter, Upsetting and viscoelasticity of vitreous SiO2: Experiments, interpretation and simulation, PhD Thesis, Technische Universität Berlin, 2006. [109]. W. D. Liu, H. H. Ruan, L. C. Zhang, Revealing structural relaxation of optical glass through the temperature dependence of young's modulus, Journal of the American Ceramic Society, Vol. 97, Issue 11, 2014, pp. 3475-3482. [110]. J. C. Dyre, N. B. Olsen, T. Christensen, Local elastic expansion model for viscous-flow activation energies of glass-forming molecular liquids, Physical Review B, Vol. 53, Issue 5, 1996, pp. 2171-2174. [111]. A. Balajee, Computational modeling of precision molding of aspheric glass optics, PhD Thesis, Clemson University, 2008. [112]. A. Y. Yi, B. Tao, F. Klocke, O. Dambon, F. Wang, Residual stresses in glass after molding and its influence on optical properties, Procedia Engineering, Vol. 19, 2011, pp. 402-406. [113]. Q. D. To, Q. C. He, M. Cossavella, K. Morcant, A. Panait, J. Yvonnet, The tempering of glass and the failure of tempered glass plates with pin-loaded joints: Modelling and simulation, Materials & Design, Vol. 29, Issue 5, 2008, pp. 943-951. [114]. Y. Chen, A. Y. Yi, L. Su, F. Klocke, G. Pongs, Numerical simulation and experimental study of residual stresses in compression molding of precision glass optical components, Journal of Manufacturing Science and Engineering, Vol. 130, Issue 5, 2008, 051012. [115]. H. N. Ritland, Relation between refractive index and density of a glass at constant temperature, Journal of the American Ceramic Society, Vol. 38, Issue 2, 1955, pp. 86-88. [116]. B. Tao, L. G. Shen, A. Yi, M. J. Li, J. Zhou, Reducing refractive index variations in compression molded lenses by annealing, Optics and Photonics Journal, Vol. 3, 2013, pp. 118-121. [117]. C. Y. Poon, B. Bhushan, Comparison of surface roughness measurements by stylus profiler, AFM and non-contact optical profiler, Wear, Vol. 190, Issue 1, 1995, pp. 76-88. [118]. F. Z. Fang, X. D. Zhang, A. Weckenmann, G. X. Zhang, C. Evans, Manufacturing and measurement of freeform optics, CIRP Annals, Vol. 62, Issue 2, 2013, pp. 823-846. 464 Chapter 15. Precision Glass Molding [119]. M. Dobosz, M. Kozuchowski, Overview of the laser-wavelength measurement methods, Optics and Lasers in Engineering, Vol. 98, 2017, pp. 107-117. [120]. F. Gao, R. K. Leach, J. Petzing, J. M. Coupland, Surface measurement errors using commercial scanning white light interferometers, Measurement Science and Technology, Vol. 19, Issue 1, 2008, 015303. [121]. E. A. Patterson, Digital photoelasticity: Principles, practice and potential measurements lecture, Strain, Vol. 38, Issue 1, 2002, pp. 27-39. [122]. K. Ramesh, T. Kasimayan, B. N. Simon, Digital photoelasticity – A comprehensive review, Journal of Strain Analysis for Engineering Design, Vol. 46, Issue 4, 2011, pp. 245-266. [123]. Y. Chen, A. Y. Yi, L. Su, F. Klocke, G. Pongs, Numerical simulation and experimental study of residual stresses in compression molding of precision glass optical components, Journal of Manufacturing Science and Engineering, Vol. 130, Issue 5, 2008. [124]. T. D. Pallicity, A. T. Vu, K. Ramesh, P. Mahajan, Birefringence measurement for validation of simulation of precision glass molding process, Journal of the American Ceramic Society, 2017, pp. 1-19. [125]. K. Ashokan, K. Ramesh, An adaptive scanning scheme for effective whole field stress separation in digital photoelasticity, Optics and Laser Technology, Vol. 41, Issue 1, 2009, pp. 25-31. [126]. T. Miyashita, Standardization for microlenses and microlens arrays, Japanese Journal of Applied Physics Part 1, Vol. 46, Issue 8B, 2007, pp. 5391-5396. [127]. Y. J. Zhu, Y. Wang, J. X. Na, Y. N. Zhi, Y. F. Fan, A framework of cloud supported collaborative design in glass lens moulds based on aspheric measurement, Proceedings of SPIE, Vol. 8841, 2013, 88411A. [128]. T.-C. Wu, Y.-C. Tsai, C.-h. Hung, Die shape optimization on molding process of optical glass lens, in Proceedings of the 33rd National Conference on Mechanical Engineering of Chinese Society of Mechanical Engineers, 2007, pp. D17-0019. [129]. M. Sellier, Optimal process design in high-precision glass forming, International Journal of Forming Processes, Vol. 9, Issue 1, 2006, pp. 61-78. [130]. M. Sellier, C. Breitbach, H. Loch, N. Siedow, An iterative algorithm for optimal mould design in high-precision compression moulding, Journal of Engineering Manufacture, Vol. 221, Issue 1, 2007, pp. 25-33. [131]. W. D. Liu, L. C. Zhang, Numerical optimization platform for precision glass molding by the simplex algorithm, Applied Optics, Vol. 56, Issue 12, 2017, pp. 3245-3250. [132]. P. Mahajan, T. Dora P., T. S. Sandeep, V. M. Trigune, Optimized design of optical surface of the mold in precision glass molding using the deviation approach, International Journal for Computational Methods in Engineering Science and Mechanics, Vol. 16, 2015, pp. 53-64. [133]. M. Sellier, Optimization in high-precision glass forming, in Progress in Industrial Mathematics at ECMI 2004 (A. Di Bucchianico, R. Mattheij, M. Peletier, Eds.), Vol. 8, Springer, Berlin, Heidelberg, 2006, pp. 381-384. [134]. K. Deb, A. Pratap, S. Agarwal, T. Meyarivan, A fast and elitist multiobjective genetic algorithm: NSGA-II, IEEE Transactions on Evolutionary Computation, Vol. 6, Issue 2, 2002, pp. 182-197. [135]. M. S. Tian, X. Y. Gong, L. Yin, H. Z. Li, W. Y. Ming, Z. Zhang, et al., Multi-objective optimization of injection molding process parameters in two stages for multiple quality characteristics and energy efficiency using Taguchi method and NSGA-II, International Journal of Advanced Manufacturing Technology, Vol. 89, Issue 1-4, 2017, pp. 241-254. [136]. Y. M. Rong, Q. Zhou, Y. Huang, Y. Chang, G. J. Zhang, X. Y. Shao, Multi-objective optimization of laser brazing with the crimping joint using ANN and NSGA-II, 465 Advances in Optics: Reviews. Book Series, Vol. 3 International Journal of Advanced Manufacturing Technology, Vol. 85, Issue 5-8, 2016, pp. 1239-1247. [137]. K. Bouacha, A. Terrab, Hard turning behavior improvement using NSGA-II and PSO-NN hybrid model, International Journal of Advanced Manufacturing Technology, Vol. 86, Issue 9-12, 2016, pp. 3527-3546. [138]. G. J. Zhang, Z. Zhang, W. Y. Ming, J. W. Guo, Y. Huang, X. Y. Shao, The multi-objective optimization of medium-speed WEDM process parameters for machining SKD11 steel by the hybrid method of RSM and NSGA-II, International Journal of Advanced Manufacturing Technology, Vol. 70, Issue 9-12, 2014, pp. 2097-2109. 466 Chapter 16. Deterministic Loose Abrasive Wear in Conventional Grinding-Polishing Machines Chapter 16 Deterministic Loose Abrasive Wear in Conventional Grinding-Polishing Machines Luis C. Alvarez-Nuñez, Carolina Keiman and Oscar Chapa1 16.1. Introduction The processes of automation of manufacturing optical components are continually in development. Optical industry requires more precise components and low costs. Optical manufacturing automation is currently undergoing dynamic developments, and the search for higher precision at lower costs leads the market. However not all the optical components demand modern equipment for their manufacturing, most of the lens and mirrors continue and they will continue being spherical. The spatial frequency error is most difficult to control in CNC polishing machines due to small polishing pad given a mid-frequency errors. This means that conventional optical manufacturing machines can still be used in order to control low-medium and high spatial frequency errors [1]. The advantage of traditional grinding-polishing machines over numeric control machines is the facility and the capacity of mass production and most important the ability to control their figure in optical components adjusting strokes parameters [2-4]. This means that the optical manufacturing using traditional machines will be been able to continue using [2, 4], if we increase their efficiency and effectiveness. Traditionally, optical manufacturing processes are classified into three stages: generation, grinding and polishing. Luis C. Alvarez-Nuñez Universidad Nacional Autónoma de México, Instituto de Astronomía, Mexico City, Mexico 467 Advances in Optics: Reviews. Book Series, Vol. 3 Undoubtedly, grinding and polishing are of great importance, since it influences directly upon an optical element’s precision, through curvature radius, figure and roughness control. To minimize grinding and polishing time the technician should adjust the machines parameters like velocity, pressure defined in Preston’s theory [5] and most important adjust stroke parameter like amplitude and off-center machine [2-4] in order to control their figure (topography), thickness (waste material) and roughness (quality polishing) in optical components. To improve the processes of production of optical components is necessary to understand the mechanism of loose abrasive wear in traditional machines. The present work is a contribution in this sense, based in a mathematical model that optimizes the machine adjustment parameter. 16.2. Abrasive Wear Theory Abrasive wear is a stochastic-statistic process used in optical manufacturing process based in mechanical and chemical interaction between abrasive-glass. The Preston’s theory [5] predicts that the abrasive wear h in a point (x, y) is proportional to the pressure P (pressure applied on upper element) and the magnitude of the relative speed V (relative instantaneous speed among two interacting points) of the optical components concerning to the tool (Eq. 16.1). This representation, the first and simpler, uses an empirical constant proportionality KP that depends mainly of the physical properties of the materials that interact (as type of glass, tool and abrasive grain) and includes the effects without distinguishing the influence of each one of them, like the size of abrasive grain, workspace, the hardness of the tool, glass, and the abrasive, etc. t h  K P  P( x, y, t ) V ( x, y, t ) dt . (16.1) 0 Subsequent research, especially those headed by Kumanin [6], introduced a deterministic equation using physical parameters for KP representing loose abrasive wear. (Eq. 16.2) h  1 . 5 E  6 kQ  S t  P( x, y, t ) V ( x, y, t ) dt , (16.2) 0 where  is the glass relative wear rate (Unity = BK7), k is the tool hardness (сast iron = 1.2), Q is the abrasive particle diameter, S is the Area under abrasive wear action. In both equations P and V are functions of position (x, y) and time t. Besides P and V, the grinding wear and polishing with free abrasive, is proportional to the viscosity and pH of the liquid in that they are suspended, the abrasive particles, the 468 Chapter 16. Deterministic Loose Abrasive Wear in Conventional Grinding-Polishing Machines concentration of this suspension and the physical properties of the individual materials that participate in the abrasive process. W. Rupp [7, 8] investigated the effect of local pressure distribution. He demonstrated that the source of curvature change is non-uniform pressure distribution near lower disk’s edge. Along this line, Wagner and Shannon [9] suggested that the original wear model proposed by Preston and Kumanin [5, 6], should be improved to include non-uniform pressure distribution at tool edge. Following this reasoning, Cordero-Davila et al [10] introduced a new model to evaluate wear rate increase due to such edge pressure distribution. This model is based on the conditions of Force and Momentum equilibrium, that interacting disks must fulfill at all time. Following this research [10] Alvarez et al [4] introducing slight modifications to Cordero-Davila’s results, dictated by the differences between their theoretical model and world machine. The research leaded by Alvarez et al [2-4] confirm that the variable affecting the wear process is the relative velocity V(x, y, t) of displacement between point pairs (on the tool and on the work disk holding the optical components, interacting through slurry abrasive grains), within the instantaneous contact area among both disks. The relative velocity was calculated between disks and was estimated the wear in upper disk assumed that the tool (lower disk) is practically undeformable and unwearable, at least as compared to the material being worked (glass). It is also assumed that the tool’s surface is uniform. In real life, tool and work can be upper and/or lower disk depending of curvature glass and figure correction based in stroke adjustment. The purpose of this work is to complete the general relative velocity equations for upper and lower disk and based in Preston’s and Kumanin’s equations estimate the wear and figure in upper disk and lower disk for grinding and polishing stage, we confirm these results comparing the final figure with interferometric data. In Section 16.5 we present a brief summary of equations based in upper disk relative velocity condition [2], in Section 16.6 we present the relative velocity equations in lower disk condition, in Section 16.7 we present a boundary conditions based in non-contact pair of points in upper and lower disk, in Section 16.8 we presents a pressure model based in Cordero's ring, and finally in Section 16.9 we present two simulations for lower and upper disk and estimate the final topography (figure) based in tool-work configurations (lenspolishing tool) and compared with real data from interferometer using CaF2 and S-FTM16 glass for FRIDA astronomical instrument [11]. 16.3. Conventional Grinding-Polishing Machines In conventional grinding-polishing machines the lower disk is motorized (spindle), rotating at a constant angular velocity . Upper free disk undergoes a reciprocating movement sliding over the lower disk, led by an arm (length L0) which is driven by an eccentric pivot that rotates at  cycles per unit time. (Fig. 16.2). 469 Advances in Optics: Reviews. Book Series, Vol. 3 Upper disk rotates freely about its central pivot, at a variable angular velocity (t), driven by the resultant of the frictional forces arising from the abrasive process that takes place within both disks contact area. Almost all machine variables are under the technician’s direct control and only one of the relative velocity components (V2) arises as consequence of the interaction of the three factors of the abrasive wear process [2, 3]. The other two velocity components (V0 and V1) are under the machine operator’s total control, as will be described in detail below. Strassbaugh’s grinding-polishing machine, model 6UR-1 and modern 6DE-DC-2 machine employed in all experiments reported herein (Fig. 16.1) possess this configuration. Fig. 16.1. Strassbaugh Machines. Left. Model 6UR-1. Right. Model 6DE-DC-2. On this machine, the operator can adjust in continuous fashion three experimental variables: pressure P applied on the upper disk through the leading arm, spindle’s rotational speed  and the arm’s oscillating rate . Oscillating arm’s central position A0 (with respect to the spindle’s axis) and oscillating amplitude +A, can also be varied continuously within some ranges, but to do so the operator must stop the machine completely. The axis of the universal cylindrical polar coordinate system is defined by the straight line joining both machine fixed axes, the spindle (a) and the arm’s oscillation axis (d), see Fig. 16.2. Its origin is located at the oscillating axis (d). In this analysis, each disk possesses its own cylindrical polar coordinate system. Their origins are located at each disk’s rotational axis, see Fig. 16.2. 470 Chapter 16. Deterministic Loose Abrasive Wear in Conventional Grinding-Polishing Machines Fig. 16.2. (a) Grinding machine, upper view; (b) Simplified diagram, showing parameters. 16.4. Relative Velocities between an Arbitrary Pair of Points Velocity vector contributions arising from a conventional grinding machine’s operation are: V0, due to the arm’s oscillation leading upper disk, V1, due to the lower disk’s rotation  driven by machine’s spindle, and V2, due to upper disk’s rotation (t) driven by abrasive wear frictional forces (Fig. 16.2a). To keep track of the process at every point, it is necessary to know both disks’ coordinate axes orientations with respect to the machine’s universal axis. Let an arbitrary upper disk point coordinates be (2, ). Let W* be the integer of (t) (upper disk rotation) from an initial instant t0=0 to t, under the sole action of abrasive process frictional forces given by Eq. (16.3), similar manner let an arbitrary lower points coordinates be (1, ) and M* be integer of  (lower disk rotation). t W *    (t ) dt . (16.3) 0 During the same time interval, the lower disk rotated a total angle of M* =  t, driven by the machine’s spindle. These two angles (W*, M*), after subtracting complete rotations (modulo 2) are given in universal polar coordinates by Eqs. (16.4) and (16.5). W   2    (t ) dt  Int  0 t   (t) dt  , t 0 M   2  t  Int   t   . (16.4) (16.5) 471 Advances in Optics: Reviews. Book Series, Vol. 3 16.5. Upper Disk Relative Velocity 16.5.1. First Relative Velocity Contribution (V0) Approximate Calculation This contribution arises exclusively from upper disk translation across the lower disk’s surface (both disks assumed static) driven by the oscillating arm by means of a bar engaged to upper disk’s central pivot. At any small time interval t, it can be assumed that the upper disk undergoes a simple arc circle displacement. The driving arm’s oscillation can be approximated as Eq. (16.6), (Fig. 16.2a) and exact movement is given by 4 arms configuration [2]. A A  A 0  A sin ( 2  γ t ) , (16.6) where  is the eccentric rotational speed (RPM) driving the arm, A0 is the arm’s mean angular position (central), and A is the arm’s oscillation amplitude (Fig. 16.2a). It is assumed that during a small time interval t, at an arbitrary instant t, only the oscillating arm actuates ( = 0, (t) = 0), and that at that instant, the arm’s approximate angular position is given by Eq. (16.7), where T0 = 1/ is the eccentric period.  t  t    . A A  A 0  A sin 2    int   T T  0   0 (16.7) Time varying arm angular position AT can be represented either by Eq. (16.7) (AT = AA, Approximate value) or AT = AE = 7. Exact value, see [2]. At time t, the universal polar coordinates of the interacting pair of points combine to form their position vector: 0 = (0, 0) (Fig. 16.2a and Fig. 16.3a). Fig. 16.3. V0 computation. (a) Radial and auxiliary components; (b) angles. 472 Chapter 16. Deterministic Loose Abrasive Wear in Conventional Grinding-Polishing Machines The point pair radial coordinate will be given by Eq. (16.8),   0  L 0 2  2 L 0  2 cos(   W *  AT )   2 2  1/ 2 . (16.8) The angular coordinate 0 of the pair of points is given by Eq. (16.9), see Fig. 16.3b and [2],  ^  R0    0  AT    AT  arctan .  L0  R0    (16.9) Position vector 0 for the point pair always remains on the x, y plane Eq. (16.10) ^ . ^ ^ ^ 0  i 0x  j 0y  0  i sin 0  j cos 0  .   (16.10) The first relative velocity contribution V0 for the pair of points is given by the vector resulting of the arm’s instantaneous angular velocity  times the position vector 0 [2]. The relative velocity first contribution V0 (Fig. 16.3) is given by Eq. (16.11).  ^  i V0    ρ0   0   0 x  ^ j 0 0 y ^  k ^   α     0   ˆi cos  0  j sin  0  .    0  (16.11) 16.5.2. Second Relative Velocity Contribution (V1) Second relative velocity contribution originates exclusively from the lower disk rotation, as driven by the machine’s spindle at a constant rotational speed . For this calculation, are assumed static both arm and upper disk, during a small time interval t ((t) = 0 and  = 0). The only requirement to compute this velocity contribution is the radial distance between the interacting points and the spindle’s axis, i.e. their position vector in lower disk’s polar coordinate axis. The line 1 joining the interacting points and lower disk rotational axis, is given by Eq. (16.12)  2 ρ1  L1  2 L1 ρ2 cos(  W *  B)  ρ22  1/ 2 . (16.12) The previous results allow us to obtain the angular coordinate of a contacting pair of points in a universal polar coordinate system as 1 = B +  (Fig. 16.4b) is given by Eq. (16.13) 473 Advances in Optics: Reviews. Book Series, Vol. 3  ^  R1  1  B  arctan .  L1  R1    (16.13) Fig. 16.4. V1. Computation. (a) Radial and auxiliary line segments; (b) angles. Since point pair always remain on the x, y plane, its position vector is given by Eq. (16.14) ^ ^ ^ . ^  1  i  1 x  j  1 y   1 ( i sin  1  j cos  1 ) . (16.14) On the other hand, lower disk angular velocity vector remains constant, perpendicular to lower disk’s face and positive, it can be written:  = kM* ( k = 1 ). This means that the second relative velocity contribution vector is given by Eq. (16.15)  ^  i V1  Ω  ρ1   0    1x  ^ j 0  1y ^  k ^   ^      1   i cos  1  j sin  1  .    0  (16.15) Note that expression Eq. (16.15) correspond to lower disk point velocity, moving with respect to an upper disk’s point that remains instantaneously static. To compute the wear produced on the upper disk’s surface, this velocity contribution Eq. (16.15) must be taken as negative. This fact will be applied latter to the vector sum of all three relative velocity contributions. 16.5.3. Third Relative Velocity Contribution (V2) The third relative velocity vector contribution originates from the rotation of the upper disk, which spins under the action of frictional forces arising from the loose abrasive wear process taking place between disks [2, 3]. 474 Chapter 16. Deterministic Loose Abrasive Wear in Conventional Grinding-Polishing Machines Let us assume now that during a small time interval t this velocity is the only one acting upon the pair of points ( = 0,  = 0). The upper disk’s point polar coordinates are given by Eqs. (16.16) and (16.17), see Fig. 16.5. 2  W *   , ^ (16.16) . ^ 2  i 2 x  j 2 y ^ ^    2  i sin  2  j cos  2  .   (16.17) Fig. 16.5. V2. Computation. (a) Radial and auxiliary line segments, (b) angles. Due to machine setup, this velocity vector expression is given by Eq. (16.18)  ^  i ^ V 2  ω(t) k  ρ2   0   2 x  ^ j 0 2 y ^  k  ^ ^  (t)    (t )  2   i cos  2  j sin  2  .    0   (16.18) The only remaining unknown in Eq. (16.18) is (t). A number of theoretical and experimental contributions [3] have pointed at this variable’s importance. However, these results are difficult to calculate in real time, or apply only to particular cases. An approximate and exact (t) equation based on experimental results was tested, described in [2, 3]; it proved to be appropriate approximation simulations of the grinding and polishing process. Note: 2 and 2 is the arbitrary point position in upper disk that is evaluated and repeated N times for simulated total wear in upper disk. 475 Advances in Optics: Reviews. Book Series, Vol. 3 16.5.4. Vector Addition of Three Relative Velocity Contributions The relative velocity between upper disk-lower disk point pair is equal to V0, V1, and V2 vector addition, see Fig. 16.9. The upper disk point (Cartesian coordinates) relative velocity components are given by Eq. (16.19) VTx  V0 x  V1x  V2 x , VTy  V0 y  V1 y  V2 y . (16.19) The minus sign affecting V1 stands for the required opposite direction for this contribution to effect on the upper disk point, as pointed out above. The relative velocity magnitude in upper disk is given by Eq. (16.20), to be applied in Preston’s or Kumanin’s wear equations, Eqs. (16.1) and (16.2). VT  (VTx )2  (VTy )2 , (16.20) 16.6. Lower Disk Relative Velocity Lower disk velocity contribution can be calculated depending of tool-work configuration and is given for 3 velocities showing next. 16.6.1. Approximate Calculation of the First Relative Velocity Component (V0) At time t, the universal polar coordinates of the interacting pair of points combine to form their position vector: 0 = (0, 0) (Fig. 16.6). Fig. 16.6. V0. Computation. (a) Radial and auxiliary components. (b) angles. The point pair radial coordinate will be given by Eq. (16.21) ρ0  476 G  Ry , cos  0 (16.21) Chapter 16. Deterministic Loose Abrasive Wear in Conventional Grinding-Polishing Machines where the upper disk’s center position radial and tangential components are given by Eqs. (16.22) and (16.23) respectively, see also Fig. 16.3a, Rx  ρ1 sin( M *  ) Ry  ρ1 cos(M *  ) . , (16.22) The angular and radial coordinate 0 and 0 of the pair of points is given by Eq. (16.23), see Fig. 16.6b,  Rx  ,  Ry  G  (16.23)  0  arctan  ^ . ^ ^ ^ 0  i 0x  j 0y  0  i sin 0  j cos 0  .   (16.24) The first relative velocity contribution V0 for the pair of points is given by the vector resulting of the arm’s instantaneous angular velocity  times the position vector 0. The relative velocity first contribution V0 (Fig. 16.6) is given by Eq. (16.25)  ^  i V0    ρ0   0   0 x  ^ j 0 0 y ^  k ^   α     0   ˆi cos  0  j sin  0  .    0  (16.25) 16.6.2. Calculation of the Relative Velocity Second Component (V1) This relative velocity contribution is produced exclusively by the lower disk rotation at constant rotational speed . For the present calculation, consider static both the arm and the upper disk rotation, during a small time interval t. Lower disk point polar coordinates are given in Eq. (16.26) (Fig. 16.7) 1  M *   . (16.26) Since the angular velocity of the lower disk remains constant, perpendicular to lower disk´s face and takes place in a positive direction, we can write:  = kM* (| k | = 1 ). On the other hand, radial coordinates always remain in the x, y plane, we have in Eq. (16.27): ^ ^ ^ . ^ 1  i 1x  j 1 y  1 ( i sin 1  j cos 1 ) . (16.27) This means that the second relative velocity contribution vector expression is given in Eq. (16.28) 477 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 16.7. Auxiliary line segments, required for computation of V1.  ^  i V1    1   0   1x  ^ j 0 1 y ^  k ^  ^      1   i cos 1  j sin 1  .    0  (16.28) Note: 1 and 1 is the arbitrary point position in upper disk that is evaluated and repeated N times for simulated total wear in lower disk. 16.6.3. General Expression for the Third Relative Velocity Component (V2) Radial distance between the pair of contacting points referenced to upper disk 2 and its angle 2 referenced to universal coordinate axis and center of upper disk is our only unknown variable to compute this velocity contribution. The third vector contribution to the relative velocity among contacting point pairs, originates from the upper disk’s rotation, which spins under the action of frictional forces arising from the loose abrasive wear process taking place between disks. Let’s assume now that during a small time interval t this velocity component is the only one acting on the point pair. Due to arm’s oscillation, the distance between both disks rotation axes L1 is given by Eq. (16.29) (Fig. 16.8) L1   rx 2  ry  2 1/2 where radial components are given in Eq. (16.30) 478 , (16.29) Chapter 16. Deterministic Loose Abrasive Wear in Conventional Grinding-Polishing Machines rx  L0 sin( AT ) , ry  L0 cos( AT )  G . (16.30) Fig. 16.8. Auxiliary line segments, required for computation of V2. Angle B (Fig. 16.8) between L1 and universal polar coordinate axis is given in Eq. (16.31) r  B  arctan  x  . r   y  (16.31) Employing the radial and angular components of lower disk center position (L1, 1) we can obtain radial component of upper disk (Fig. 16.8) and Eq. (16.32).   2  L12  12  2 L1 1 cos( 3 )  1/ 2 , (16.32) where angular position 3 is given in Eq. (16.33)  3  1  B , (16.33) 1  M *   . (16.34) and 1 is given in Eq. (16.34) Angular coordinates 2 of the pair of points is given by Eq. (16.35) 2  B  4 , (16.35) where angular 4 coordinates is given in Eq. (16.36) 479 Advances in Optics: Reviews. Book Series, Vol. 3 ^  R1  4  tan   .  R1    (16.36) 1 ^ Auxiliary components R 2 , R2 are given by Eq. (16.37) R1  1 cos( 3 )  L1 , ^ R1  1 sin( 3 ) . (16.37) On the other hand, radial coordinates Eq. (16.38) always remain in the x, y plane, we have: ^ ^ ^ . ^  2  i  2 x  j  2 y   2 ( i sin  2  j cos  2 ) . (16.38) Due to machine setup, the angular velocity vector expression is: (t) = k (t), wherefrom is shown in Eq. (16.39):  ^  i V2   (t )   2   0   2 x  ^ j 0 2 y ^  k  ^ ^  (t )    (t )  2   i cos  2  j sin  2  .    0   (16.39) Note that these calculations correspond to the point on the upper disk, moving with respect to lower disk’s point, which remains instantaneously static. To compute wear produced on lower disk’s surface, this velocity contribution must take as the negative of the expression found. We’ll apply this fact latter, to the vector sum of all three relative velocity components. 16.6.4. Vector Addition of Three Relative Velocity Contributions The relative velocity among a given pair of points in contact, is the vector addition of the three contributions found, namely V0, V1, and V2, see Fig. 16.9. The lower disk point (Cartesian coordinates) relative velocity equations, are given by Eq. (16.40) V Tx  V 0 x  V 1 x  V 2 x , V Ty  V 0 y  V 1 y  V 2 y . (16.40) The minus sign affecting V2 stands for the required opposite sense for this contribution to effect on the lower disk point, as pointed out above. The relative velocity magnitude among arbitrary pair of points in lower disk is given by Eq. (16.41), to be applied in Preston’s or Kumanin’s wear equations, Eqs. (16.1) and (16.2). 480 Chapter 16. Deterministic Loose Abrasive Wear in Conventional Grinding-Polishing Machines VT  (V Tx ) 2  (V Ty ) 2 . (16.41) Fig. 16.9. Final relative velocity vector contributions. 16.7. Boundary Conditions in Abrasive Wear Process Depending upon both disk sizes, arm’s length L0, oscillation amplitude A and central position A0, there can be moments when an upper disk fraction slides beyond the lower one’s edge. At these time intervals, an upper disk and lower disk region will not experience abrasive wear whatsoever. The upper disk no-wear condition takes place when the magnitude of the upper disk point position vector becomes larger than the lower disk radius i.e. 2 > R1 and for lower disk no-wear condition when 1 > R2 upper disk radius (Fig. 16.9). 16.8. Pressure Distribution within Disks Contact Area Instantaneous pressure among point pairs is not uniformly distributed, neither is constant, but changes continuously as a function of upper disk’s position over the lower one assuming contacting surfaces match each other. Applying the results of Cordero-Davila [10] and modify the notation for to use in our system is possible to determine the pressure in any point of contact in upper disk and lower disk [4]. 481 Advances in Optics: Reviews. Book Series, Vol. 3 16.9. Arm Stroke Adjustments (Controlling Curvature Radius and Figure) Three of the arm stroke adjustments most frequently employed in optical fabrication, to drive the curvature towards a desired value, are described here [4]. The tendency followed by the upper disk surface curvature or lower disk curvature is a direct consequence of the abrasive wear induced in each case and depends of relative velocities and pressure distribution due to machine adjustments. Labels in parenthesis (Fig. 16.10) indicate upperdisk’s surface change tendencies as grinded and polishing against lower disk. Fig. 16.10. Stroke settings and upper disk surface change tendency (in parenthesis). Lower disk’s surface changes in a complementary manner. <1/(3-4) Diameter (Convex), =1/(3-4) Diameter (Flat), >1/(3-4) (Concave). The relationship 1/(3-4) (upper disk diameter) < eccentric position > 1/(3-4) (upper disk diameter) is based in technician experience and can be modify slightly depending of curvature control and tool-work diameter configuration. The purpose of this arm stroke adjustment is for grinding process maintain and/or correct curvature radius to nominal value and reduce the internal fracture, and for polishing stage is correct the figure (maintaining the curvature radius) and reduce the roughness in glass. 16.10. Simulation and Real Optical Manufacturing In order to apply the wear equations in real world (using Kumanin’s equation), were computed the theoretical results presented in this research to the lens surface from FRIDA project [11]. One lens surface using S-FTM16 glass was polished and controlling their figure using normal stroke to achieve /4 P-V quality surface, the lens (work) was located in lower disk and tool (pitch) in upper disk position, the adjustments machines was established in normal stroke see Fig. 16.11, the results are shown in Fig. 16.11a and simulation results in Fig. 16.11b. Simulations results can be compared with figure and P-V quality, the results between real and simulated results in figure and P-V are closer (P-V 158 nm (simulation) real 0.239 waves = 151 nm (real)). Same conditions, one lens surface using CaF2 glass was polished in upper disk position and tool (pitch) was placed in lower disk position, the adjustments machines was 482 Chapter 16. Deterministic Loose Abrasive Wear in Conventional Grinding-Polishing Machines established in normal stroke see Fig. 16.12, the results are shown in Fig. 16.12a and simulation results in Fig. 16.12b. (a) Interferometric data’s surface (b) Simulations (lower disk) Fig. 16.11. Computed wear and figure over lower disk. (a) Normal Stroke A0 = 4.5˚, A = +3˚, (Sampling interval: t = 0.1 seg Load = 1 kgf,  = 33  = 20), 40 mm diameter work-tool. (b) 100×100 simulated points. (a) Interferometric data’s surface; (b) Simulations (upper disk) Fig. 16.12. Computed wear and figure over upper disk: (a) Normal Stroke A0 = 6.5, A = +4.5˚, (Sampling interval: t = 0.1 seg Load = 1 kgf  = 30  = 24), 40 mm diameter work-tool. (b) 100×100 simulated points. 483 Advances in Optics: Reviews. Book Series, Vol. 3 Results in figure and P-V are closer with real and simulation (P-V 125 nm (simulation); 0.189 waves = 119.5 nm (real)). Note the good agreement between simulated and real results for figure estimation. 16.11. Concluding Remarks The advantage of numerical simulations in optical fabrication will allow one to compute the most suitable machine settings for each desired curvature. In optical manufacturing processes, curvature control accuracy during grinding-polishing stages depends upon the type of work performed. For average and low quality large scale industrial production, quality control seeks only to maintain surface curvatures within relatively wide tolerance limits. For precision optics, constant curvature value for each surface must be controlled during the whole process, frequently adjusting machine parameters. Employing relative velocity equations given in this research, it will be feasible to simulate both trajectory and relative velocity for any upper disk point as upper disk slides over the lower disk and vice versa, lower disk point as upper disk slides. Knowing relative velocity and pressure distribution now is possible to calculate abrasive wear and figure for arbitrary lower and upper disk points as well using Kumanin’s wear equations, results can be applied with good results in both grinding and polishing stage in special this numerical simulation can be useful due to can estimate and avoid if the machine parameters are in resonance 1x Ω (spindle velocity) = 1x  (eccentric velocity), this resonances are the principal cause of errors in manufacturing due to produce surface aberration like astigmatism, ashtray, trefoil etc. For grinding stage is possible to estimate and control curvature radius and wear (waste material) and optimize the optical production process. For polishing stage we can estimate the final figure estimating the best machines parameters in consequence to save time especially using soft materials like CaF2 and improve optical quality surface. For future works we will motorize upper disk and simulate several velocities and conditions in order to control high precision optical manufacturing process and obtain a general wear equation (grinding and polishing) based in machine parameters and toolwork properties (glass-abrasive) based in theoretical and experimental results. Acknowledgment Valuable support from the administrative staff at the IA-UNAM. Special thanks to Dr. Jesus Gonzalez and Lic. Angelina Salmeron. Valuable support to FRIDA’s Engineering Team at IA-UNAM to provide optical lenses in special to Beatriz Sanchez, Salvador Cuevas and Carlos Espejo. 484 Chapter 16. Deterministic Loose Abrasive Wear in Conventional Grinding-Polishing Machines References [1]. S. Li, Y. Dai, Large and Middle-Scale Aperture Aspheric Surfaces, John Wiley & Sons, 2017. [2]. L. C. Álvarez-Nuñez, R. B. Flores-Hernández, Relative velocity and loose abrasive wear in conventional grinding machines, Optik, Vol. 120, Issue 16, 2009, pp. 845-854. [3]. L. C. Álvarez-Nuñez, R. B. Flores-Hernández, Free upper-disk rotational speed under loose abrasive grinding in conventional machines, Optik, Vol. 121, Issue 2, 2010, pp. 195-205. [4]. L. C. Álvarez-Nuñez, R. B. Flores-Hernández, Loose abrasive edge wear effects and final surface topography in conventional grinding machines, Optik, Vol. 121, Issue 3, 2010, pp. 217-229. [5]. F. W. Preston, The theory and design of plate glass finishing machines, J. Glass Tech., Vol. 11, 1927, pp. 214-256. [6]. K. G. Kumanin, Generation of Optical Surfaces, Focal Library, New York, 1962. [7]. W. Rupp, Loose abrasive grinding of optical surfaces, Appl. Opt., Vol. 11, Issue 12, 1972, pp. 2797-2810. [8]. V. Rupp, The development of optical surface during grinding process, Appl. Opt., Vol. 4, Issue 6, 1965, pp. 743-748. [9]. R. E. Wagner, R. R. Shannon, Fabrication of aspheric using a mathematical model for material removal, Appl. Opt., Vol. 13, Issue 7, 1974, pp. 1683-1689. [10]. A. Cordero-Davila, J. Gonzalez-Garcia, M. Pedrayes-Lopez, L. A. Aguilar-Chiu, J. CuautleCortes, C. Robledo-Sanchez, Edge effects with the Preston equation for a circular tool and workpiece, Appl. Opt., Vol. 43, Issue 6, 2004, pp. 1250-1254. [11]. B. Sánchez, C. Keiman, C. Espejo, S. Cuevas, L. C. Álvarez, O. Chapa, R. Flores-Meza, J. Fuentes, L. Garcés, G. Lara, J. A. López, R. Rodríguez, A. Watson, V. Bringas, A. Corrales, D. Lucero, A. Rodríguez, B. Rodríguez, D. Torres, J. Uribe, FRIDA diffraction limited NIR Instrument, the challenges of its verification processes, Proceedings of SPIE, Vol. 9150, 2014, 91501. 485 Chapter 17. Quantitative Phase Microscopy and Tomography with Spatially Incoherent Light Chapter 17 Quantitative Phase Microscopy and Tomography with Spatially Incoherent Light Azeem Ahmad and Dalip Singh Mehta1 17.1. Introduction Coherence properties of light sources play a crucial role in various optical techniques such as profilometry, digital holography (DH), quantitative phase microscopy (QPM) and optical coherence tomography (OCT) [1-5]. Coherence is broadly classified into two categories: temporal and spatial coherence [6-8]. The spatial coherence is further divided into two sub-categories: lateral and longitudinal spatial coherence [6-8]. Most of the offaxis DH and QPM techniques so far employed a highly temporally and spatially coherent (i.e., laser) light source to obtain the interference pattern easily throughout the field of view (FOV) of camera [9]. The QPM techniques provide the measurement of different parameters associated with biological objects, such as cell dynamics (i.e., thickness and refractive index fluctuations) and a cell’s dry mass density (i.e., nonaqueous content) [3]. For the quantification of these parameters off-axis digital interference microscopy is widely preferred as it can recover information related to specimen from a single interferogram [9-11]. This makes it suitable to study dynamical behaviour of biological cells or tissues, which could be a good indicator of various cell’s diseases like malaria, sickle cell anemia etc. [3]. However, high temporal and spatial coherence properties of laser light source degrade the image quality due to coherent noise and parasitic fringe formation due to multiple reflections from surfaces of the optical components [12]. As a consequence, it reduces phase and height measurement accuracy of the specimens. To improve the measurement accuracy, broadband light sources like white light (halogen lamp) and light emitting diodes (LEDs) have been used extensively in the field of QPM, DH and profilometry of biological and industrial objects [13-15]. The spatial phase sensitivity of such types of light sources comparative to lasers is very high due to their low temporal coherence length [16]. It is well known that the interference pattern occurs only when the optical path length difference between object and reference beam is within the coherence length of the light source [16]. Therefore, it is difficult to obtain interference Azeem Ahmad Department of Physics, Indian Institute of Technology Delhi, Hauz Khas, New Delhi, India 487 Advances in Optics: Reviews. Book Series, Vol. 3 pattern quickly while employing low coherent light sources (coherence length ~ 2-6 µm). Moreover, obtaining high fringe density of the interference signal and exploitation of whole camera FOV cannot be done simultaneously in case of low coherence based interferometric techniques [17]. To overcome this limitation, on-axis interferometric configurations attracted strong attention of many researchers, which can utilize whole FOV of camera at the cost of low fringe density of the modulated signal [13, 18]. Fourier transform based single-shot phase recovery of the specimens is difficult to implement with such low fringe density [19]. Therefore, on-axis interferometric configurations generally require multi-shot phase retrieval algorithms for noise (DC and twin image) free recovery of complex field related to specimens at full detector’s resolution [20]. This limits the ability to study the live cell dynamics of the biological cells and adds complexity to the system. In addition, use of the spectrally broad band light sources in DH and QPM systems require chromatic aberration corrected optical components. Further, most of the OCT systems exploit the low temporal coherence properties of light source to perform non-contact, non-invasive optical sectioning of biological cells or tissues [4, 5]. According to Wiener-Khinchin Theorem, temporal coherence function and source spectrum form Fourier transform pairs [6, 7]. In other words, larger the bandwidth of source temporal frequency spectrum, smaller will be the coherence length of light source or vice versa. Therefore, broadband temporal frequency spectrum light sources such as super-luminescent diodes (SLD), tungsten halogen lamps, and broadband Ti: Sapphire lasers are used for obtaining high-axial resolution sectioning in OCT imaging [4, 5]. However, the main disadvantages while using these broad band light sources in OCT is the requirement of dispersion-compensation mechanism for dispersion correction, and inhomogeneous spectral response of highly absorbing specimen or medium [4, 8, 21]. These limitations compel us to devise a straightforward and cost effective method to overcome aforementioned issues related to various OCT and QPM techniques. The use of a spectrally narrow, i.e., monochromatic (temporally highly coherent) and spatially extended, i.e., spatially incoherent (pseudo-thermal) light source may have advantages over all commercially available light sources [8, 22]. To date, there have been made several attempts to use spatial coherence properties of light sources in the field of profilometry and OCT [1, 8, 22-24]. Safrani and Abdulhalim used broad band thermal light source in conjunction with a narrow bandpass filter (spectral bandwidth ~10 nm) but it is still a large bandwidth compared to laser [25]. Takeda and Rosen have demonstrated surface profilometry from a synthesized pseudo-thermal source generated from a direct laser using spatial light modulator [1]. These types of light sources do not require any dispersion compensation mechanism for dispersion corrections while imaging biological specimens, which have strong dispersion or inhomogeneous spectral response [1]. The light sources having high temporal coherence (spectrally narrow) and low spatial coherence (wide angular spectrum) have been synthesized by many authors just by passing the laser light through the rotating diffuser [1, 2, 26] or stationary diffuser, followed by vibrating a multiple multimode fiber bundle (MMFB) [12, 23, 27]. The spatially low-coherent light sources generated by employing aforementioned procedures or illuminating a speckle field [28] can also be helpful for the reduction of speckle noise 488 Chapter 17. Quantitative Phase Microscopy and Tomography with Spatially Incoherent Light significantly as in case of broadband light sources. A significant number of works have been reported previously for investigating coherence properties of such light sources [26, 29-33]. This chapter, first, provides a brief introduction about different types of coherence and their use in different optical systems (Section 17.2). Here, we will focus on the mathematical formulation of temporal and spatial coherence properties of light sources. In Section 17.3 a highly efficient method to synthesize a low spatial and high temporal coherent (pseudo thermal) light source is presented. In Section 17.4 two different phase retrieval algorithms such as five step and Fourier transform methods are described. In Section 17.5 characterization of system parameters like spatial phase sensitivity and transverse as well as axial resolution of QPM and OCT techniques is studied. In Section 17.6 influence of coherence on the spatial phase sensitivity of QPM is presented. Finally in Sections 17.7 and 17.8 successful implementation of pseudo thermal light source to perform quantitative phase imaging (QPI) and OCT of various industrial and biological cells is demonstrated. 17.2. Concepts of Coherence The coherence properties of the light sources have a significant role in various optical techniques. In the coherence theory of optical fields, Wiener–Khintchin theorem is used for the determination of temporal coherence (TC) function [6, 7]. For determination lateral spatial coherence (SC) function the van-Cittert–Zernike theorem is used [34]. The longitudinal spatial coherence (LSC), which is different from the lateral SC, can be determined from the more generalized form of van-Cittert–Zernike theorem [8, 30]. The distinction among all types of coherence and their role in the microscopic systems are briefly discussed below. 17.2.1. Temporal Coherence Temporal coherence describes fixed or constant phase relationship, i.e., correlation between light vibrations at two different moments of time. According to WienerKhintchin theorem, autocorrelation or temporal coherence function Γ ∆ ∗ 〈 ∆ 〉 and source power spectral density forms Fourier transform pairs and given by the following relation [7, 8]. Γ ∆ exp 2 ∆ , (17.1) is the source spectral distributaion where Γ ∆ is the temporal coherence function, and ∗ ∆ . function, and ∆ is the temporal delay between optical fields The full width half maximum (FWHM) of temporal coherence function provides information about the coherence length. It can be seen from Eq. (17.1), larger spectral bandwidth of the light source leads to smaller coherence length or vice versa. Michelson interferometer with a collimated light beam is generally used to realize temporal 489 Advances in Optics: Reviews. Book Series, Vol. 3 coherence function experimentally as shown in Fig. 17.1(a) [35]. The light beams reflected from mirrors M1 and M2 generate interference pattern in the camera FOV, only if the optical path difference (OPD) between them is within the temporal coherence length of light source. The scanning of either mirror M1 or M2 provides temporal coherence function related to the light source. If one replaces say mirror M2 with a multilayered biological specimen then only those layers of the sample produce interference signal which come under the temporal coherence gate of the light source. Therefore, conventional OCT systems utilize low temporal coherent light sources to perform high axial resolution optical sectioning of biological cells or tissues [4]. In addition, the use of low coherent light source in DH and QPM reduces the problems of coherent noise, speckle noise, and parasitic fringe formation from the captured interferometric images [13, 16]. Michelson Interferometer Young’s Interferometer Screen CCD L BS Scan Extended source S1 M1 Point source S2 M2 (a) (b) Fig. 17.1. Schematic diagrams of Michelson and Young interferometer. L:Lens, BS: Beam splitter, : Path lengths, CCD: Charge coupled device, is spatial extent M1 – 2: Mirrors, of the light source, and S1 – 2: double slits. 17.2.2. Spatial Coherence Spatial coherence describes the correlation of optical fields at two different spatial locations, situated either transverse or longitudinal direction of the beam propagation, at the same moment of time [7, 8]. It can be seen from the previous section, the spectral bandwidth of the light source controls the temporal coherence function. In contrast, spatial coherence function is decided by angular frequency spectrum, i.e., number of spatial frequency contained in the light source [7, 8].In other words, source size controls the spatial coherence function. Fig. 17.2 is the manifestation to determine tranverse and longitudinal spatial coherence function of an extended source. 490 Chapter 17. Quantitative Phase Microscopy and Tomography with Spatially Incoherent Light (a) (b) Fig. 17.2. (a) Manifestation to determine tranverse and longitudinal spatial coherence function of an extended source. , : is an extended light source; (b) Spatial periods and spatial frequencies of a plane wave propagating along the direction [7]. 17.2.2.1. Transverse Spatial Coherence For transverse (lateral) spatial coherence, the correlation function Γ , , ∆ 0 ∗ 〈 , , 〉 between the light fieldsoriginated from source , at two different spatial points , and , located in , plane (Fig. 17.2) is considered [7, 8]. Van-Cittert-Zernike theorem relates the correlation or transverse spatial 0 ’ to the spatial frequency spectrum of the source as coherence function ‘Γ , , ∆ follows [7, 8]: Γ , ,∆ 0 , exp , (17.2) , is the spatial frequency spectrum of the light source, and are the where position vectors of points and located in , plane, and are the spatial frequencies of the light field along and directions, respectively. The spatial frequencies and for a plane wave propagating along the direction cos , cos can be written as (Fig. 17.2(b)) [7] cos cos cos cos , , (17.3) (17.4) where , are the spatial periods of the propagating field along and directions, , are the angles between the direction of propagating field and and axes respectively, and is the wavelength of optical field.The transverse spatial coherence lengths can be defined as follows: 491 Advances in Optics: Reviews. Book Series, Vol. 3 where ∆ and ∆ ∆ , (17.5) , ∆ are the transverse spatial frequency range. Finally, the expressions for tranverse (lateral) spatial coherence lengths of the optical field along and directions, in terms of and , can be defined as follows: , (17.6) . The details about the mathematical description to achieve the above relationships can be found elsewhere [7]. Young’s interferometer is the most commonly used experimental technique to determine the transverse spatial coherence function associated with the light source [35]. The experimental scheme of Young’s interferometer is illustrated in Fig. 17.1(b). 17.2.2.2. Longitudinal Spatial Coherence The generalized Van-Cittert–Zernike theorem [8], relates LSC to the spatial structure (i.e., angular frequency spectrum) of the quasi monochromatic extended light source, analogous to the Wiener-Khintchine theorem [6, 35], which states that temporal coherence function and the source spectrum form Fourier transform pairs. According to the generalized Van-Cittert–Zernike theorem [7, 8, 30], correlation or LSC function 0 ’ is defined as follows: ‘Γ δ , ∆ Γ δ ,∆ 0 exp , (17.7) 0 is the longitudinal spatial coherence function and is the where Γ δ , ∆ angular frequency spectrum of the light source, ( is the separation between spatial points and situated in two different observation planes, and is the longitudinal spatial frequency [7]. cos cos , (17.8) is the spatial period along the z-direction. is the angle between the direction where of propagating field and axis and is the wavelength of light source. The longitudinal coherence length is defined as: where ∆ ∆ , (17.9) is the longitudinal spatial frequency range. The general expression of the longitudinal coherence length ( ) which depends on both the angular frequency and temporal frequency spectrum of the light source, as follows [7]: 492 Chapter 17. Quantitative Phase Microscopy and Tomography with Spatially Incoherent Light ∆ 2 , (17.10) where is a half of the angular spectrum width, is the central wavelength, and ∆ is related to the temporal spectrum width of the source. If the size of the light source is small (i.e. a point source) then the coherence length or the coherence time or simply the longitudinal coherence properties of the light vibrations can be purely determined by the temporal frequency spectrum (spectral distribution) of the light source [7]. On the other hand, if the size of the light source is large (extended) and then the temporal frequency spectrum is narrow (Quasi-monochromatic, ∆ ≪ longitudinal spatial coherence properties of the light source can be purely determined by the angular frequency spectrum of the light source [7, 29, 33]. Thus, for sufficiently narrow temporal frequency spectrum width, second term in the above expression can be neglected and is determined by the angular frequency spectrum only: 2 2 . (17.11) In this case, the axial resolution (i.e., 2) is dominated by LSC rather than temporal coherence length of the source. In this chapter, this phenomenon is experimentally demonstrated and applied for QPI and high-resolution spatial coherence gated optical sectioning of biological cells. 17.3. Synthesis of Low Spatial and High Temporal Coherent light Source To synthesize a low spatial and high temporal coherent (pseudo thermal) light source, an effective speckle reduction scheme, i.e., combined effect of spatial, angular and temporal diversity is adopted [23, 36]. First, a green laser (temporal coherence length ~10 cm) was made incident onto the beam splitter BS1, which splits the beam into two beams as shown in Fig. 17.3. One of the beams is passed through a microscope objective MO1 to get a diverging beam and the other one is coupled into a 50/50 fiber based beam splitter using microscope objective MO2. The three diverging beams thus achieved are projected onto a common area (spot size ~6 mm) of the stationary diffuser. All the beams are made incident on the diffuser at angles – 40°, 0°, and 40°, called as angular diversity. The diffuser scattered all three beams and generates uncorrelated speckle patterns, which are coupled into MMFB (hundreds of fibers, each fiber having core diameter 0.1 mm), by a condenser lens (focal length ~17.5 mm). The direct laser beam and the two other beams coming from the fiber coupler are not travelled equal optical path length; therefore, uncorrelated speckle patterns will be present at the output of the diffuser. At least two of them add on the intensity basis and subsequently reduced speckle contrast is observed [36]. In other words, the output of the diffuser contained wide range of spatial frequency components, which leads to the short LSC length. The source 493 Advances in Optics: Reviews. Book Series, Vol. 3 spatial frequency spectrum can also be modified by employing other speckle reduction schemes such as wavelength and polarization diversity [36]. The output spatial frequency spectrum after the diffuser is further modified by vibrating MMFB having each fiber of equal length. The modification of output angular spectrum of MMFB can be achieved either by changing the length of each fiber in bundle or by changing the illumination strategy [36, 37]. As it is difficult to fabricate a fiber bundle having each fiber of different length, therefore angular illumination strategy is preferred. All modes propagated inside each fiber of the bundle will experience different phase delays depending upon the entrance angles of light beams at the input of MMFB. Therefore, a large number of independent point sources are generated at the output of MMFB, thus generating M independent speckle patterns which add on an intensity basis. Further, vibration to MMFB scrambles all the modes and produce uniform intensity at the output port of MMFB. The combined effects of all aforementioned strategies modify the spatial frequency spectrum and thus generate a pseudo thermal light source having very short LSC length from a high temporally coherent laser. The effective LSC length is reduced due to the combined effects: (1) superposition of the three angular beams at the diffuser plane, (2) various path lengths within the MMFB, (3) mode coupling within the multimode fiber, and (4) numerical aperture (NA) of the imaging lens as utilized in next section. Obtaining such a short LSC length leads to the elimination of parasitic/spurious fringes and speckle contrast in the laser based optical setups. Similar procedure is adopted for He-Ne laser (temporal coherence length ~ 15 cm) to obtain short LSC length pseudo thermal light source. Therefore, with these achievements a laser source can be used for the QPI and OCT of delicate biological samples, which are demonstrated in this chapter. Fig. 17.3. Synthesis of low spatial and high temporal coherent light source with the combined effect of spatial, angular, and temporal diversity. BS1: beam splitter, MO1 – 2: microscope objectives, 50/50: fiber coupler, MMFB: multiple multi-mode fiber bundle. 494 Chapter 17. Quantitative Phase Microscopy and Tomography with Spatially Incoherent Light 17.3.1. Experimental Details To perform QPI and OCT of biological cells, the scattered light at the output of MMFB is coupled into the NIKON microscope (Nikon Eclipse 50i) as shown in Fig. 17.4. The lens L2 nearly collimated the scattered light beam and made incident onto a beam splitter BS2 which directed the beam towards the attached Mirau interferometer objective lens. Mirau interferometer is a highly compact interferometer in the form of an objective lens, which has the capability to produce interferometric signal very quickly even with a very short coherence length light sources such as halogen lamp and LEDs. It contains inbuilt beam splitter which splits the input beam into two beams. Fig. 17.4. Schematic diagram of the experimental set-up; BS1 – 2: beam splitters, MO1 – 2: microscope objectives, 50/50: fiber coupler, MMFB: multiple multi-mode fiber bundle and CCD: charge coupled device [27]. One of the beams goes towards inbuilt reference mirror to produce reference beam and the other one goes towards the sample to generate object beam. Both the beams recombine at the same beam splitter and generate interference pattern, which is finally projected onto a CCD camera plane. The interference pattern is recorded by a CCD camera [Lumenera Infinity 2, 1392×1024 pixels, pixel size: 4∶65×4∶65 μm2]. In order to record five phase shifted interferograms, the Mirau interferometric objective lens is attached with a piezoelectric transducer (PZT) (Piezo, Jena, MIPOS 3), which is driven by an amplifier as shown in Fig. 17.4. PZT moves the inbuilt reference mirror, as well as the imaging lens vertically to introduce a constant phase shift between consecutive interferograms. The five π∕2 phase-shifted interferograms are then recorded by a CCD camera (~10 fps) within 1 s and stored in a personal computer for further processing. During the recording of all frames, samples were within the depth of field of the objective lens. Subsequently, five frame phase shifting algorithm is utilized to extract phase information related to the specimen [38]. The image post processing took ∼50-60 s. 495 Advances in Optics: Reviews. Book Series, Vol. 3 17.4. Phase Retrieval Algorithm 17.4.1. Five Step Algorithm Different QPM techniques work on the principle of interferometry in order to measure the phase shift produced by the biological sample. For quantitative phase recovery, five frame phase shifting algorithm is widely preferred over the other phase shifting interferometry because of moderate phase error and acquisition time [38]. There is trade-off between phase shift error and acquisition time, i.e., if one tries to decrease the phase shift error by increasing the number of frames, the acquisition time gets increased or vice versa. Hariharan proposed a five frame phase shifting algorithm for the phase measurement with acceptable phase measurement error [38].The two-dimensional intensity modulation produced by the superposition of object and reference waves can be written as follows [27, 38]: , 2 , , , cos ∆ , , 3 , (17.12) . (17.13) , and I , are the intensities of the reference and sample beam, where I respectively, n (n = 1-5) corresponds to the different phase shifted interferograms, ∆ϕ , is the phase difference between the sample and reference arm, α is the phase shift between the consecutive phase shifted interference patterns for wavelength λ. Phase difference between the sample (cells + medium) and reference arm can be given by the following expression [38]: ∆ , , , , , , The phase map of the sample (cells + outside medium) ‘φ , ’ can be obtained by subtracting the phase of the reference field from ‘∆ϕ , ’. The reconstructed phase maps can be utilized to determine the refractive index profiles or height map of sample for individual wavelengths using the following expression [27]: , , , ∗2 , , (17.14) where n x, y and n x, y are the refractive indices of cell and medium. h(x, y) is the cell thickness. The Eq. (17.14) can be rewritten into the following form: , ∆ , . (17.15) This expression can be further utilized to quantify corresponding height maps of biological cells. 496 Chapter 17. Quantitative Phase Microscopy and Tomography with Spatially Incoherent Light 17.4.2. Fourier Transform Algorithm Fourier transform algorithm has the capability to recover phase information related to specimen from a single interferometric image. The 2D intensity variation of an interferogram can be expressed as follows [19]: , , ,  cos 2 , (17.16) , , and , are the background (DC) and the modulation terms, where respectively. Spatially-varying phase  , contains information about the specimen. , are the spatial carrier frequencies of interferogram along , axes. In practical , , , and  , are slowly varying functions applications, it is envisaged that compared to the variation introduced by the spatial carrier frequencies , and . The above intensity modulation can be rewritten in the following form for convenience , where ∗ , , , , exp exp 2 2 , exp , The Fourier transform of Eq. 17.17 is given as follows: , , , (17.17) . (17.18) ∗ , , . (17.19) The term , is simply a background (DC) term at the origin in the Fourier plane. The term , corresponds to +1 order term contains information about , . Similarly, ∗ , is –1 order term the object and situated at , which carry complex conjugate information about of the specimen. situated at After applying Fourier filtering of zero and –1 order terms, Eq. (17.19) can be reduced into the following form: , , . (17.20) The filtered spectrum is, first, shifted at the origin and then inverse Fourier transformed , , subsequently the wrapped phase map from the to retrieve the complex signal following expression [19]: , tan , , , (17.21) where and are the imaginary and real part of the complex signal. The reconstructed wrapped phase map lies between – to . These discontinuities are then corrected by 497 Advances in Optics: Reviews. Book Series, Vol. 3 minimum LP–norm two dimensional (2D) phase unwrapping algorithm [39].The retrieved unwrapped phase map can be further used to calculate corresponding height map by employing Eq. (17.15). 17.5. Characterization of System Parameters 17.5.1. Temporal and Spatial Phase Noise High temporal and spatial phase sensitivity is a primary requirement of any QPM setup. The temporal phase sensitivity provides information about the stability of an interferometer, which further leads to measure membrane fluctuations/stiffness of various biological cells [3]. Accurate measurement of membrane fluctuations could be a good indicator of various diseases like sickle cell anemia and cancer etc. However, spatial phase sensitivity of QPM is a measure of spatial phase variation in the reconstructed phase map, which can be measured by imaging a standard flat mirror [3]. Ideally, for a standard flat object, the measured phase variation subsequently the height variation should be equal to zero. The major source of spatial phase variation could be due to the high coherent nature of light source being used to image the specimen [12]. In other words, interference between multiple reflections from the optical surfaces and light scattering from the dust particles etc. degrade the image quality and finally reflect in the phase images. Eventually, spatially phase sensitivity limits the phase measurement accuracy of QPM systems. To measure the temporal and spatial phase sensitivity of the present QPM, we imaged a standard flat mirror, subsequently, a 1 min time lapse interferometric movie was recorded under the stable environmental condition. Fourier transform method, as described in Section 17.4.2, is further utilized to retrieve phase maps corresponding to whole interferometric movie. The variation of phase value of a given pixel as a function of time is a measure of temporal phase sensitivity of the system. Temporal phase noise of the setup is measured in all three different cases (single, double, and triple beam). Since, temporal phase noise of the setup does not depend on the coherence length of the light source, i.e., angular multiplexing of the beam. It only depends on the stability of the interferometer. Therefore, temporal phase noise of the system is found to be almost same in all three cases. The temporal phase noise of the setup is obtained less than 5 mrad, which corresponds to the 0.2 nm sensitivity in the measurement of temporal height variation, for example, cell’s membrane fluctuations, as shown in Fig. 17.5(a). To measure the spatial phase noise of the system, one of the interferogram of flat mirror is utilized [12]. The corresponding recovered phase map is depicted in Fig. 17.5(b). It can be seen from Fig. 17.5(b), recovered phase values are not same at each spatial location of the phase image. The spatial phase noise, i.e., root mean square (RMS) standard deviation, of the system is measured to be ~10 mrad and found to be quite less than that for coherent laser. 498 Chapter 17. Quantitative Phase Microscopy and Tomography with Spatially Incoherent Light Phase (mrad.) Temporal phase sensitivity Spatial phase sensitivity 8 0 6 50 4 100 2 0 0 Phase (mrad.) 20 0 150 -20 20 40 Time (Sec) 60 (a) 200 0 100 200 (b) Fig. 17.5. Phase sensitivity of the present setup. (a) Temporal phase sensitivity. (b) Spatial phase sensitivity. Color bar represents phase in mrad. Reproduced from [27], with the permission of OSA Publishing, © 2016 OSA. 17.5.2. Transverse Resolution For the measurement of transverse resolution of the system, we captured an image of a standard USAF test target using an objective lens of NA 0.3 (10×) and wavelength 532 nm (Fig. 17.6(a)). The field of view of our experimental setup was 2.0×1.5 mm2 for 10×, objective lens. Fig. 17.6(b) represents the line profile of resolution target along the red line enclosed in blue color box. It is clear from Figs. 17.6(a) and (b) that we are able to resolve the 6th element of 7th group. The measured lateral resolution was obtained to be 2.19 µm, and calculated value is found to be 1.09 µm using the formula (0.61× )/NA. Fig. 17.6. Characterization of the transverse resolution of the microscope. (a) Standard USAF resolution chart image recorded by employing objective lens having NA 0.3 (10×) at 532 nm wavelength. (b) Normalized line profile of 5th and 6th element in 7th group enclosed in blue color box. Reproduced from [23], with the permission of AIP Publishing, © 2015 AIP. 17.5.3. Axial Resolution The axial resolution determined from short LSC length rather than TC length plays an important role in high-resolution topography and tomography of industrial and biological objects. As already discussed in the introduction section, the use of pseudo thermal light 499 Advances in Optics: Reviews. Book Series, Vol. 3 source does not require any dispersion compensation and chromatic aberration corrected optics, which are otherwise mandatory in case of broadband light sources [40]. For the measurement of axial resolution, a flat mirror as a test sample is placed under the microscope and scanned vertically. The sequential interferograms are then recorded as a function of vertical scanning of the sample mirror. The sample mirror is translated vertically in a step of 1 µm for the measurement of LSC function of the light source. It is observed that the fringe visibility reduces as the sample mirror go away from the zero OPD position. Further, fringe visibility is plotted as a function of sample mirror vertical position for the measurement of LSC length. The FWHM of the fringe visibility curve thus obtained provides information about the LSC length of synthesized light source. Once the information about coherence envelope is obtained, the axial resolution (half of LSC length) of the microscope can be determined. For the determination of axial resolution as a function of wavelength, two Mirau interferometric objective lenses having NA 0.3 (10×) and NA 0.4 (20×) were used in the experimental setup. Here, two different temporally high coherent lasers: DPSS laser ) and He-Ne laser ( ~15 ) are utilized for measurement. Figs. 17.7(a) and ( ~6 (b) depict the LSC functions for NA 0.3 (10×) and 0.4 (20×) Mirau interferometric objective lenses at 532 nm wavelength respectively. Similarly, the LSC functions for NA 0.3 (10×) and 0.4 (20×) objective lenses at 632 nm wavelength are illustrated in Figs. 17.7(c) and (d) respectively. Fig. 17.7. Measurement of the LSC length of setup. The LSC length is measured for the objective lens (a) NA 0.3 (10×), and (b) NA 0.4 (20×) at 532 nm wavelength; (c) NA 0.3 (10×), and (b) NA 0.4 (20×) at 632 nm wavelength. Reproduced from [23], with the permission of AIP Publishing, © 2015 AIP. 500 Chapter 17. Quantitative Phase Microscopy and Tomography with Spatially Incoherent Light For 532 nm wavelength, axial resolution is measured of the order of 9.5 µm and 4 µm for objective lenses of NA 0.3(10×) and 0.4 (20×), respectively, as shown in Figs. 17.7(a) and (b).The calculated values are found to be 5.77 µm and 3.18 µm, for NA 0.3 and 0.4, respectively, using Eq. (17.11), at the central wavelength of 532 nm. Similarly, the measured axial resolution for red wavelength (632 nm) is found to be 11 µm and 5 µm for objective lenses of NA 0.3 (10×) and 0.4 (20×), respectively, as shown in Figs. 17.7(c) and (d) and the calculated values are found to be equal to 6.86 µm and 3.78 µm, respectively, using Eq. (17.11) at the central wavelength 632 nm. The short LSC length thus achieved may find potential application in QPM and high resolution optical sectioning of multilayered biological specimens. 17.6. Spatial Phase Noise Comparison in Case of Direct Laser and Synthesized Light Source 17.6.1. Standard Flat Mirror For the comparison of spatial phase sensitivity of the phase microscope while using two different light sources such as direct laser (type ‘A’) and synthesized source(type ‘B’), a standard flat mirror is placed under the microscope and corresponding interferometric images are acquired by the CCD camera. These interferometric images are further postprocessed for the phase recovery using Fourier transform based phase retrieval algorithm Eq. (17.21). Figs. 17.8(a) and (d) illustrate the interferometric images of the standard flat mirror for type ‘A’ and type ‘B’ light sources, respectively. Fig. 17.8. Spatial phase noise comparison of QPM generated due to direct laser and synthesized light source. (a, d) interferogram of the standard flat mirror, (b, e) recovered phase map (the color bar is in rad.), and (c, f) corresponding height map (the color bar is in nm) for two different light sources: He-Ne laser and synthesized light source, respectively [12]. 501 Advances in Optics: Reviews. Book Series, Vol. 3 The corresponding reconstructed phase and height maps of the standard flat mirror are shown in Figs. 17.8 (b, e) and (c, f) respectively. The spatial phase sensitivity of the microscope using type ‘A’ and type ‘B’ light sources are found to be approximately equal to 112 mrad and 8.3 mrad, respectively. The corresponding sensitivities in the height measurements are found to be 5.62 nm and 0.42 nm. It is quite evident from Figs. 17.8(b, e) that the coherence property of the light source plays an important role in generation of spatial phase noise. It further leads to introduce an unwanted error in the phase measurement of specimen. Approximately 90 % enhancement in the spatial phase sensitivity is observed while using synthesized light source rather than direct laser [12]. The synthesized light source is further used to perform QPM and OCT of some industrial as well as biological objects. 17.6.2. Human Red Blood Cells The effect of coherence properties of light sources: direct laser and synthesized light, are further studied on the QPI of RBCs. As mentioned earlier that synthesized light source does not generate coherent noise in the images. It is depicted from Fig. 17.9(a), strong non-uniform intensity distribution is observed in case of direct laser due to high coherent nature of the light source. In contrast, synthesized light source produces quite uniform intensity distribution throughout the image (Fig. 17.9(d)). Figs. 17.9(b) and (e) present the interferometric images of RBCs, which are recorded using direct laser and synthesized light source respectively. Fig. 17.9. QPI of RBCs using direct laser and synthesized light source and comparison. (a, d) noninterferometric, (b, e) interferometric images, and (d, f) recovered phase maps of human RBCs for two different light sources: He-Ne laser and synthesized light source respectively. The color bars represent phase in rad. 502 Chapter 17. Quantitative Phase Microscopy and Tomography with Spatially Incoherent Light The insets of Figs. 17.9 (b) and (e) clearly depict the fringe quality difference in the interferometric images generated from highly coherent and low spatial coherent light source. The synthesized low spatial coherent light source produces quite smooth interferometric fringes. These recorded interferometric images are further processed to generate corresponding phase maps of human RBCs. The recovered phase maps of RBCs corresponding to highly coherent and pseudo thermal light sources are illustrated in Figs. 17.9 (c) and (f). It is observed from the phase images that high coherent light source produces unwanted spatial phase variation in the reconstruction, which is almost negligible in case of synthesized pseudo thermal light source. 17.7. Quantitative Phase Imaging of Industrial and Biological Cells Using Pseudo Thermal Light Source 17.7.1. QPI of Standard Waveguide To quantify the phase measurement accuracy of the present technique, experiment is performed on a standard strip waveguide chip of known height. The core layer of strip waveguide chip (height ~220 10 nm) is made of tantalum pentoxide (Ta2O5). Ta2O5 has refractive index value around 2.1359 at an operating wavelength 632 nm. The height profile of the waveguide chip is first measured from a standard surface profilometer (P-7 stylus profiler). The height of Ta2O5 layer was found to be equal to 224.8 nm as shown in Fig. 17.10(a). Further, the experiment is performed on the same waveguide chip using the proposed technique shown in Fig 17.4. Fig. 17.10(b) shows one of the interferogram selected from a set of five phase shifted interferograms, which are recorded using 50× (NA 0.55) Mirau interferometric objective lens. These five phase shifted interferograms are further utilized to measure the phase as well as the height profile of waveguide chip. The phase and height map of the chip are calculated using Eqs. (17.14) and (17.15), as shown in Figs. 17.10(c) and (d), respectively. The height of the chip is found to be equal to 225.5 nm, which is very close to the value obtained using surface profilometer. 17.7.2. QPI of Human RBCs Next, the experiment was performed for the QPI of RBCs using a Mirau-interferometric objective of magnifications 50× (NA = 0.55). The 50× objective provides a sufficient field-of-view and required transverse resolution for imaging of Human RBCs. A set of five phase shifted interferograms are recorded with and without vibrating the MMFB for the reconstruction of RBCs phase map as shown in Fig. 17.11. Figs. 17.11(a) and (b) clearly depict the difference in the interferometric image quality of RBCs without and with vibrating MMFB. It can be seen that the image quality is quite bad in case of static MMFB, i.e., RBCs are not visible in FOV of camera. Whereas, vibrating MMFB significantly improves the image as well as fringe contrast as shown in Fig. 17.11(b). In order to reduce the speckle contrast, MMFB is vibrated at a constant frequency ~15 Hz. 503 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 17.10. (a) Height map of waveguide chip measured from surface profilometer, (b) interferogram, (c) unwrapped phase map, and (d) height map of the same chip using present setup at 632 nm wavelength. Reproduced from [27], with the permission of OSA Publishing, © 2016 OSA. Fig. 17.11. Effect of non-vibrating and vibrating MMFB on the QPI of RBCs. (a, b) Interferograms, and (c, d) corresponding phase map of Human RBCs, without and with vibrating MMFB respectively. Reproduced from [27], with the permission of OSA Publishing, © 2016 OSA. 504 Chapter 17. Quantitative Phase Microscopy and Tomography with Spatially Incoherent Light To reconstruct the phase map of RBCs five equally π/2 phase shifted interferograms were recorded by the CCD camera. The exactly π/2 phase shift is introduced between the reference and sample arm with the help of PZT. Figs. 17.11(a) and (b) show one of the interferogram from five phase shifted interferograms without and with vibrating MMFB, respectively. The five phase shifted interferograms are then utilized to calculate wrapped phase maps using five step phase retrieval algorithm given in Eq. (17.14). The Minimum LP-norm two-dimensional phase unwrapping algorithm [39] is used to remove the discontinuities from the wrapped phase map of sample. Figs. 17.11(c) and (d) shows the unwrapped phase maps of RBCs (image size ~48 μm × 60 μm) at 632 nm wavelength corresponding to vibrating and non-vibrating MMFB respectively. In Fig. 17.11(d), the artifacts present in the reconstructed phase image of RBCs could arise due to the phase shift error between consecutive data frames. The major sources of these errors are the uncalibrated phase shifter and environmental fluctuations. Ideally, phase shift between the two consecutive frames must be equal to π/2 or a constant number ‘α’ for the artifacts free phase images. If somehow PZT does not introduce exact phase shift between the frames then such artifacts can be present in the reconstructed phase images. It can be visualized from Fig. 17.11(c), it is difficult to retrieve phase information related to biological cells in case of static MMFB, which is completely filled with the speckle noise. There are various computational approaches present in the literature which have the capability to recover phase information from speckle images [41, 42]. But this is not the scope of this chapter. The speckle noise is effectively reduced with vibrating MMFB due to the temporal averaging of the speckle patterns (Fig. 17.11(b)). Therefore, vibrating MMFB is a smart choice for the speckle free phase imaging of biological cells with high spatial phase sensitivity. The maximum phase value of the RBCs is found to be 3.042 rad. with vibrating MMFB. 17.7.3. QPI of Onion Cells Similarly, the experiment was performed for the QPI of onion cells at 632 nm wavelength with vibrating MMFB. This time 10× (NA = 0.33) Mirau objective lens is preferred to image onion cells as the size of onion cells are bigger than RBCs. The 10× objective lens provides a sufficient field of view and required transverse resolution for imaging onion cells. Fig. 17.12(a) shows one of the interferogram selected from a set of five phase shifted interferograms recorded at 632 nm wavelength. The unwrapped phase map for onion layer is obtained by following the similar procedure opted for RBCs (Fig. 17.12(b)). It is clear from the phase map; onion skin is not uniform due to different thicknesses of the cells at different positions. Therefore, the OPD is also different at each position. The height map of onion cells is obtained by using the value of ∆n equal to 0.3345 in Eq. (17.15) at red wavelength. This value is obtained by using refractive index values of onion layer and outside medium equal to 1.3345 and 1.0 at red wavelength, respectively [43]. Figs. 17.12(c) and (d) show height map and corresponding line profile of onion cells marked with blue box, respectively. The maximum height of the cell is found to be ~3 μm at the center of the cell. 505 Advances in Optics: Reviews. Book Series, Vol. 3 Fig. 17.12. QPI of onion cells. (a) Interferogram of onion cells, (b) unwrapped phase map, (c) height maps of few onion cells, and (d) corresponding line profile at red wavelength. Reproduced from [27], with the permission of OSA Publishing, © 2016 OSA. 17.8. Profilometry and Optical Coherence Tomography 17.8.1. Profilometry of Standard Gauge Block and Indian Five Rupee Coin To demonstrate tomographic capability of the setup, the experiment is performed on standard gauge blocks. Two standard gauge blocks with height difference (i.e., step height) of 5 µm are placed under the microscope. It is clear from Figs. 17.13 (a) and (b), that initially high contrast fringe is formed on the left gauge block while the right side gauge block is out of LSC length of the light source. As the sample stage is moved upward by ~5 µm, the high contrast fringes sifted from left to right side gauge block (Fig. 17.13 (b)). The reconstructed height map of the gauge blocks obtained from these two interference patterns is shown in Fig. 17.13 (c). The feasibility of present setup is demonstrated by measuring the height of letter ‘R’ imprinted on a five rupee Indian coin. A 20× Mirau interferometric objective lens is used to generate interferogram, which occurs due to superposition of scattered light from the coin and the reference beam. The field of view 20× objective lens is 910×680 µm2 (1392×1040 pixels). For height measurement of the coin, the sample stage is translated vertically in a step size of 1 µm and a series of corresponding interferograms are captured. For phase recovery, five equal phase shifted interferograms are then recorded and analyzed using Eq. (17.13) at each vertical position of the sample stage. The phase shift between data frames is introduced with the help of PZT attached to the objective lens. Subsequently, the phase ambiguities are corrected using Minimum LP-norm 2D phase 506 Chapter 17. Quantitative Phase Microscopy and Tomography with Spatially Incoherent Light unwrapping algorithm[39].In order to acquire en-face OCT images of the feature on coin, the sample platform is scanned vertically, until the depth of 19 µm. Since, the LSC lengths obtained for DPSS and He-Ne laser are 8 µm and 10 µm respectively. Therefore, high contrast interference fringes are observed only when the condition of zero OPD is satisfied. In other words, foreground and background of letter ‘R’ will not produce interference pattern simultaneously. As the sample stage is translated vertically, the fringe contrast shifted from the foreground to background of letter ‘R’. In this way, height of letter ‘R’ is measured to be 19 µm. The 3D volumetric images with dimensions (X) 557 × (Y) 437 × (Z) 19 µm3 of the letter ‘R’ for green and red wavelengths are shown in Figs. 17.13(d) and (e) respectively. Fig. 17.13. Profilometry of industrial objects. (a, b) interference pattern on the standard gauge block, and (c) reconstructed height map, Volumetric image of letter ‘R’ of the coin using 20× microscope objective at (d) 532 nm, and (e) 632 nm wavelength. Reproduced from [23], with the permission of AIP Publishing, © 2015 AIP. 17.8.2. OCT of Multilayered Onion Sample To demonstrate the feasibility of setup, we performed experiment on the multilayered onion sample. For the sample preparations, two onion slices are placed one over the other on a reflecting mirror. The onion layers are placed one over the other in a way that the upper and bottom layers are oriented perpendicular to each other as shown in Figs. 17.14(a) and (d). The interferometric images of each onion layer are recorded using 20× (NA 0.4)interferometric objective lens at wavelength 532 nm as depicted in Figs. 17.14 (a) and (b). It is clear from Fig. 17.14 (a), the interference fringes are formed only for the light reflected from top layer due to short LSC length of the light source. As 507 Advances in Optics: Reviews. Book Series, Vol. 3 the sample stage is moved vertically, the interference pattern is shifted from the upper layer to the bottom layer of the onion sample (Fig. 17.14 (b)). Phase (rad.) 0 0 (a) 100 Phase (rad.) 0 (b) 2 100 (c) 40 100 30 0 200 200 300 0 300 0 20 200 10 -2 200 400 200 300 0 400 200 400 Phase (rad.) 0 0 (d) 100 100 200 200 Phase (rad.) 0 (e) 2 40 (f) 30 100 0 20 200 10 -2 300 0 200 400 300 0 200 400 300 0 200 400 0 Fig. 17.14. Optical sectioning of multilayered onion sample. (a, d) Interferograms of top and bottom layer of onion sample, (b, e) corresponding wrapped phase map, and (c, f) unwrapped phase maps of the top and bottom layers respectively. The color bars represent phase in rad. Reproduced from [23], with the permission of AIP Publishing, © 2015 AIP. For phase recovery, five phase shifted interferograms corresponding to each onion layers are recorded. The calculated wrapped phase maps are illustrated in Figs. 17.14(b) and (e). The wrapped phase maps of top and bottom layers are then corrected using 2D phase unwrapping algorithm as depicted in Figs. 17.14(c) and (f) respectively. Finally, it can be seen from the recovered phase images, short LSC length instead of TC length can be utilized for the optical sectioning of multilayered biological specimens. 17.9. Conclusions This chapter provides an overview about the coherence properties of the light sources and further their role in QPM and OCT techniques is studied. Most of the OCT techniques generally utilized broadband light sources to obtain high axial-resolution and speckle free imaging. However, broadband light sources require dispersion compensation mechanism for dispersion correction which adds complexity to the system. In addition, chromatic aberration corrected optical components are also mandatory for the noise free images. To overcome these limitations, a narrowband light source like laser could be a possible solution. But these light sources also suffer from the problem of poor axial resolution due to high temporal coherence length and coherent or speckle noise, which reduces sensitivity of the system. 508 Chapter 17. Quantitative Phase Microscopy and Tomography with Spatially Incoherent Light To avoid these problems associated with broadband and narrowband light sources both, a spatially low and temporally high coherent source with the combined effect of spatial, angular and temporal diversity is synthesized. To exhibit potential of the technique, QPI of biological cells (RBCs and onion cells) using such light source is demonstrated. The spatial phase sensitivity of the synthesized light source is found to be quite high compared to direct laser. Moreover, it significantly reduces the problem of parasitic/spurious fringe formation and speckle noise from the images. Further, LSC length of synthesized pseudo thermal light source is found to be equal to ~8 µm with 20× objective lens at 532 nm wavelength. The short LSC length instead of TC length is further utilized to perform high axial-resolution sectioning of multilayered objects. The present system does not require any dispersion compensation optical system for biological samples as a highly monochromatic light source is used. In addition, it is useful for obtaining morphological information about the cells and tissues with good accuracy and precision. Acknowledgements The authors are thankful to Department of Atomic Energy (DAE), Board of Research in Nuclear Sciences (BRNS) for financial grant no. 34/14/07/BRNS and DST, Govt. of India, Project No.: SB/S2/LOP-026/2013. References [1]. J. Rosen, M. Takeda, Longitudinal spatial coherence applied for surface profilometry, Applied Optics, Vol. 39, 2000, pp. 4107-4111. [2]. F. Dubois, M.-L. N. Requena, C. Minetti, O. Monnom, E. Istasse, Partial spatial coherence effects in digital holographic microscopy with a laser source, Applied Optics, Vol. 43, 2004, pp. 1131-1139. [3]. G. Popescu, Quantitative Phase Imaging of Cells and Tissues, McGraw Hill Professional, 2011. [4]. W. Drexler, J. G. Fujimoto, Optical Coherence Tomography: Technology and Applications, Springer, 2015. [5]. H. M. Subhash, Full-field and single-shot full-field optical coherence tomography: A novel technique for biomedical imaging applications, Advances in Optical Technologies, Vol. 2012, 2012, 435408. [6]. L. Mandel, E. Wolf, Optical Coherence and Quantum Optics, Cambridge University Press, 1995. [7]. V. Ryabukho, D. Lyakin, A. Grebenyuk, S. Klykov, Wiener–Khintchin theorem for spatial coherence of optical wave field, Journal of Optics, Vol. 15, 2013, 025405. [8]. I. Abdulhalim, Spatial and temporal coherence effects in interference microscopy and full‐ field optical coherence tomography, Annalen der Physik, Vol. 524, 2012, pp. 787-804. [9]. M. K. Kim, Digital Holographic Microscopy: Principles, Techniques, and Applications, Springer, New York, 2011. [10]. T. Colomb, F. Montfort, J. Kühn, N. Aspert, E. Cuche, A. Marian, F. Charrière, S. Bourquin, P. Marquet, C. Depeursinge, Numerical parametric lens for shifting, magnification, and complete aberration compensation in digital holographic microscopy, J. Opt. Soc. Am. A, Vol. 23, 2006, pp. 3177-3190. 509 Advances in Optics: Reviews. Book Series, Vol. 3 [11]. K. Lee, K. Kim, J. Jung, J. Heo, S. Cho, S. Lee, G. Chang, Y. Jo, H. Park, Y. Park, Quantitative phase imaging techniques for the study of cell pathophysiology: from principles to applications, Sensors, Vol. 13, 2013, pp. 4170-4191. [12]. A. Ahmad, V. Dubey, V. Singh, A. Butola, T. Joshi, D. S. Mehta, Reduction of spatial phase noise in the laser based digital holographic microscopy for the quantitative phase measurement of biological cells, Proceedings of SPIE, Vol. 10414, 2017, 104140H. [13]. D. Singh Mehta, V. Srivastava, Quantitative phase imaging of human red blood cells using phase-shifting white light interference microscopy with colour fringe analysis, Appl. Phys. Lett., Vol. 101, 2012, 203701. [14]. V. Dubey, G. Singh, V. Singh, A. Ahmad, D. S. Mehta, Multispectral quantitative phase imaging of human red blood cells using inexpensive narrowband multicolor LEDs, Applied Optics, Vol. 55, 2016, pp. 2521-2525. [15]. R. Guo, B. Yao, J. Min, M. Zhou, X. Yu, M. Lei, S. Yan, Y. Yang, D. Dan, LED-based digital holographic microscopy with slightly off-axis interferometry, Journal of Optics, Vol. 16, 2014, 125408. [16]. B. Bhaduri, H. Pham, M. Mir, G. Popescu, Diffraction phase microscopy with white light, Optics Letters, Vol. 37, 2012, pp. 1094-1096. [17]. P. Girshovitz, N. T. Shaked, Doubling the field of view in off-axis low-coherence interferometric imaging, Light: Science and Applications, Vol. 3, 2014, p. e151. [18]. V. Dubey, V. Singh, A. Ahmad, G. Singh, D. S. Mehta, White light phase shifting interferometry and color fringe analysis for the detection of contaminants in water, Proceedings of SPIE, Vol. 9718, 2016, 97181F. [19]. M. Takeda, H. Ina, S. Kobayashi, Fourier-transform method of fringe-pattern analysis for computer-based topography and interferometry, J. Opt. Soc. Am. A, Vol. 72, 1982, pp. 156-160. [20]. H. Schreiber, J. H. Bruning, Phase shifting interferometry, Chapter 14, in Optical Shop Testing (D. Malacara, Ed.), 3rd Edition, John Wiley & Sons Inc., Hoboken, NJ, USA, 2006, pp. 547-666. [21]. L. Vabre, A. Dubois, A. C. Boccara, Thermal-light full-field optical coherence tomography, Optics Letters, Vol. 27, 2002, pp. 530-532. [22]. D. S. Mehta, V. Srivastava, S. Nandy, A. Ahmad, V. Dubey, Full-field optical coherence tomography and microscopy using spatially incoherent monochromatic light, Chapter 10, in Handbook of Full-Field Optical Coherence Microscopy, Pan Stanford Publishing Pte. Ltd., 2016, pp. 357-392. [23]. A. Ahmad, V. Srivastava, V. Dubey, D. Mehta, Ultra-short longitudinal spatial coherence length of laser light with the combined effect of spatial, angular, and temporal diversity, Appl. Phys. Lett., Vol. 106, 2015, 093701. [24]. P. Pavliček, M. Halouzka, Z. Duan, M. Takeda, Spatial coherence profilometry on tilted surfaces, Applied Optics, Vol. 48, 2009, pp. H40-H47. [25]. A. Safrani, I. Abdulhalim, Ultrahigh-resolution full-field optical coherence tomography using spatial coherence gating and quasi-monochromatic illumination, Optics Letters, Vol. 37, 2012, pp. 458-460. [26]. M. Gokhler, J. Rosen, General configuration for using the longitudinal spatial coherence effect, Optics Communications, Vol. 252, 2005, pp. 22-28. [27]. A. Ahmad, V. Dubey, G. Singh, V. Singh, D. S. Mehta, Quantitative phase imaging of biological cells using spatially low and temporally high coherent light source, Opt. Lett., Vol. 41, 2016, pp. 1554-1557. [28]. Y. Park, W. Choi, Z. Yaqoob, R. Dasari, K. Badizadegan, M. S. Feld, Speckle-field digital holographic microscopy, Optics Express, Vol. 17, 2009, pp. 12285-12292. [29]. V. Ryabukho, D. Lyakin, M. Lobachev, Longitudinal pure spatial coherence of a light field with wide frequency and angular spectra, Optics Letters, Vol. 30, 2005, pp. 224-226. 510 Chapter 17. Quantitative Phase Microscopy and Tomography with Spatially Incoherent Light [30]. I. Abdulhalim, Competence between spatial and temporal coherence in full field optical coherence tomography and interference microscopy, Journal of Optics A: Pure and Applied Optics, Vol. 8, 2006, pp. 952-958. [31]. V. Ryabukho, D. Lyakin, V. Lychagov, Longitudinal coherence length of an optical field, Optics and Spectroscopy, Vol. 107, 2009, pp. 282-287. [32]. V. Ryabukho, A. Kal’yanov, D. Lyakin, V. Lychagov, Influence of the frequency spectrum width on the transverse coherence of optical field, Optics and Spectroscopy, Vol. 108, 2010, pp. 979-984. [33]. V. Ryabukho, D. Lyakin, The effects of longitudinal spatial coherence of light in interference experiments, Optics and Spectroscopy, Vol. 98, 2005, pp. 273-283. [34]. C. McCutchen, Generalized source and the Van Cittert–Zernike theorem: a study of the spatial coherence required for interferometry, Journal of the Optical Society of America, Vol. 56, 1966, pp. 727-733. [35]. J. W. Goodman, Statistical Optics, John Wiley & Sons, 2015. [36]. J. W. Goodman, Speckle Phenomena in Optics: Theory and Applications, Roberts and Company Publishers, 2007. [37]. D. S. Mehta, D. N. Naik, R. K. Singh, M. Takeda, Laser speckle reduction by multimode optical fiber bundle with combined temporal, spatial, and angular diversity, Applied Optics, Vol. 51, 2012, pp. 1894-1904. [38]. P. Hariharan, B. Oreb, T. Eiju, Digital phase-shifting interferometry: a simple errorcompensating phase calculation algorithm, Appl. Opt., Vol. 26, 1987, pp. 2504-2506. [39]. D. C. Ghiglia, L. A. Romero, Minimum Lp-norm two-dimensional phase unwrapping, J. Opt. Soc. Am. A, Vol. 13, 1996, pp. 1999-2013. [40]. J. Wang, J.-F. Léger, J. Binding, A. C. Boccara, S. Gigan, L. Bourdieu, Measuring aberrations in the rat brain by coherence-gated wavefront sensing using a Linnik interferometer, Biomedical Optics Express, Vol. 3, 2012, pp. 2510-2525. [41]. M. Trusiak, K. Patorski, M. Wielgus, Adaptive enhancement of optical fringe patterns by selective reconstruction using FABEMD algorithm and Hilbert spiral transform, Optics Express, Vol. 20, 2012, pp. 23463-23479. [42]. M. B. Bernini, A. Federico, G. H. Kaufmann, Normalization of fringe patterns using the bidimensional empirical mode decomposition and the Hilbert transform, Applied Optics, Vol. 48, 2009, pp. 6862-6869. [43]. W. Wang, C. Li, Measurement of the light absorption and scattering properties of onion skin and flesh at 633 nm, Postharvest Biology and Technology, Vol. 86, 2013, pp. 494-501. 511 Index Index 3D micro-stereolithography, 205 A Abbe number, 438 Abrasive Wear, 467, 468, 481 Acceptance angle, 316, 325, 326, 338, 340, 351, 355, 364, 366, 372, 373, 376, 378, 382, 385, 392, 401-403, 405, 407, 408 active optical materials, 214 Additive white Gaussian noise. See AWGN adjacent heights, 422 Adults module, 40 aerosol structures multi-layering, 431 akinetic beam scanner, 213 Akinetic Beam Scanner, 216 AliveCor, 27 altitude range, 425, 429 Analysis Linear Discriminant, 188 of variance, 188 Quadratic Discriminant, 188 Angular distribution, 316, 318, 332, 354, 394 divergence, 316, 322, 346, 347, 367, 401, 402 resolution, 321, 339, 346, 355, 356, 377, 394 sector, 421, 425, 426 arm stroke, 482 aspherical optical lenses, 435 AWGN, 259 Axial Resolution, 499 B Baby's cradle module, 40 backscatter signal, 419, 423, 429, 430 bacteria identification, 21–20 resistant, 169 band-pass filtering, 145 basic algorithms, 419 beam propagation method, 117 splitter/combiner, 219 beat length, 108 BER, 248, 255, 260, 261, 269 biosensors optical, 169 bit error rate. See BER BodyMedia’s, 27 Bragg condition, 302 frequency, 200 resonant dip, 203 Bragg-reflection waves, 200 burning agricultural, 418 area, 418, 432 C CCD camera, 349, 358, 359, 361, 385 CellNovo, 27 cells bacterial, 171 Chromatic aberration, 392 Cisco’s Connected Factory, 28 classification, 188 Cloud, 24, 25, 28, 38 CMOS fabrication, 104 CNC machines, 435 coefficient 2D transmission, 173 Coherence, 489 coherent gradient sensor, 129 colony bacterial, 170 Communication Models, 24 component analysis, 231 Component Analysis, 232, 233 independent component analysis, 232 Independent Component Analysis, 232 Principal Component Analysis, 232 Concentration ratio geometric, 315, 326, 328, 330 optical, 315, 328, 331, 332, 401, 407 Concentrator 513 Advances in Optics: Reviews. Book Series, Vol. 3 2D, 328, 401, 402, 405 3D, 317, 326, 327, 329, 338-342, 359, 401-405, 407 characterization, 316-318, 320, 325, 331, 342, 346, 348, 362, 373, 394, 395, 411 CPC, 316, 317, 326, 329, 335, 338-344, 355, 356, 359, 361, 362, 365-367, 369-385, 395, 401-407 direct collimated absorptance, 325 reflectance, 325 transmittance, 325 direct Lambertian absorptance, 330 concentration ratio, 331 optical conductance, 336 optical conductivity, 337 reflectance, 330 transmittance, 330 inverse irradiation, 317, 318, 319 inverse Lambertian optical conductance, 336 optical conductivity, 337 transmittance, 333 optical losses, 327-329, 343, 344 PhoCUS, 344, 348, 353, 359, 364, 391394 photovoltaic, 315-317, 320 reflective, 316, 317, 359, 402 refractive, 317, 353, 364 Rondine, 343-345, 350-352, 361-364, 367-369, 385-391, 395, 412, 414 thermodynamic, 315, 316, 321 transmission efficiency, 316, 318-320, 323, 324, 328-333, 338-342, 346, 348, 352, 354, 363, 366, 376, 377, 387, 388, 390, 394, 401, 402, 406, 409, 410 transmission properties, 316, 317 confocal microscopy, 450 constitutive model, 445-447, 458 Contrast. Véase Visibility Cost, 315, 350, 395, 411 coupling coefficients, 106 CPC squared, 342, 343, 369, 378 truncated, 338, 342, 343, 381, 383 cross-sensitivity, 86 Cryogenic Medium, 134 514 Cryogenic Vacuum Chamber, 138 curvature radius, 468, 482, 484 curvatures, 133 D data CALIPSO, 418 data points clusters of, 425 delay phase, 175 detection limit, 113 Threshold Level, 264 developer AZ303, 304 DHL’s IoT Tracking and Monitoring, 28 Differential signalling Channel modelling, 266, 283 Diffraction efficiency, 301 Diffractive optics, 435 Disabled and elderly module, 40 distance, 231, 233, 238-243 dynamic heart rate estimation, 231, 233 E effective refraction index, 92 electroforming. See Silver spray Electroless nickel-phosphor, 439 electro-optic coupler, 217 switch, 214, 216, 217, 222 envelope function, 421, 422, 429 Ericsson Maritime ICT, 28 etching, 435 Étendue, 327, 329, 336 ethanol, 104 extrinsic FP fiber temperature sensors, 93 F factors antibacterial, 190 features extraction, 188 FFT method, 141 Fiber Bragg Grating, 35, 36 optic interferometers, 34 Optic Sensors, 21, 28, 30, 33 optical, 173 Index finite difference method, 111 element, 444, 445 Fire Cedar, 423 Observation, 423 Roaring Lion, 423, 432 Fitbit Flex, 27 Five Step Algorithm, 496 Fizeau interferometer, 450, 451 Flux distribution, 315, 388-390 Focal length, 338, 349, 350, 358, 367, 368, 391, 402, 403 focused ion beams, 443 Fourier Transform Algorithm, 497 spectroscopy, 220 four-wave mixing (FWM), 52 Free space optical. See FSO spectral range, 70, 120 frequency domain, 145 Fresnel equations, 318 lenses, 445 reflection, 85, 95 FSO, 248 Aperture averaging factor, 258 Atmospheric attenuation, 254 Differential signalling, 261 Fog attenuation, 254 Geometrical loss, 254 Miscellaneous attenuation, 255 Pointing errors, 256 Scintillation/turbulence, 255 System diagram, 251, 253 Turbulence, 256 G Generalized Van-Cittert–Zernike theorem, 492 glass, 436-440, 444-446, 448, 450, 451, 457, 458, 468, 469, 482, 484 Glassy carbon, 440 glucose, 104 grinding, 435, 436, 439-443, 467-471, 475, 482, 484 H health and well-being applications, 29 Health and wellness, 21, 26 Health and Wellness Application, 28 heat treatment, 443 He-Cd laser, 305 height interval, 420, 426 He-Ne laser, 304 hetero-core fiber, 33 High energy beams, 443 holographic optical elements HOE, 301 solar concentration, 311 weapon sight, 301 holography digital, 193 homogeneous broadening, 49, 60 homogeneous mechanism, 111 horizontal layering smoke-plume, 427, 432 well-defined, 426 well-developed, 427, 428 I iHealth, 27 impulse excitation method, 446 increased backscattering layers of, 423, 425, 432 index of refraction, 438 inhomogeneous broadening, 50 injection height, 427 plume, 427 integrated optics, 213 intensity-modulated sensor, 30 intercept function, 420 interference mechanisms, 104 interferometric techniques, 450 Internet of Things, 21, 22, 24, 41 intrinsic FP fiber sensors, 99 Inverse method, 316, 320, 321, 338, 353355, 357-359, 376, 379, 381-383, 393, 394, 408, 409 ion-sliced lithium niobate film, 214 IoT Systems for the Family, 38 iRobot’s Roomba, 28 Irradiation diffuse, 317, 331 direct, 317-320, 333, 366, 378, 394 515 Advances in Optics: Reviews. Book Series, Vol. 3 lambertian, 316, 317, 320, 321, 331, 333, 336 K Kumanin, 468, 469, 476, 480, 482, 484 L L-3 EO-tech, 301 Lambertian diffuser, 351, 352, 376, 378 light, 321, 350, 354, 359-361, 411 source, 316, 317, 320-322, 331, 335, 338, 354, 359, 363, 364, 367, 368, 389 lapping, 435 Laser, 360, 364, 366, 367 beam, 323, 342, 344, 359, 360, 365-367, 369, 373, 374, 376 method, 365, 370, 371, 377 lens arrays, 436, 439 surface, 482 lidar ground-based, 418 scanning, 418, 419, 421, 423, 426, 428, 432 backscatter signal, 429 profiling, 417 light diffraction on bacterial colonies, 169, 178 LN-on-silicon waveguide platform, 221 Longitudinal Spatial Coherence, 492 function, 492 loss factor, 106 M Machine Support Vector, 188 Mach-Zehnder interferometer, 221 Manchester code, 290 maximum fringe contrast, 90 measurement interruptions, 425 Medtronic’s, 27 metal waveguides, 197 metal-rod-array, 204 method data processing, 419 516 gradient, 419 Method direct, 316, 347-349, 358, 363, 392, 395 direct collimated, 319, 323, 324 direct Lambertian, 321-323, 329 direct local collimated, 318, 338 inverse Lambertian, 319, 320, 323, 325, 333, 353 inverse local Lambertian, 320 mixed Lambertian, 323 Parretta, 359, 388 Parretta-Herrero, 319, 321, 367, 368, 388 PH-Method, 319, 321, 367, 368, 388390 P-Method, 319, 320, 359, 388-390 methodology advanced data processing, 419 Michelson interferometer, 221, 490 Micro and macrobending, 31 Micro-lens arrays, 435 microring resonators, 104 microscope transmission, 173 microspectrometer, 221 milling, 436, 443 Mirau interferometer, 495 Mirror film, 343, 345 parabolic, 320, 321, 325, 347, 348, 350, 351, 363, 367, 368, 388, 395, 402, 405 Radiant, 343, 345 surface, 364, 388 MMI, 104 modal field, 197 model classification, 188 modeling biophysical, 170 models smoke dispersion, 417 modified factor, 141 molecular sensing, 197, 204, 206 multi-beam interference, 92 multichannel sensors, 104 multicore fibers, 37 multilayer mediums, 139 structures, 428 multimode interference, 106 multi-objective optimization, 454, 457 multiplied fringe, 144 mutual information, 231, 233, 235, 242 Index N Nest, 28 Nonimaging optics, 378, 396 normalized functions, 429, 430 NRZ-OOK, 253 numerical platform, 454 simulations, 484 Nyquiest sampling theorem, 217, 225 O objective functions, 453 Optical axis, 316, 319, 339, 342, 343, 355, 361, 364, 367, 369, 373, 377-379, 405, 407 channel Channel correlation radius, 256 Correlation length, 256 Spatial coherence radius, 257 coherence tomography, 213 Coherence Tomography, 506, 507 communication, 220 efficiency, 315, 316, 324, 325, 331, 347, 348, 355, 358, 359, 364, 366, 367, 370373, 380, 381, 387, 389-392, 394, 405408 glass, 435, 436, 438, 439, 442, 445, 447449 imaging techniques, 213 manufacturing, 467, 468, 484 path, 321, 332, 357 length, 92 sensors, 103 simulations, 317, 338, 346, 365, 388, 405 spectroscopy, 220 Optics primary, 344, 346, 359 refractive, 316, 329, 344, 359, 364 secondary, 316, 344, 346 optimization, 436, 444, 453-458 oxidation-induced deterioration, 437 oxidized silicon wafer, 214 P Gamma-Gamma, 258 Log-normal, 258 phase error, 146 imaging, 502, 503 Retrieval Algorithm, 496 unwrapping, 145 Philips’ Hue Light Bulbs and Bridge, 27 Photodetector, 347-350, 365, 366, 373 photoelasticity experiment, 146 Photo-emf effect, 155, 156, 158 sensor, 158, 160-164, 166 signal, 158, 161, 162, 163 photolithography, 435 Photoplethysmography, 230 Photoresist Positive photoressit, 302 photoresist (PR), 302 pigment concentration, 231 plume concentration vertical structure of, 429 heights quantifying, 417 injection height, 418 Pointing errors. See FSO Pointing errors polishing, 435, 441, 443, 467, 468-470, 475, 482, 484 Polycarbonate, 301 Tetrabromobisphenol A, 302 polynomial fitting function, 149 prescribed fires, 417, 426, 431 pressure distribution, 469, 482, 484 Preston, 468, 469, 476, 480 Print-out effect, 301 darkening problem, 301 profile derivation, 444-446, 450 profiles bacterial colony, 175 Profilometry, 506 propagation constants, 199 proportional-integral-derivative, 444 Q Qualcomm Life’s, 27 quality factor, 120 parallel-plate waveguide, 197 PDF 517 Advances in Optics: Reviews. Book Series, Vol. 3 R Radiometer, 349 Ralph Lauren’s Polo Tech Shirt, 28 Ray-tracing, 355, 405 Reflection, 317, 318, 343, 352, 372, 388 light intensity, 92 Reflections, 318, 344, 365, 366, 372, 373, 412 Refraction, 318, 327, 329 refractive index, 198, 223 sensing Method of Measuring Fluid RI in FP Cavity, 84, 88 residual stresses, 444, 445, 448, 449, 454, 456-458 resonance wavelength, 112 shift, 113 Reticle image, 301 Ring, 27 Ronchi grating, 154 roughness, 468, 482 S Samsung SmartThings, 28 scanning lidar setup schematics of, 423 scans vertical, 426 sensitivity, 110 enhancement, 111 sensor application, 29 shoving model, 446 signal square range-corrected, 429 Silicon, 345, 415 waveguide, 104 Silver Halide AgH, 301 SIMO, 259 simplex method, 454 simulated image, 147 Single-input multiple-output. See SIMO single-output. See SISO SISO, 259 smoke plume actual height, 421 518 boundaries, 431 column, 426 concentration, 430 vertically stratified, 427, 432 smooth filtering method, 148 soaking, 437, 444, 445 Solar cell, 315, 318, 320, 329, 338, 342, 344, 346, 351, 352, 393, 394 Sparse Interferometric Fringes, 144 Spatial Coherence, 490 heterodyne spectroscopy, 221 Phase Noise Comparison, 501 spherical mirror, 141 spindle, 469-471, 473, 484 stereolithography, 205 Strain sensing Cylindrical FP Cavity Strain Sensor, 76 Prolate Spheroidal FP Cavity Strain Sensor, 71 Spherical FP Cavity Strain Sensor, 75 Strassbaugh, 470 stress-optical law, 149 structural relaxation, 445, 447, 448 surface finish quality, 446 plasmon polaritons, 198 Symmetry, 316, 325, 326, 329, 333, 339, 342, 348, 359, 385, 387, 395 absolute error limits, 137 relative error limit, 137 T Talbot distance, 153, 154, 160, 164, 166 effect, 153-155, 163, 165, 166 self images, 154 temperature fluctuations, 114 sensing Coating Type FP Fiber Optic Sensor, 97 Splicing Type FP Fiber Optic Sensor, 93 Temperature Sensitivity for the Fiber FP Strain Sensor, 83 Temporal and Spatial Phase Noise, 498 Temporal Coherence, 489 The intensity of the interference fringe, 70 Index The interference fringe contrast of the reflecting spectrum, 70 The last mile access, 247 the minimum interference intensity, 70 thermal drift, 114 expansion coefficient, 83, 84, 85, 93 shrinkage, 454 thermal-expansion effect, 95 thermo-optic coefficient, 84 THz fiber sensor, 197 hybrid plasmonic waveguide, 198 spectroscopy, 199 spoof surface plasmons, 201 surface plasmonic waves, 201 wave polarization, 199 waveguides, 197 THz-phase sensitive response, 204 tilt of the transparent window, 138 TNM model, 446, 448 transfer matrix, 105, 115 Transmission curve, 316, 317, 325, 326 transmittance amplitude, 172 Transverse Resolution, 499 Spatial Coherence, 491 spatial coherence function, 491 tungsten carbide, 440 two-way transmission, 420 U UV curable photopolymer, 205 V Van-Cittert-Zernike theorem, 491 variation of refractive index, 134 vertical scans, 424 video sequences, 230, 231, 233 viscosity, 437, 445-447 Vitality, 27 Vogel-Fulcher-Tammann, 445 W wave length, 143 waveguide sensitivity, 110 factor, 118 wheel grinding, 441 Wiener Khintchin theorem, 489 wildfires, 417-419, 423, 428, 432 locations of, 424 wrapped phase, 145 Y Young interferometer, 490 519