GMAT 9600 Principles of Remote Sensing Dr. Linlin GE Senior Lecturer Satellite Navigation And Positioning Group School of Surveying & Spatial Information Systems University of New South Wales Phone: +61-2-9385 4177 Fax: +61-2-9313 7493 Email: l.ge@unsw.edu.au http://www.gmat.unsw.edu.au/linlinge C-band radar coverage Towns are white. Lakes are dark.
What is Remote Sensing (RS)? formal and comprehensive definition The acquisition and measurement of data/information on some property(ies) of a phenomenon, object, or material by a recording device not in physical, intimate contact with the feature(s) under surveillance; techniques involve amassing knowledge pertinent to environments by measuring force fields, electromagnetic radiation, or acoustic energy employing cameras, radiometers and scanners, lasers, radio frequency receivers, radar systems, sonar, thermal devices, seismographs, magnetometers, gravimeters, scintillometers, and other instruments. What is Remote Sensing (RS)? Remote Sensing involves gathering data and information about the physical "world" by detecting and measuring radiation, particles, and fields associated with objects located beyond the immediate vicinity of the sensor device(s). What is Remote Sensing (RS)? Remote Sensing is a technology for sampling electromagnetic radiation to acquire and interpret non-immediate geospatial data from which to extract information about features, objects, and classes on the Earth's land surface, oceans, and atmosphere (and, where applicable, on the exteriors of other bodies in the solar system, or, in the broadest framework, celestial bodies such as stars and galaxies).
RS: the History (Photographic) 1759 First statements by Lambert (France) of principles underlying photogrammetry 1839 First ever photographs by Daguerre and Nepce, in France 1840 French used photos in making topographic maps. 1850's Photographs important in documenting exploration of the U.S. West (through 1870's). 1855 Maxwell proposed proof of trichromatic color vision by photographic experiments (1861; Sutton). 1858 Pictures of Paris from cameras mounted in free and captive balloons. 1880's Camera airborne on kites in England, France, Russia. 1895 First color separations produced. 1900 Ives invented three-lens multispectral 1909 Wilbur Wright took first photos (movies) from an airplane. RS: the History (Photographic con.) 1920 Aerial photos used by petroleum geologists for exploration. 1930 First aerial spectrophotography of the Earth by Krinov and colleagues (Russia). 1930's Extensive use of aerial photos in Earth sciences and agriculture. 1931 Testing of aerial IR sensitive film from stratospheric 1943 Rapid advances in development of black-and-white and color IR (CIR) film for camouflage detection and haze penetration. 1950's Term "Remote Sensing" first used (generally ascribed to Evelyn Pruitt). 1962 United States and Russian nine-lens multispectral cameras; Itek camera (ten lens) in 1963. 1964 NASA inaugurated programs in testing usefulness of multiband photography for Earth resources. 1965 Multispectral additive color system developed by Yost and Wenderoth. 1967First practical uses of UV photography. 1968 multispectral photography experiment on Apollo-9 RS: the History (non-photographic) 1800 Discovery of the IR spectral region by Sir William Herschel. 1879 Use of the bolometer by Langley to make temperature measurements of electrical objects. 1889 Hertz demonstrated reflection of radio waves from solid objects. 1916 Aircraft tracked in flight by Hoffman using thermopiles to detect heat effects. 1930 Both British and Germans work on systems to locate airplanes from their thermal patterns at night. RS: the History (non-photographic con.) 1940 Development of incoherent radar systems by the British and United States to detect and track aircraft and ships during W.W.II. 1950's Extensive studies of IR systems at University of Michigan and elsewhere. 1951 First concepts of a moving coherent radar system. 1953 Flight of an X-band coherent radar. 1954 Formulation of synthetic aperture concept (SAR) in radar. 1960's Development of various detectors which allowed building of imaging and non-imaging radiometers, scanners, spectrometers and polarimeters.
RS: the History (Space Imaging Systems) 1891 First proposal (Rahrmann) for using a rocket as a photo platform. 1908 Maul (Germany) develops gyro-stabilized camera mounted on rocket (launched by 1912). 1946 Space pictures obtained from V-2 rockets launched at White Sands Proving Ground (New Mexico). 1957 Launch of Sputnik 1 by USSR. 1960 Images obtained from U.S. meteorological satellite TIROS-1. 1961 Orbital photographs from unmanned Mercury spacecraft MA - 4, followed by astronaut photographs from MA -8 and MA -9. 1964 Nimbus research meteorological satellite program begins; TV and other sensors. 1965 First manned Gemini flight (GT-3), with some color photos. 1965 Gemini GT-4 space photography experiments. RS: the History (Space Imaging Systems con.) 1965 Recommendation of ERTS (Landsat) program by U.S. Department of Interior to NASA. 1966 Launch of ATS series of geosynchronous satellites, with imaging sensors, followed by SMS (GOES) series beginning in 1974 1967 Apollo mission Earth-orbital flights (Apollo-6, -7) culminating in SO65 multispectral photography experiment on Apollo-9 (1968) 1972 Launch of ERTS-1 (Landsat); Landsat-2 (1975); Landsat-3 (1978) 1972-7373 Skylab launch; series of experiments by astronauts with EREP (Earth Resources Experiment Package). 1975 Apollo-Soyuz flight; some photography. 1978 Seasat-1 launched in June (failed after 99 days). 1978 Heat Capacity Mapping Mission (HCMM), first AEM. 1979 Nimbus-7 launched. EM vs eye As a species, we've been literally blind to the universe around us. If the known electromagnetic spectrum were scaled up to stretch around the Earth's circumference, the human eye would see a portion equal to the diameter of a pencil. Our ability to build detectors that see for us where we can't see, and computers that bring the invisible information back to our eyesight, will ultimately contribute to our survival on Earth and in space. m Wavelength Micrometer (µm) = 10-6 m Solar radiation: 0.1 4 µm (ultraviolet, visible, infrared) visible: 0.4 0.7 µm Nanometer (nm) = 10-9 m
(Lillesand and Kiefer, 2000, Remote sensing and image interpretation.) Black and white True color photograph Blue: blue Green: green Red: red Photograph Standard false color photograph Blue: green response Green: red response Red: near infrared response (NIR) JERS-1 Principal Component / NDVI Composite SPOT Principal Component Image NASDA 1995 CNES 1995
NOAA AVHRR NDVI Landsat MSS (Multi--Spectral Scanner ) (Multi CSIRO Division of Oceanography 7
Landsat TM (Thematic Mapper ) Image MSU Channel 2 Trend Map Comparison RSS UAH RSS UAH Remote Sensing Instruments Aerial Photography Color Infrared Film (CIR) Thermal Infrared Multispectral Scanner (TIMS) Airborne Oceanographic Lidar (AOL) Synthetic Aperture Radar (SAR) Microwave Radar
RS applications Geologic Applications - Stratigraphy, Structure, Mineral & Petroleum Exploration Vegetation Applications - Agriculture, Forestry, and Ecology Urban and Land Use Applications Disaster Prevention and Monitoring Environmental Monitoring Seasat Image Central Jamaica in the Caribbean obtained through cloud cover on August 8, 1978 "The application of Remote Sensing is only limited by the imagination of those who use it." Topography Thermography: Night Vision Systems
Animation: Multi-pass DInSAR results Mining Subsidence The West Cliff D-Line!