Hyperspectral imagingHyperspectral imaging collects and processes information from across the electromagnetic spectrum. The goal of hyperspectral imaging is to obtain the spectrum for each pixel in the image of a scene, with the purpose of finding objects, identifying materials, or detecting processes. There are three general types of spectral imagers. There are push broom scanners and the related whisk broom scanners (spatial scanning), which read images over time, band sequential scanners (spectral scanning), which acquire images of an area at different wavelengths, and snapshot hyperspectral imagers, which uses a staring array to generate an image in an instant.
SuperlensA superlens, or super lens, is a lens which uses metamaterials to go beyond the diffraction limit. The diffraction limit is a feature of conventional lenses and microscopes that limits the fineness of their resolution depending on the illumination wavelength and the numerical aperture NA of the objective lens. Many lens designs have been proposed that go beyond the diffraction limit in some way, but constraints and obstacles face each of them. In 1873 Ernst Abbe reported that conventional lenses are incapable of capturing some fine details of any given image.
MetamaterialA metamaterial (from the Greek word μετά meta, meaning "beyond" or "after", and the Latin word materia, meaning "matter" or "material") is any material engineered to have a property that is rarely observed in naturally occurring materials. They are made from assemblies of multiple elements fashioned from composite materials such as metals and plastics. These materials are usually arranged in repeating patterns, at scales that are smaller than the wavelengths of the phenomena they influence.
Plasmonic metamaterialA plasmonic metamaterial is a metamaterial that uses surface plasmons to achieve optical properties not seen in nature. Plasmons are produced from the interaction of light with metal-dielectric materials. Under specific conditions, the incident light couples with the surface plasmons to create self-sustaining, propagating electromagnetic waves known as surface plasmon polaritons (SPPs). Once launched, the SPPs ripple along the metal-dielectric interface. Compared with the incident light, the SPPs can be much shorter in wavelength.
OpticsOptics is the branch of physics that studies the behaviour and properties of light, including its interactions with matter and the construction of instruments that use or detect it. Optics usually describes the behaviour of visible, ultraviolet, and infrared light. Because light is an electromagnetic wave, other forms of electromagnetic radiation such as X-rays, microwaves, and radio waves exhibit similar properties.
History of opticsOptics began with the development of lenses by the ancient Egyptians and Mesopotamians, followed by theories on light and vision developed by ancient Greek philosophers, and the development of geometrical optics in the Greco-Roman world. The word optics is derived from the Greek term τα ὀπτικά meaning 'appearance, look'. Optics was significantly reformed by the developments in the medieval Islamic world, such as the beginnings of physical and physiological optics, and then significantly advanced in early modern Europe, where diffractive optics began.
Metamaterial absorberA metamaterial absorber is a type of metamaterial intended to efficiently absorb electromagnetic radiation such as light. Furthermore, metamaterials are an advance in materials science. Hence, those metamaterials that are designed to be absorbers offer benefits over conventional absorbers such as further miniaturization, wider adaptability, and increased effectiveness. Intended applications for the metamaterial absorber include emitters, photodetectors, sensors, spatial light modulators, infrared camouflage, wireless communication, and use in solar photovoltaics and thermophotovoltaics.
Image editingImage editing encompasses the processes of altering s, whether they are digital photographs, traditional photo-chemical photographs, or illustrations. Traditional analog image editing is known as photo retouching, using tools such as an airbrush to modify photographs or editing illustrations with any traditional art medium. Graphic software programs, which can be broadly grouped into vector graphics editors, raster graphics editors, and 3D modelers, are the primary tools with which a user may manipulate, enhance, and transform images.
Image resolutionImage resolution is the level of detail an holds. The term applies to digital images, film images, and other types of images. "Higher resolution" means more image detail. Image resolution can be measured in various ways. Resolution quantifies how close lines can be to each other and still be visibly resolved. Resolution units can be tied to physical sizes (e.g. lines per mm, lines per inch), to the overall size of a picture (lines per picture height, also known simply as lines, TV lines, or TVL), or to angular subtense.
Nonimaging opticsNonimaging optics (also called anidolic optics) is the branch of optics concerned with the optimal transfer of light radiation between a source and a target. Unlike traditional imaging optics, the techniques involved do not attempt to form an of the source; instead an optimized optical system for optimal radiative transfer from a source to a target is desired.
PixelIn digital imaging, a pixel (abbreviated px), pel, or picture element is the smallest addressable element in a raster image, or the smallest addressable element in a dot matrix display device. In most digital display devices, pixels are the smallest element that can be manipulated through software. Each pixel is a sample of an original or synthetic image; more samples typically provide more accurate representations of the original. The intensity of each pixel is variable.
Terahertz metamaterialA terahertz metamaterial is a class of composite metamaterials designed to interact at terahertz (THz) frequencies. The terahertz frequency range used in materials research is usually defined as 0.1 to 10 THz. This bandwidth is also known as the terahertz gap because it is noticeably underutilized. This is because terahertz waves are electromagnetic waves with frequencies higher than microwaves but lower than infrared radiation and visible light.
Cardinal point (optics)In Gaussian optics, the cardinal points consist of three pairs of points located on the optical axis of a rotationally symmetric, focal, optical system. These are the focal points, the principal points, and the nodal points. For ideal systems, the basic imaging properties such as image size, location, and orientation are completely determined by the locations of the cardinal points; in fact only four points are necessary: the focal points and either the principal or nodal points.
Metamaterial antennaMetamaterial antennas are a class of antennas which use metamaterials to increase performance of miniaturized (electrically small) antenna systems. Their purpose, as with any electromagnetic antenna, is to launch energy into free space. However, this class of antenna incorporates metamaterials, which are materials engineered with novel, often microscopic, structures to produce unusual physical properties. Antenna designs incorporating metamaterials can step-up the antenna's radiated power.
Spectral imagingSpectral imaging is imaging that uses multiple bands across the electromagnetic spectrum. While an ordinary camera captures light across three wavelength bands in the visible spectrum, red, green, and blue (RGB), spectral imaging encompasses a wide variety of techniques that go beyond RGB. Spectral imaging may use the infrared, the visible spectrum, the ultraviolet, x-rays, or some combination of the above.
Transformation opticsTransformation optics is a branch of optics which applies metamaterials to produce spatial variations, derived from coordinate transformations, which can direct chosen bandwidths of electromagnetic radiation. This can allow for the construction of new composite artificial devices, which probably could not exist without metamaterials and coordinate transformation. Computing power that became available in the late 1990s enables prescribed quantitative values for the permittivity and permeability, the constitutive parameters, which produce localized spatial variations.
Active-pixel sensorAn active-pixel sensor (APS) is an , which was invented by Peter J.W. Noble in 1968, where each pixel sensor unit cell has a photodetector (typically a pinned photodiode) and one or more active transistors. In a metal–oxide–semiconductor (MOS) active-pixel sensor, MOS field-effect transistors (MOSFETs) are used as amplifiers. There are different types of APS, including the early NMOS APS and the now much more common complementary MOS (CMOS) APS, also known as the CMOS sensor.
Multispectral imagingMultispectral imaging captures image data within specific wavelength ranges across the electromagnetic spectrum. The wavelengths may be separated by filters or detected with the use of instruments that are sensitive to particular wavelengths, including light from frequencies beyond the visible light range, i.e. infrared and ultra-violet. It can allow extraction of additional information the human eye fails to capture with its visible receptors for red, green and blue.
Image sensorAn image sensor or imager is a sensor that detects and conveys information used to form an . It does so by converting the variable attenuation of light waves (as they pass through or reflect off objects) into signals, small bursts of current that convey the information. The waves can be light or other electromagnetic radiation. Image sensors are used in electronic imaging devices of both analog and digital types, which include digital cameras, camera modules, camera phones, optical mouse devices, medical imaging equipment, night vision equipment such as thermal imaging devices, radar, sonar, and others.
Ray (optics)In optics, a ray is an idealized geometrical model of light or other electromagnetic radiation, obtained by choosing a curve that is perpendicular to the wavefronts of the actual light, and that points in the direction of energy flow. Rays are used to model the propagation of light through an optical system, by dividing the real light field up into discrete rays that can be computationally propagated through the system by the techniques of ray tracing. This allows even very complex optical systems to be analyzed mathematically or simulated by computer.