Search This Blog

Saturday, August 6, 2022

Digital camera

From Wikipedia, the free encyclopedia

Front and back of Canon PowerShot A95 (c.2004), a once typical pocket-sized compact camera, with mode dial, optical viewfinder, and articulating screen.
 
Hasselblad 503CW with Ixpress V96C digital back, an example of a professional digital camera system

A digital camera is a camera that captures photographs in digital memory. Most cameras produced today are digital, largely replacing those that capture images on photographic film. Digital cameras are now widely incorporated into mobile devices like smartphones with the same or more capabilities and features of dedicated cameras (which are still avaliable). High-end, high-definition dedicated cameras are still commonly used by professionals and those who desire to take higher-quality photographs.

Digital and digital movie cameras share an optical system, typically using a lens with a variable diaphragm to focus light onto an image pickup device. The diaphragm and shutter admit a controlled amount of light to the image, just as with film, but the image pickup device is electronic rather than chemical. However, unlike film cameras, digital cameras can display images on a screen immediately after being recorded, and store and delete images from memory. Many digital cameras can also record moving videos with sound. Some digital cameras can crop and stitch pictures and perform other elementary image editing.

History

The first semiconductor image sensor was the charge-coupled device (CCD), invented by Willard S. Boyle and George E. Smith at Bell Labs in 1969, based on MOS capacitor technology. The NMOS active-pixel sensor was later invented by Tsutomu Nakamura's team at Olympus in 1985, which led to the development of the CMOS active-pixel sensor (CMOS sensor) by Eric Fossum's team at the NASA Jet Propulsion Laboratory in 1993.

In the 1960s, Eugene F. Lally of the Jet Propulsion Laboratory was thinking about how to use a mosaic photosensor to capture digital images. His idea was to take pictures of the planets and stars while travelling through space to give information about the astronauts' position. As with Texas Instruments employee Willis Adcock's film-less camera (US patent 4,057,830) in 1972, the technology had yet to catch up with the concept.

The Cromemco Cyclops was an all-digital camera introduced as a commercial product in 1975. Its design was published as a hobbyist construction project in the February 1975 issue of Popular Electronics magazine. It used a 32×32 metal-oxide-semiconductor (MOS) image sensor, which was a modified MOS dynamic RAM (DRAM) memory chip.

Steven Sasson, an engineer at Eastman Kodak, invented and built a self-contained electronic camera that used a CCD image sensor in 1975. Around the same time, Fujifilm began developing CCD technology in the 1970s. Early uses were mainly military and scientific; followed by medical and news applications.

Nikon was interested in digital photography since the mid-1980s. In 1986, while presenting to Photokina, Nikon introduced an operational prototype of the first SLR-type electronic camera (Still Video Camera), manufactured by Panasonic. The Nikon SVC was built around a sensor 2/3″ charge-coupled device of 300,000 pixels. Storage media, a magnetic floppy disk inside the camera allows recording 25 or 50 B&W images, depending on the definition. In 1988, Nikon released the first commercial electronic single-lens reflex camera, the QV-1000C.

At Photokina 1988, Fujifilm introduced the FUJIX DS-1P, the first fully digital camera, capable of saving data to a semiconductor memory card. The camera's memory card had a capacity of 2 MB of SRAM (static random-access memory), and could hold up to ten photographs. In 1989, Fujifilm released the FUJIX DS-X, the first fully digital camera to be commercially released. In 1996, Toshiba's 40 MB flash memory card was adopted for several digital cameras.

The first commercial camera phone was the Kyocera Visual Phone VP-210, released in Japan in May 1999. It was called a "mobile videophone" at the time, and had a 110,000-pixel front-facing camera. It stored up to 20 JPEG digital images, which could be sent over e-mail, or the phone could send up to two images per second over Japan's Personal Handy-phone System (PHS) cellular network. The Samsung SCH-V200, released in South Korea in June 2000, was also one of the first phones with a built-in camera. It had a TFT liquid-crystal display (LCD) and stored up to 20 digital photos at 350,000-pixel resolution. However, it could not send the resulting image over the telephone function, but required a computer connection to access photos. The first mass-market camera phone was the J-SH04, a Sharp J-Phone model sold in Japan in November 2000. It could instantly transmit pictures via cell phone telecommunication. By the mid-2000s, higher-end cell phones had an integrated digital camera and by the early 2010s, almost all smartphones had an integrated digital camera.

Image sensors

The two major types of digital image sensor are CCD and CMOS. A CCD sensor has one amplifier for all the pixels, while each pixel in a CMOS active-pixel sensor has its own amplifier. Compared to CCDs, CMOS sensors use less power. Cameras with a small sensor use a back-side-illuminated CMOS (BSI-CMOS) sensor. The image processing capabilities of the camera determine the outcome of the final image quality much more than the sensor type.

Sensor resolution

The resolution of a digital camera is often limited by the image sensor that turns light into discrete signals. The brighter the image at a given point on the sensor, the larger the value that is read for that pixel. Depending on the physical structure of the sensor, a color filter array may be used, which requires demosaicing to recreate a full-color image. The number of pixels in the sensor determines the camera's "pixel count". In a typical sensor, the pixel count is the product of the number of rows and the number of columns. For example, a 1,000 by 1,000 pixel sensor would have 1,000,000 pixels, or 1 megapixel.

Resolution options

Firmwares' resolution selector allows the user to optionally lower the resolution, to reduce the file size per picture and extend lossless digital zooming. The bottom resolution option is typically 640×480 pixels (0.3 Megapixels).

A lower resolution extends the number of remaining photos in free space, postponing the exhaustion of space storage, which is of use where no further data storage device is available, and for captures of lower significance, where the benefit from less space storage consumption outweighs the disadvantage from reduced detail.

Image sharpness

The final quality of an image depends on all optical transformations in the chain of producing the image. Carl Zeiss, a German optician, points out that the weakest link in an optical chain determines the final image quality. In the case of the digital camera, a simple way to describe this concept is that the lens determines the maximum sharpness of the image while the image sensor determines the maximum resolution. The illustration on the right can be said to compare a lens with very poor sharpness on a camera with high resolution, to a lens with good sharpness on a camera with lower resolution.

Methods of image capture

At the heart of a digital camera is a CCD or a CMOS image sensor.
 
Digital camera, partly disassembled. The lens assembly (bottom right) is partially removed, but the sensor (top right) still captures an image, as seen on the LCD screen (bottom left).

Since the first digital backs were introduced, there have been three main methods of capturing the image, each based on the hardware configuration of the sensor and color filters.

Single-shot capture systems use either one sensor chip with a Bayer filter mosaic, or three separate image sensors (one each for the primary additive colors red, green, and blue) which are exposed to the same image via a beam splitter (see Three-CCD camera).

Multi-shot exposes the sensor to the image in a sequence of three or more openings of the lens aperture. There are several methods of application of the multi-shot technique. The most common was originally to use a single image sensor with three filters passed in front of the sensor in sequence to obtain the additive color information. Another multiple shot method is called Microscanning. This method uses a single sensor chip with a Bayer filter and physically moves the sensor on the focus plane of the lens to construct a higher resolution image than the native resolution of the chip. A third version combines these two methods without a Bayer filter on the chip.

The third method is called scanning because the sensor moves across the focal plane much like the sensor of an image scanner. The linear or tri-linear sensors in scanning cameras utilize only a single line of photosensors, or three lines for the three colors. Scanning may be accomplished by moving the sensor (for example, when using color co-site sampling) or by rotating the whole camera. A digital rotating line camera offers images consisting of a total resolution that is very high.

The choice of method for a given capture is determined largely by the subject matter. It is usually inappropriate to attempt to capture a subject that moves with anything but a single-shot system. However, the higher color fidelity and larger file sizes and resolutions that are available with multi-shot and scanning backs make them more attractive for commercial photographers who are working with stationary subjects and large-format photographs.

Improvements in single-shot cameras and image file processing at the beginning of the 21st century made single shot cameras almost completely dominant, even in high-end commercial photography.

Filter mosaics, interpolation, and aliasing

The Bayer arrangement of color filters on the pixel array of an image sensor.

Most current consumer digital cameras use a Bayer filter mosaic in combination with an optical anti-aliasing filter to reduce the aliasing due to the reduced sampling of the different primary-color images. A demosaicing algorithm is used to interpolate color information to create a full array of RGB image data.

Cameras that use a beam-splitter single-shot 3CCD approach, three-filter multi-shot approach, color co-site sampling or Foveon X3 sensor do not use anti-aliasing filters, nor demosaicing.

Firmware in the camera, or a software in a raw converter program such as Adobe Camera Raw, interprets the raw data from the sensor to obtain a full color image, because the RGB color model requires three intensity values for each pixel: one each for the red, green, and blue (other color models, when used, also require three or more values per pixel). A single sensor element cannot simultaneously record these three intensities, and so a color filter array (CFA) must be used to selectively filter a particular color for each pixel.

The Bayer filter pattern is a repeating 2x2 mosaic pattern of light filters, with green ones at opposite corners and red and blue in the other two positions. The high proportion of green takes advantage of properties of the human visual system, which determines brightness mostly from green and is far more sensitive to brightness than to hue or saturation. Sometimes a 4-color filter pattern is used, often involving two different hues of green. This provides potentially more accurate color, but requires a slightly more complicated interpolation process.

The color intensity values not captured for each pixel can be interpolated from the values of adjacent pixels which represent the color being calculated.

Sensor size and angle of view

Cameras with digital image sensors that are smaller than the typical 35 mm film size have a smaller field or angle of view when used with a lens of the same focal length. This is because angle of view is a function of both focal length and the sensor or film size used.

Kids 50mm 100mm.jpg

The crop factor is relative to the 35mm film format. If a smaller sensor is used, as in most digicams, the field of view is cropped by the sensor to smaller than the 35 mm full-frame format's field of view. This narrowing of the field of view may be described as crop factor, a factor by which a longer focal length lens would be needed to get the same field of view on a 35 mm film camera. Full-frame digital SLRs utilize a sensor of the same size as a frame of 35 mm film.

Common values for field of view crop in DSLRs using active pixel sensors include 1.3x for some Canon (APS-H) sensors, 1.5x for Sony APS-C sensors used by Nikon, Pentax and Konica Minolta and for Fujifilm sensors, 1.6 (APS-C) for most Canon sensors, ~1.7x for Sigma's Foveon sensors and 2x for Kodak and Panasonic 4/3-inch sensors currently used by Olympus and Panasonic. Crop factors for non-SLR consumer compact and bridge cameras are larger, frequently 4x or more.

Relative sizes of sensors used in most current digital cameras.
 
Table of sensor sizes
Type Width (mm) Height (mm) Size (mm²)
1/3.6" 4.00 3.00 12.0
1/3.2" 4.54 3.42 15.5
1/3" 4.80 3.60 17.3
1/2.7" 5.37 4.04 21.7
1/2.5" 5.76 4.29 24.7
1/2.3" 6.16 4.62 28.5
1/2" 6.40 4.80 30.7
1/1.8" 7.18 5.32 38.2
1/1.7" 7.60 5.70 43.3
2/3" 8.80 6.60 58.1
1" 12.8 9.6 123
4/3" 18.0 13.5 243
APS-C 25.1 16.7 419
35 mm 36 24 864
Back 48 36 1728

Types of digital cameras

Digital cameras come in a wide range of sizes, prices and capabilities. In addition to general purpose digital cameras, specialized cameras including multispectral imaging equipment and astrographs are used for scientific, military, medical and other special purposes.

Compacts

DSC-W170 is a subcompact camera with lens assembly retracted
 
Disassembled compact digital camera

Compact cameras are intended to be portable (pocketable) and are particularly suitable for casual "snapshots".

Many incorporate a retractable lens assembly that provides optical zoom. In most models, an auto actuating lens cover protects the lens from elements. Most ruggedized or water-resistant models do not retract, and most with superzoom capability do not retract fully.

Compact cameras are usually designed to be easy to use. Almost all include an automatic mode, or "auto mode", which automatically makes all camera settings for the user. Some also have manual controls. Compact digital cameras typically contain a small sensor which trades-off picture quality for compactness and simplicity; images can usually only be stored using lossy compression (JPEG). Most have a built-in flash usually of low power, sufficient for nearby subjects. A few high end compact digital cameras have a hotshoe for connecting to an external flash. Live preview is almost always used to frame the photo on an integrated LCD. In addition to being able to take still photographs almost all compact cameras have the ability to record video.

Compacts often have macro capability and zoom lenses, but the zoom range (up to 30x) is generally enough for candid photography but less than is available on bridge cameras (more than 60x), or the interchangeable lenses of DSLR cameras available at a much higher cost. Autofocus systems in compact digital cameras generally are based on a contrast-detection methodology using the image data from the live preview feed of the main imager. Some compact digital cameras use a hybrid autofocus system similar to what is commonly available on DSLRs.

Typically, compact digital cameras incorporate a nearly silent leaf shutter into the lens but play a simulated camera sound for skeuomorphic purposes.

For low cost and small size, these cameras typically use image sensor formats with a diagonal between 6 and 11 mm, corresponding to a crop factor between 7 and 4. This gives them weaker low-light performance, greater depth of field, generally closer focusing ability, and smaller components than cameras using larger sensors. Some cameras use a larger sensor including, at the high end, a pricey full-frame sensor compact camera, such as Sony Cyber-shot DSC-RX1, but have capability near that of a DSLR.

A variety of additional features are available depending on the model of the camera. Such features include GPS, compass, barometers and altimeters.

Starting in 2011, some compact digital cameras can take 3D still photos. These 3D compact stereo cameras can capture 3D panoramic photos with dual lens or even single lens for play back on a 3D TV.

In 2013, Sony released two add-on camera models without display, to be used with a smartphone or tablet, controlled by a mobile application via WiFi.

Rugged compacts

Rugged compact cameras typically include protection against submersion, hot and cold conditions, shock and pressure. Terms used to describe such properties include waterproof, freezeproof, heatproof, shockproof and crushproof, respectively. Nearly all major camera manufacturers have at least one product in this category. Some are waterproof to a considerable depth up to 100 feet (30 m); others only 10 feet (3 m), but only a few will float. Ruggeds often lack some of the features of ordinary compact camera, but they have video capability and the majority can record sound. Most have image stabilization and built-in flash. Touchscreen LCD and GPS do not work under water.

Action cameras

GoPro and other brands offer action cameras which are rugged, small and can be easily attached to helmet, arm, bicycle, etc. Most have wide angle and fixed focus, and can take still pictures and video, typically with sound.

360-degree cameras

The 360-degree camera can take picture or video 360 degrees using two lenses back-to-back and shooting at the same time. Some of the cameras are Ricoh Theta S, Nikon Keymission 360 and Samsung Gear 360. Nico360 was launched in 2016 and claimed as the world's smallest 360-degree camera with size 46 x 46 x 28 mm (1.8 x 1.8 x 1.1 in) and price less than $200. With virtual reality mode built-in stitching, Wifi, and Bluetooth, live streaming can be done. Due to it also being water resistant, the Nico360 can be used as action camera.

There are tend that action cameras have capabilities to shoot 360 degrees with at least 4K resolution.

Bridge cameras

Sony DSC-H2
 

Bridge cameras physically resemble DSLRs, and are sometimes called DSLR-shape or DSLR-like. They provide some similar features but, like compacts, they use a fixed lens and a small sensor. Some compact cameras have also PSAM mode. Most use live preview to frame the image. Their usual autofocus is by the same contrast-detect mechanism as compacts, but many bridge cameras have a manual focus mode and some have a separate focus ring for greater control.

Big physical size and small sensor allow superzoom and wide aperture. Bridge cameras generally include an image stabilization system to enable longer handheld exposures, sometimes better than DSLR for low light condition.

As of 2014, bridge cameras come in two principal classes in terms of sensor size, firstly the more traditional 1/2.3" sensor (as measured by image sensor format) which gives more flexibility in lens design and allows for handholdable zoom from 20 to 24 mm (35 mm equivalent) wide angle all the way up to over 1000 mm supertele, and secondly a 1" sensor that allows better image quality particularly in low light (higher ISO) but puts greater constraints on lens design, resulting in zoom lenses that stop at 200 mm (constant aperture, e.g. Sony RX10) or 400 mm (variable aperture, e.g. Panasonic Lumix FZ1000) equivalent, corresponding to an optical zoom factor of roughly 10 to 15.

Some bridge cameras have a lens thread to attach accessories such as wide-angle or telephoto converters as well as filters such as UV or Circular Polarizing filter and lens hoods. The scene is composed by viewing the display or the electronic viewfinder (EVF). Most have a slightly longer shutter lag than a DSLR. Many of these cameras can store images in a raw format in addition to supporting JPEG. The majority have a built-in flash, but only a few have a hotshoe.

In bright sun, the quality difference between a good compact camera and a digital SLR is minimal but bridge cameras are more portable, cost less and have a greater zoom ability. Thus a bridge camera may better suit outdoor daytime activities, except when seeking professional-quality photos.

Mirrorless interchangeable-lens cameras

Olympus OM-D E-M1 Mark II introduced 2016
 
Nikon Z7 introduced 2018
 

In late 2008, a new type of camera emerged, called a mirrorless interchangeable-lens camera. It is technically a DSLR camera that does not require a reflex mirror, a key component of the former. While a typical DSLR has a mirror that reflects light from the lens up to the optical viewfinder, in a mirrorless camera, there is no optical viewfinder. The image sensor is exposed to light at all times, giving the user a digital preview of the image either on the built-in rear LCD screen or an electronic viewfinder (EVF).

These are simpler and more compact than DSLRs due to not having a lens reflex system. MILCs, or mirrorless cameras for short, come with various sensor sizes depending on the brand and manufacturer, these include: a small 1/2.3 inch sensor, as is commonly used in bridge cameras such as the original Pentax Q (more recent Pentax Q versions have a slightly larger 1/1.7 inch sensor); a 1-inch sensor; a Micro Four Thirds sensor; an APS-C sensor found in Sony NEX series and α "DSLR-likes", Fujifilm X series, Pentax K-01, and Canon EOS M; and some, such as the Sony α7, use a full frame (35 mm) sensor, with the Hasselblad X1D being the first medium format mirrorless camera. Some MILCs have a separate electronic viewfinder to compensate the lack of an optical one. In other cameras, the back display is used as the primary viewfinder in the same way as in compact cameras. One disadvantage of mirrorless cameras compared to a typical DSLR is its battery life due to the energy consumption of the electronic viewfinder, but this can be mitigated by a setting inside the camera in some models.

Olympus and Panasonic released many Micro Four Thirds cameras with interchangeable lenses that are fully compatible with each other without any adapter, while others have proprietary mounts. In 2014, Kodak released its first Micro Four Third system camera.

As of March 2014, mirrorless cameras are fast becoming appealing to both amateurs and professionals alike due to their simplicity, compatibility with some DSLR lenses, and features that match most DSLRs today.

Modular cameras

Sony Alpha ILCE-QX1, an example of a modular, lens-style camera, introduced in 2014

While most digital cameras with interchangeable lenses feature a lens-mount of some kind, there are also a number of modular cameras, where the shutter and sensor are incorporated into the lens module.

The first such modular camera was the Minolta Dimâge V in 1996, followed by the Minolta Dimâge EX 1500 in 1998 and the Minolta MetaFlash 3D 1500 in 1999. In 2009, Ricoh released the Ricoh GXR modular camera.

At CES 2013, Sakar International announced the Polaroid iM1836, an 18 MP camera with 1"-sensor with interchangeable sensor-lens. An adapter for Micro Four Thirds, Nikon and K-mount lenses was planned to ship with the camera.

There are also a number of add-on camera modules for smartphones, they are called lens-style cameras (lens camera or smart lens). They contain all the essential components of a digital camera inside a DSLR lens-shaped module, hence the name, but lack any sort of viewfinder and most controls of a regular camera. Instead, they are connected wirelessly and/or mounted to a smartphone to be used as its display output and operate the camera's various controls.

Lens-style cameras include:

  • Sony Cyber-shot QX series "Smart Lens" or "SmartShot" cameras, announced and released in mid 2013 with the Cyber-shot DSC-QX10. In January 2014, a firmware update was announced for the DSC-QX10 and DSC-QX100. In September 2014, Sony announced the Cyber-shot DSC-QX30 as well as the Alpha ILCE-QX1, the former an ultrazoom with a built-in 30x optical zoom lens, the latter opting for an interchangeable Sony E-mount instead of a built-in lens.
  • Kodak PixPro smart lens camera series, announced in 2014. These include: the 5X optical zoom SL5, 10X optical zoom SL10, and the 25X optical zoom SL25; all featuring 16 MP sensors and 1080p video recording, except for the SL5 which caps at 720p.
  • ViviCam IU680 smart lens camera from Sakar-owned brand, Vivitar, announced in 2014.
  • Olympus Air A01 lens camera, announced in 2014 and released in 2015, the lens camera is an open platform with an Android operating system and can detach into 2 parts (sensor module and lens), just like the Sony QX1, and all compatible Micro Four Thirds lenses can then be attached to the built-in lens mount of the camera's sensor module.

Digital single-lens reflex cameras (DSLR)

Cutaway of an Olympus E-30 DSLR
 

Digital single-lens reflex cameras (DSLR) is a camera with a digital sensor that utilizes a reflex mirror to split or direct light into the viewfinder to produce an image. The reflex mirror finds the image by blocking light to the camera's sensor and then reflecting it into the camera's pentaprism which allows it to be seen through the viewfinder. When the shutter release is fully pressed the reflex mirror pulls out horizontally below the pentaprism briefly darkening the viewfinder and then opening up the sensor for exposure which creates the photo. The digital image is produced by the sensor which is an array of photoreceptors on a microchip capable of recording light values. Many modern DSLRs offer the ability for "live view" or the framing of the subject emitted from the sensor onto a digital screen.

The sensor also known as a full-frame sensor is much larger than the other types, typically 18mm to 36mm on the diagonal (crop factor 2, 1.6, or 1). The larger sensor permits more light to be received by each pixel; this, combined with the relatively large lenses provides superior low-light performance. For the same field of view and the same aperture, a larger sensor gives shallower focus. DSLRs can equip interchangeable lenses for versatility by removing it from the lens mount of the camera, typically a silver ring on the front side of DSLRs. These lenses work in tandem with the mechanics of the DSLR to adjust aperture and focus. Autofocus is accomplished using sensors in the mirror box and on most modern lenses can be activated from the lens itself which will trigger upon shutter release.

Digital Still Cameras (DSC)

Digital Still Camera (DSC), such as the Sony DSC cameras, is a type of camera that doesn't use a reflex mirror. DSCs are like point-and-shoot cameras and is, the most common type of cameras, due to its comfortable price and its quality.

Here are a list of DSCs: List of Sony Cyber-shot cameras

Fixed-mirror DSLT cameras

Cameras with fixed semi-transparent mirrors, also known as DSLT cameras, such as the Sony SLT cameras, are single-lens without a moving reflex mirror as in a conventional DSLR. A semi-transparent mirror transmits some of the light to the image sensor and reflects some of the light along the path to a pentaprism/pentamirror which then goes to an optical view finder (OVF) as is done with a reflex mirror in DSLR cameras. The total amount of light is not changed, just some of the light travels one path and some of it travels the other. The consequences are that DSLT cameras should shoot a half stop differently from DSLR. One advantage of using a DSLT camera is the blind moments a DSLR user experiences while the reflecting mirror is moved to send the light to the sensor instead of the viewfinder do not exist for DSLT cameras. Because there is no time at which light is not traveling along both paths, DSLT cameras get the benefit of continuous auto-focus tracking. This is especially beneficial for burst-mode shooting in low-light conditions and also for tracking when taking video.

Digital rangefinders

A rangefinder is a device to measure subject distance, with the intent to adjust the focus of a camera's objective lens accordingly (open-loop controller). The rangefinder and lens focusing mechanism may or may not be coupled. In common parlance, the term "rangefinder camera" is interpreted very narrowly to denote manual-focus cameras with a visually-read out optical rangefinder based on parallax. Most digital cameras achieve focus through analysis of the image captured by the objective lens and distance estimation, if it is provided at all, is only a byproduct of the focusing process (closed-loop controller).

Line-scan camera systems

A San Francisco cable car, imaged using an Alkeria Necta N4K2-7C line scan camera with a shutter speed of 250 microseconds, or 4000 frames per second.

A line-scan camera traditionally has a single row of pixel sensors, instead of a matrix of them. The lines are continuously fed to a computer that joins them to each other and makes an image. This is most commonly done by connecting the camera output to a frame grabber which resides in a PCI slot of an industrial computer. The frame grabber acts to buffer the image and sometimes provide some processing before delivering to the computer software for processing. Industrial processes often require height and width measurements performed by digital line-scan systems.

Multiple rows of sensors may be used to make colored images, or to increase sensitivity by TDI (time delay and integration).

Many industrial applications require a wide field of view. Traditionally maintaining consistent light over large 2D areas is quite difficult. With a line scan camera all that is necessary is to provide even illumination across the “line” currently being viewed by the camera. This makes sharp pictures of objects that pass the camera at high speed.

Such cameras are also commonly used to make photo finishes, to determine the winner when multiple competitors cross the finishing line at nearly the same time. They can also be used as industrial instruments for analyzing fast processes.

Line-scan cameras are also extensively used in imaging from satellites (see push broom scanner). In this case the row of sensors is perpendicular to the direction of satellite motion. Line-scan cameras are widely used in scanners. In this case, the camera moves horizontally.

Stand alone camera

Stand alone cameras can be used as remote camera. One kind weighs 2.31 ounces (65.5 g), with a periscope shape, IPx7 water-resistance and dust-resistance rating and can be enhanced to IPx8 by using a cap. They have no viewfinder or LCD. Lens is a 146 degree wide angle or standard lens, with fixed focus. It can have a microphone and speaker, And it can take photos and video. As a remote camera, a phone app using Android or iOS is needed to send live video, change settings, take photos, or use time lapse.

Superzoom Cameras

Digital superzoom cameras are digital cameras that can zoom in very far. These superzoom cameras are suitable for people who have nearsightedness.

The HX series is a series containing Sony's superzoom cameras like HX20V, HX90V and the newest HX99. HX stands for HyperXoom.

Light-field camera

This type of digital camera captures information about the light field emanating from a scene; that is, the intensity of light in a scene, and also the direction that the light rays are traveling in space. This contrasts with a conventional digital camera, which records only light intensity.

Integration into other devices

Many devices have a built-in digital camera, including, for example, smartphones, mobile phones, PDAs and laptop computers. Built-in cameras generally store the images in the JPEG file format.

Mobile phones incorporating digital cameras were introduced in Japan in 2001 by J-Phone. In 2003 camera phones outsold stand-alone digital cameras, and in 2006 they outsold film and digital stand-alone cameras. Five billion camera phones were sold in five years, and by 2007 more than half of the installed base of all mobile phones were camera phones. Sales of separate cameras peaked in 2008.

Notable digital camera manufacturers

There are many manufacturers that lead in the production of digital cameras (commonly DSLRs). Each brand embodies different mission statements that differ them from each other outside of the physical technology that they produce. While majority manufacturer share modern features amongst their production of cameras, some specialize in specific details either physically on camera or within the system and image quality.

A Nikon D200 camera with a Nikon 17-55 mm / 2,8 G AF-S DX IF-ED lens and a Nikon SB-800 flash. Flashes are used as attachment to a camera to provide light to the image, timed with the shutter of the camera.
 
Canon EF 70-200 f/2.8 lens mounted on a Canon 7D camera body. Lenses of varying lengths can be equipped onto main camera bodies to provide different perspectives for an image taken.

Market trends

Chart of sale of smartphones (with built-in cameras) compared to digital cameras 2009–2013 showing smartphone sale soaring while camera sale is stagnating
Sale of smartphones compared to digital cameras 2009–2013

Sales of traditional digital cameras have declined due to the increasing use of smartphones for casual photography, which also enable easier manipulation and sharing of photos through the use of apps and web-based services. "Bridge cameras", in contrast, have held their ground with functionality that most smartphone cameras lack, such as optical zoom and other advanced features. DSLRs have also lost ground to Mirrorless interchangeable-lens camera (MILC)s offering the same sensor size in a smaller camera. A few expensive ones use a full-frame sensor, just like DSLR professional cameras.

In response to the convenience and flexibility of smartphone cameras, some manufacturers produced "smart" digital cameras that combine features of traditional cameras with those of a smartphone. In 2012, Nikon and Samsung released the Coolpix S800c and Galaxy Camera, the first two digital cameras to run the Android operating system. Since this software platform is used in many smartphones, they can integrate with some of the same services (such as e-mail attachments, social networks and photo sharing sites) that smartphones do and use other Android-compatible software.

In an inversion, some phone makers have introduced smartphones with cameras designed to resemble traditional digital cameras. Nokia released the 808 PureView and Lumia 1020 in 2012 and 2013; the two devices respectively run the Symbian and Windows Phone operating systems, and both include a 41-megapixel camera (along with a camera grip attachment for the latter). Similarly, Samsung introduced the Galaxy S4 Zoom, having a 16-megapixel camera and 10x optical zoom, combining traits from the Galaxy S4 Mini with the Galaxy Camera. Panasonic Lumix DMC-CM1 is an Android KitKat 4.4 smartphone with 20MP, 1" sensor, the largest sensor for a smartphone ever, with Leica fixed lens equivalent of 28 mm at F2.8, can take RAW image and 4K video, has 21 mm thickness. Furthermore, in 2018 Huawei P20 Pro is an android Oreo 8.1 has triple Leica lenses in the back of the smartphone with 40MP 1/1.7" RGB sensor as first lens, 20MP 1/2.7" monochrome sensor as second lens and 8MP 1/4" RGB sensor with 3x optical zoom as third lens. Combination of first lens and second lens will produce bokeh image with larger high dynamic range, whereas combination of mega pixel first lens and optical zoom will produce maximum 5x digital zoom without loss of quality by reducing the image size to 8MP.

Light-field cameras were introduced in 2013 with one consumer product and several professional ones.

After a big dip of sales in 2012, consumer digital camera sales declined again in 2013 by 36 percent. In 2011, compact digital cameras sold 10 million per month. In 2013, sales fell to about 4 million per month. DSLR and MILC sales also declined in 2013 by 10–15% after almost ten years of double digit growth. Worldwide unit sales of digital cameras is continuously declining from 148 million in 2011 to 58 million in 2015 and tends to decrease more in the following years.

Film camera sales hit their peak at about 37 million units in 1997, while digital camera sales began in 1989. By 2008, the film camera market had died and digital camera sales hit their peak at 121 million units in 2010. In 2002, cell phones with an integrated camera had been introduced and in 2003 the cell phone with an integrated camera had sold 80 million units per year. By 2011, cell phones with an integrated camera were selling hundreds of millions per year, which were causing a decline in digital cameras. In 2015, digital camera sales were 35 million units or only less than a third of digital camera sales numbers at their peak and also slightly less than film camera sold number at their peak.

Connectivity

Transferring photos

Many digital cameras can connect directly to a computer to transfer data:-

  • Early cameras used the PC serial port. USB is now the most widely used method (most cameras are viewable as USB mass storage), though some have a FireWire port. Some cameras use USB PTP mode for connection instead of USB MSC; some offer both modes.

  • Cameras with integrated Wi-Fi or specific Wi-Fi adapters mostly allow camera control, especially shutter release, exposure control and more (tethering) from computer or smartphone apps additionally to the transfer of media data.

  • Cameraphones and some high-end stand-alone digital cameras also use cellular networks to connect for sharing images. The most common standard on cellular networks is the MMS Multimedia Messaging Service, commonly called "picture messaging". The second method with smartphones is to send a picture as an email attachment. Many old cameraphones, however, do not support email.

A common alternative is the use of a card reader which may be capable of reading several types of storage media, as well as high speed transfer of data to the computer. Use of a card reader also avoids draining the camera battery during the download process. An external card reader allows convenient direct access to the images on a collection of storage media. But if only one storage card is in use, moving it back and forth between the camera and the reader can be inconvenient. Many computers have a card reader built in, at least for SD cards.

Printing photos

Many modern cameras support the PictBridge standard, which allows them to send data directly to a PictBridge-capable computer printer without the need for a computer.

Wireless connectivity can also provide for printing photos without a cable connection.

An instant-print camera, is a digital camera with a built-in printer. This confers a similar functionality as an instant camera which uses instant film to quickly generate a physical photograph. Such non-digital cameras were popularized by Polaroid with the SX-70 in 1972.

Displaying photos

Many digital cameras include a video output port. Usually sVideo, it sends a standard-definition video signal to a television, allowing the user to show one picture at a time. Buttons or menus on the camera allow the user to select the photo, advance from one to another, or automatically send a "slide show" to the TV.

HDMI has been adopted by many high-end digital camera makers, to show photos in their high-resolution quality on an HDTV.

In January 2008, Silicon Image announced a new technology for sending video from mobile devices to a television in digital form. MHL sends pictures as a video stream, up to 1080p resolution, and is compatible with HDMI.

Some DVD recorders and television sets can read memory cards used in cameras; alternatively several types of flash card readers have TV output capability.

Weather-sealing and waterproofing

Cameras can be equipped with a varying amount of environmental sealing to provide protection against splashing water, moisture (humidity and fog), dust and sand, or complete waterproofness to a certain depth and for a certain duration. The latter is one of the approaches to allow underwater photography, the other approach being the use of waterproof housings. Many waterproof digital cameras are also shockproof and resistant to low temperatures.

Some waterproof cameras can be fitted with a waterproof housing to increase the operational depth range. The Olympus 'Tough' range of compact cameras is an example.

Modes

Many digital cameras have preset modes for different applications. Within the constraints of correct exposure various parameters can be changed, including exposure, aperture, focusing, light metering, white balance, and equivalent sensitivity. For example, a portrait might use a wider aperture to render the background out of focus, and would seek out and focus on a human face rather than other image content.

Few cameras are equipped with a voice note (audio-only) recording feature.

Scene modes

Vendors implement a variety scene modes in cameras' firmwares for various purposes, such as a "landscape mode" which prevents focusing on rainy and/or stained window glass such as a windshield, and a "sports mode" which reduces motion blur of moving subjects by reducing exposure time with the help of increased light sensitivity. Firmwares may be equipped with the ability to select a suitable scene mode automatically through artificial intelligence.

Image data storage

A CompactFlash (CF) card, one of many media types used to store digital photographs
 
Digital camera (Panasonic Lumix DMC-TZ10) user interface, indicating the approximate count of remaining photos.

Many camera phones and most stand alone digital cameras store image data in flash memory cards or other removable media. Most stand-alone cameras use SD format, while a few use CompactFlash or other types. In January 2012, a faster XQD card format was announced. In early 2014, some high end cameras have two hot-swappable memory slots. Photographers can swap one of the memory card with camera-on. Each memory slot can accept either Compact Flash or SD Card. All new Sony cameras also have two memory slots, one for its Memory Stick and one for SD Card, but not hot-swapable.

The approximate count of remaining photos until space exhaustion is calculated by the firmware throughout use and indicated in the viewfinder, to prepare the user for an impending necessary hot swap of the memory card, and/or file offload.

A few cameras used other removable storage such as Microdrives (very small hard disk drives), CD single (185 MB), and 3.5" floppy disks (e. g. Sony Mavica). Other unusual formats include:

  • Onboard (internal) flash memory — Cheap cameras and cameras secondary to the device's main use (such as a camera phone). Some have small capacities such as 100 Megabytes and less, where intended use is buffer storage for uninterrupted operation during a memory card hot swap.
  • SuperDisk (LS120) used in two Panasonic digital cameras, the PV-SD4090 and PV-SD5000, which allowed them to use both SuperDisk and 3.5" floppy disks
  • PC Card hard drives — early professional cameras (discontinued)
  • Thermal printer — known only in one model of camera that printed images immediately rather than storing
  • Zink technology — printiing images immediately rather than storing

Most manufacturers of digital cameras do not provide drivers and software to allow their cameras to work with Linux or other free software. Still, many cameras use the standard USB mass storage and/or Media Transfer Protocol, and are thus widely supported. Other cameras are supported by the gPhoto project, and many computers are equipped with a memory card reader.

File formats

The Joint Photography Experts Group standard (JPEG) is the most common file format for storing image data. Other file types include Tagged Image File Format (TIFF) and various Raw image formats.

Many cameras, especially high-end ones, support a raw image format. A raw image is the unprocessed set of pixel data directly from the camera's sensor, often saved in a proprietary format. Adobe Systems has released the DNG format, a royalty-free raw image format used by at least 10 camera manufacturers.

Raw files initially had to be processed in specialized image editing programs, but over time many mainstream editing programs, such as Google's Picasa, have added support for raw images. Rendering to standard images from raw sensor data allows more flexibility in making major adjustments without losing image quality or retaking the picture.

Formats for movies are AVI, DV, MPEG, MOV (often containing motion JPEG), WMV, and ASF (basically the same as WMV). Recent formats include MP4, which is based on the QuickTime format and uses newer compression algorithms to allow longer recording times in the same space.

Other formats that are used in cameras (but not for pictures) are the Design Rule for Camera Format (DCF), an ISO specification, used in almost all camera since 1998, which defines an internal file structure and naming. Also used is the Digital Print Order Format (DPOF), which dictates what order images are to be printed in and how many copies. The DCF 1998 defines a logical file system with 8.3 filenames and makes the usage of either FAT12, FAT16, FAT32 or exFAT mandatory for its physical layer in order to maximize platform interoperability.

Most cameras include Exif data that provides metadata about the picture. Exif data may include aperture, exposure time, focal length, date and time taken. Some are able to tag the location.

Directory and file structure

In order to guarantee interoperability, DCF specifies the file system for image and sound files to be used on formatted DCF media (like removable or non-removable memory) as FAT12, FAT16, FAT32, or exFAT. Media with a capacity of more than 2 GB must be formatted using FAT32 or exFAT.

The filesystem in a digital camera contains a DCIM (Digital Camera IMages) directory, which can contain multiple subdirectories with names such as "123ABCDE" that consist of a unique directory number (in the range 100...999) and five alphanumeric characters, which may be freely chosen and often refer to a camera maker. These directories contain files with names such as "ABCD1234.JPG" that consist of four alphanumeric characters (often "100_", "DSC0", "DSCF", "IMG_", "MOV_", or "P000"), followed by a number. Handling of directories with possibly user-created duplicate numbers may vary among camera firmwares.

DCF 2.0 adds support for DCF optional files recorded in an optional color space (that is, Adobe RGB rather than sRGB). Such files must be indicated by a leading "_" (as in "_DSC" instead of "100_" or "DSC0").

Thumbnail files

To enable loading many images in miniature view quickly and efficiently, and to retain meta data, some vendors' firmwares generate accompanying low-resolution thumbnail files for videos and raw photos. For example, those of Canon cameras end with .THM. JPEG can already store a thumbnail image standalone.

Batteries

Digital cameras have become smaller over time, resulting in an ongoing need to develop a battery small enough to fit in the camera and yet able to power it for a reasonable length of time.

Digital cameras utilize either proprietary or standard consumer batteries. As of March 2014, most cameras use proprietary lithium-ion batteries while some use standard AA batteries or primarily use a proprietary Lithium-ion rechargeable battery pack but have an optional AA battery holder available.

Proprietary

The most common class of battery used in digital cameras is proprietary battery formats. These are built to a manufacturer's custom specifications. Almost all proprietary batteries are lithium-ion. In addition to being available from the OEM, aftermarket replacement batteries are commonly available for most camera models.

Standard consumer batteries

Digital cameras that utilize off-the-shelf batteries are typically designed to be able to use both single-use disposable and rechargeable batteries, but not with both types in use at the same time. The most common off-the-shelf battery size used is AA. CR2, CR-V3 batteries, and AAA batteries are also used in some cameras. The CR2 and CR-V3 batteries are lithium based, intended for a single use. Rechargeable RCR-V3 lithium-ion batteries are also available as an alternative to non-rechargeable CR-V3 batteries.

Some battery grips for DSLRs come with a separate holder to accommodate AA cells as an external power source.

Conversion of film cameras to digital

When digital cameras became common, many photographers asked whether their film cameras could be converted to digital. The answer was not immediately clear, as it differed among models. For the majority of 35 mm film cameras the answer is no, the reworking and cost would be too great, especially as lenses have been evolving as well as cameras. For most a conversion to digital, to give enough space for the electronics and allow a liquid crystal display to preview, would require removing the back of the camera and replacing it with a custom built digital unit.

Many early professional SLR cameras, such as the Kodak DCS series, were developed from 35 mm film cameras. The technology of the time, however, meant that rather than being digital "backs" the bodies of these cameras were mounted on large, bulky digital units, often bigger than the camera portion itself. These were factory built cameras, however, not aftermarket conversions.

A notable exception is the Nikon E2 and Nikon E3, using additional optics to convert the 35 mm format to a 2/3 CCD-sensor.

A few 35 mm cameras have had digital camera backs made by their manufacturer, Leica being a notable example. Medium format and large format cameras (those using film stock greater than 35 mm), have a low unit production, and typical digital backs for them cost over $10,000. These cameras also tend to be highly modular, with handgrips, film backs, winders, and lenses available separately to fit various needs.

The very large sensor these backs use leads to enormous image sizes. For example, Phase One's P45 39 MP image back creates a single TIFF image of size up to 224.6 MB, and even greater pixel counts are available. Medium format digitals such as this are geared more towards studio and portrait photography than their smaller DSLR counterparts; the ISO speed in particular tends to have a maximum of 400, versus 6400 for some DSLR cameras. (Canon EOS-1D Mark IV and Nikon D3S have ISO 12800 plus Hi-3 ISO 102400 with the Canon EOS-1Dx's ISO of 204800).

Digital camera backs

In the industrial and high-end professional photography market, some camera systems use modular (removable) image sensors. For example, some medium format SLR cameras, such as the Mamiya 645D series, allow installation of either a digital camera back or a traditional photographic film back.

  • Area array
    • CCD
    • CMOS
  • Linear array
    • CCD (monochrome)
    • 3-strip CCD with color filters

Linear array cameras are also called scan backs.

  • Single-shot
  • Multi-shot (three-shot, usually)

Most earlier digital camera backs used linear array sensors, moving vertically to digitize the image. Many of them only capture grayscale images. The relatively long exposure times, in the range of seconds or even minutes generally limit scan backs to studio applications, where all aspects of the photographic scene are under the photographer's control.

Some other camera backs use CCD arrays similar to typical cameras. These are called single-shot backs.

Since it is much easier to manufacture a high-quality linear CCD array with only thousands of pixels than a CCD matrix with millions, very high resolution linear CCD camera backs were available much earlier than their CCD matrix counterparts. For example, you could buy an (albeit expensive) camera back with over 7,000 pixel horizontal resolution in the mid-1990s. However, as of 2004, it is still difficult to buy a comparable CCD matrix camera of the same resolution. Rotating line cameras, with about 10,000 color pixels in its sensor line, are able, as of 2005, to capture about 120,000 lines during one full 360 degree rotation, thereby creating a single digital image of 1,200 Megapixels.

Most modern digital camera backs use CCD or CMOS matrix sensors. The matrix sensor captures the entire image frame at once, instead of incrementing scanning the frame area through the prolonged exposure. For example, Phase One produces a 39 million pixel digital camera back with a 49.1 x 36.8 mm CCD in 2008. This CCD array is a little smaller than a frame of 120 film and much larger than a 35 mm frame (36 x 24 mm). In comparison, consumer digital cameras use arrays ranging from 36 x 24 mm (full frame on high end consumer DSLRs) to 1.28 x 0.96 mm (on camera phones) CMOS sensor.

Weakly interacting massive particles

From Wikipedia, the free encyclopedia
https://en.wikipedia.org/wiki/Weakly_interacting_massive_particles

Weakly interacting massive particles (WIMPs) are hypothetical particles that are one of the proposed candidates for dark matter.

There exists no formal definition of a WIMP, but broadly, a WIMP is a new elementary particle which interacts via gravity and any other force (or forces), potentially not part of the Standard Model itself, which is as weak as or weaker than the weak nuclear force, but also non-vanishing in its strength. Many WIMP candidates are expected to have been produced thermally in the early Universe, similarly to the particles of the Standard Model according to Big Bang cosmology, and usually will constitute cold dark matter. Obtaining the correct abundance of dark matter today via thermal production requires a self-annihilation cross section of , which is roughly what is expected for a new particle in the 100 GeV mass range that interacts via the electroweak force.

Experimental efforts to detect WIMPs include the search for products of WIMP annihilation, including gamma rays, neutrinos and cosmic rays in nearby galaxies and galaxy clusters; direct detection experiments designed to measure the collision of WIMPs with nuclei in the laboratory, as well as attempts to directly produce WIMPs in colliders, such as the LHC.

Because supersymmetric extensions of the Standard Model of particle physics readily predict a new particle with these properties, this apparent coincidence is known as the "WIMP miracle", and a stable supersymmetric partner has long been a prime WIMP candidate. However, recent null results from direct-detection experiments along with the failure to produce evidence of supersymmetry in the Large Hadron Collider (LHC) experiment has cast doubt on the simplest WIMP hypothesis.

Theoretical framework and properties

WIMP-like particles are predicted by R-parity-conserving supersymmetry, a popular type of extension to the Standard Model of particle physics, although none of the large number of new particles in supersymmetry have been observed. WIMP-like particles are also predicted by universal extra dimension and little Higgs theories.

Model parity candidate
SUSY R-parity lightest supersymmetric particle (LSP)
UED KK-parity lightest Kaluza-Klein particle (LKP)
little Higgs T-parity lightest T-odd particle (LTP)

The main theoretical characteristics of a WIMP are:

Because of their lack of electromagnetic interaction with normal matter, WIMPs would be invisible through normal electromagnetic observations. Because of their large mass, they would be relatively slow moving and therefore "cold". Their relatively low velocities would be insufficient to overcome the mutual gravitational attraction, and as a result, WIMPs would tend to clump together. WIMPs are considered one of the main candidates for cold dark matter, the others being massive compact halo objects (MACHOs) and axions. These names were deliberately chosen for contrast, with MACHOs named later than WIMPs. In contrast to MACHOs, there are no known stable particles within the Standard Model of particle physics that have all the properties of WIMPs. The particles that have little interaction with normal matter, such as neutrinos, are all very light, and hence would be fast moving, or "hot".

As dark matter

A decade after the dark matter problem was established in the 1970s, WIMPs were suggested as a potential solution to the issue. Although the existence of WIMPs in nature is still hypothetical, it would resolve a number of astrophysical and cosmological problems related to dark matter. There is consensus today among astronomers that most of the mass in the Universe is indeed dark. Simulations of a universe full of cold dark matter produce galaxy distributions that are roughly similar to what is observed. By contrast, hot dark matter would smear out the large-scale structure of galaxies and thus is not considered a viable cosmological model.

WIMPs fit the model of a relic dark matter particle from the early Universe, when all particles were in a state of thermal equilibrium. For sufficiently high temperatures, such as those existing in the early Universe, the dark matter particle and its antiparticle would have been both forming from and annihilating into lighter particles. As the Universe expanded and cooled, the average thermal energy of these lighter particles decreased and eventually became insufficient to form a dark matter particle-antiparticle pair. The annihilation of the dark matter particle-antiparticle pairs, however, would have continued, and the number density of dark matter particles would have begun to decrease exponentially. Eventually, however, the number density would become so low that the dark matter particle and antiparticle interaction would cease, and the number of dark matter particles would remain (roughly) constant as the Universe continued to expand. Particles with a larger interaction cross section would continue to annihilate for a longer period of time, and thus would have a smaller number density when the annihilation interaction ceases. Based on the current estimated abundance of dark matter in the Universe, if the dark matter particle is such a relic particle, the interaction cross section governing the particle-antiparticle annihilation can be no larger than the cross section for the weak interaction. If this model is correct, the dark matter particle would have the properties of the WIMP.

Indirect detection

Because WIMPs may only interact through gravitational and weak forces, they are extremely difficult to detect. However, there are many experiments underway to attempt to detect WIMPs both directly and indirectly. Indirect detection refers to the observation of annihilation or decay products of WIMPs far away from Earth. Indirect detection efforts typically focus on locations where WIMP dark matter is thought to accumulate the most: in the centers of galaxies and galaxy clusters, as well as in the smaller satellite galaxies of the Milky Way. These are particularly useful since they tend to contain very little baryonic matter, reducing the expected background from standard astrophysical processes. Typical indirect searches look for excess gamma rays, which are predicted both as final-state products of annihilation, or are produced as charged particles interact with ambient radiation via inverse Compton scattering. The spectrum and intensity of a gamma ray signal depends on the annihilation products, and must be computed on a model-by-model basis. Experiments that have placed bounds on WIMP annihilation, via the non-observation of an annihilation signal, include the Fermi-LAT gamma ray telescope and the VERITAS ground-based gamma ray observatory. Although the annihilation of WIMPs into Standard Model particles also predicts the production of high-energy neutrinos, their interaction rate is too low to reliably detect a dark matter signal at present. Future observations from the IceCube observatory in Antarctica may be able to differentiate WIMP-produced neutrinos from standard astrophysical neutrinos; however, by 2014, only 37 cosmological neutrinos had been observed, making such a distinction impossible.

Another type of indirect WIMP signal could come from the Sun. Halo WIMPs may, as they pass through the Sun, interact with solar protons, helium nuclei as well as heavier elements. If a WIMP loses enough energy in such an interaction to fall below the local escape velocity, it would not have enough energy to escape the gravitational pull of the Sun and would remain gravitationally bound. As more and more WIMPs thermalize inside the Sun, they begin to annihilate with each other, forming a variety of particles, including high-energy neutrinos. These neutrinos may then travel to the Earth to be detected in one of the many neutrino telescopes, such as the Super-Kamiokande detector in Japan. The number of neutrino events detected per day at these detectors depends on the properties of the WIMP, as well as on the mass of the Higgs boson. Similar experiments are underway to detect neutrinos from WIMP annihilations within the Earth and from within the galactic center.

Direct detection

Direct detection refers to the observation of the effects of a WIMP-nucleus collision as the dark matter passes through a detector in an Earth laboratory. While most WIMP models indicate that a large enough number of WIMPs must be captured in large celestial bodies for indirect detection experiments to succeed, it remains possible that these models are either incorrect or only explain part of the dark matter phenomenon. Thus, even with the multiple experiments dedicated to providing indirect evidence for the existence of cold dark matter, direct detection measurements are also necessary to solidify the theory of WIMPs.

Although most WIMPs encountering the Sun or the Earth are expected to pass through without any effect, it is hoped that a large number of dark matter WIMPs crossing a sufficiently large detector will interact often enough to be seen—at least a few events per year. The general strategy of current attempts to detect WIMPs is to find very sensitive systems that can be scaled up to large volumes. This follows the lessons learned from the history of the discovery and (by now) routine detection of the neutrino.

Fig 1. CDMS parameter space excluded as of 2004. DAMA result is located in green area and is disallowed.

Experimental techniques

Cryogenic crystal detectors – A technique used by the Cryogenic Dark Matter Search (CDMS) detector at the Soudan Mine relies on multiple very cold germanium and silicon crystals. The crystals (each about the size of a hockey puck) are cooled to about 50 mK. A layer of metal (aluminium and tungsten) at the surfaces is used to detect a WIMP passing through the crystal. This design hopes to detect vibrations in the crystal matrix generated by an atom being "kicked" by a WIMP. The tungsten transition edge sensors (TES) are held at the critical temperature so they are in the superconducting state. Large crystal vibrations will generate heat in the metal and are detectable because of a change in resistance. CRESST, CoGeNT, and EDELWEISS run similar setups.

Noble gas scintillators – Another way of detecting atoms "knocked about" by a WIMP is to use scintillating material, so that light pulses are generated by the moving atom and detected, often with PMTs. Experiments such as DEAP at SNOLAB and DarkSide at the LNGS instrument a very large target mass of liquid argon for sensitive WIMP searches. ZEPLIN, and XENON used xenon to exclude WIMPs at higher sensitivity, with the most stringent limits to date provided by the XENON1T detector, utilizing 3.5 tons of liquid xenon. Even larger multi-ton liquid xenon detectors have been approved for construction from the XENON, LUX-ZEPLIN and PandaX collaborations.

Crystal scintillators – Instead of a liquid noble gas, an in principle simpler approach is the use of a scintillating crystal such as NaI(Tl). This approach is taken by DAMA/LIBRA, an experiment that observed an annular modulation of the signal consistent with WIMP detection (see § Recent limits). Several experiments are attempting to replicate those results, including ANAIS and DM-Ice, which is codeploying NaI crystals with the IceCube detector at the South Pole. KIMS is approaching the same problem using CsI(Tl) as a scintillator. The COSINE-100 collaboration (a merging of KIMS and DM-Ice groups) published their results on replicating the DAMA/LIBRA signal in December 2018 in journal Nature; their conclusion was that "this result rules out WIMP–nucleon interactions as the cause of the annual modulation observed by the DAMA collaboration". In 2021 new results from ANAIS-112 and COSINE-100 have both failed to replicate the DAMA/LIBRA signal.

Bubble chambers – The PICASSO (Project In Canada to Search for Supersymmetric Objects) experiment is a direct dark matter search experiment that is located at SNOLAB in Canada. It uses bubble detectors with Freon as the active mass. PICASSO is predominantly sensitive to spin-dependent interactions of WIMPs with the fluorine atoms in the Freon. COUPP, a similar experiment using trifluoroiodomethane(CF3I), published limits for mass above 20 GeV in 2011. The two experiments merged into PICO collaboration in 2012.

A bubble detector is a radiation sensitive device that uses small droplets of superheated liquid that are suspended in a gel matrix. It uses the principle of a bubble chamber but, since only the small droplets can undergo a phase transition at a time, the detector can stay active for much longer periods. When enough energy is deposited in a droplet by ionizing radiation, the superheated droplet becomes a gas bubble. The bubble development is accompanied by an acoustic shock wave that is picked up by piezo-electric sensors. The main advantage of the bubble detector technique is that the detector is almost insensitive to background radiation. The detector sensitivity can be adjusted by changing the temperature, typically operated between 15 °C and 55 °C. There is another similar experiment using this technique in Europe called SIMPLE.

PICASSO reports results (November 2009) for spin-dependent WIMP interactions on 19F, for masses of 24 Gev new stringent limits have been obtained on the spin-dependent cross section of 13.9 pb (90% CL). The obtained limits restrict recent interpretations of the DAMA/LIBRA annual modulation effect in terms of spin dependent interactions.

PICO is an expansion of the concept planned in 2015.

Other types of detectorsTime projection chambers (TPCs) filled with low pressure gases are being studied for WIMP detection. The Directional Recoil Identification From Tracks (DRIFT) collaboration is attempting to utilize the predicted directionality of the WIMP signal. DRIFT uses a carbon disulfide target, that allows WIMP recoils to travel several millimetres, leaving a track of charged particles. This charged track is drifted to an MWPC readout plane that allows it to be reconstructed in three dimensions and determine the origin direction. DMTPC is a similar experiment with CF4 gas.

The DAMIC (DArk Matter In CCDs) and SENSEI (Sub Electron Noise Skipper CCD Experimental Instrument) collaborations employ the use of scientific Charge Coupled Devices (CCDs) to detect light Dark Matter. The CCDs act as both the detector target and the readout instrumentation. WIMP interactions with the bulk of the CCD can induce the creation of electron-hole pairs, which are then collected and readout by the CCDs. In order to decrease the noise and achieve detection of single electrons, the experiments make use of a type of CCD known as the Skipper CCD, which allows for averaging over repeated measurements of the same collected charge.

Recent limits

Fig. 2: Plot showing the parameter space of dark matter particle mass and interaction cross section with nucleons. The LUX and SuperCDMS limits exclude the parameter space above the labelled curves. The CoGeNT and CRESST-II regions indicate regions which were previously thought to correspond to dark matter signals, but which were later explained with mundane sources. The DAMA and CDMS-Si data remain unexplained, and these regions indicate the preferred parameter space if these anomalies are due to dark matter.

There are currently no confirmed detections of dark matter from direct detection experiments, with the strongest exclusion limits coming from the LUX and SuperCDMS experiments, as shown in figure 2. With 370 kilograms of xenon LUX is more sensitive than XENON or CDMS. First results from October 2013 report that no signals were seen, appearing to refute results obtained from less sensitive instruments. and this was confirmed after the final data run ended in May 2016.

Historically there have been four anomalous sets of data from different direct detection experiments, two of which have now been explained with backgrounds (CoGeNT and CRESST-II), and two which remain unexplained (DAMA/LIBRA and CDMS-Si). In February 2010, researchers at CDMS announced that they had observed two events that may have been caused by WIMP-nucleus collisions.

CoGeNT, a smaller detector using a single germanium puck, designed to sense WIMPs with smaller masses, reported hundreds of detection events in 56 days. They observed an annual modulation in the event rate that could indicate light dark matter. However a dark matter origin for the CoGeNT events has been refuted by more recent analyses, in favour of an explanation in terms of a background from surface events.

Annual modulation is one of the predicted signatures of a WIMP signal, and on this basis the DAMA collaboration has claimed a positive detection. Other groups, however, have not confirmed this result. The CDMS data made public in May 2004 exclude the entire DAMA signal region given certain standard assumptions about the properties of the WIMPs and the dark matter halo, and this has been followed by many other experiments (see Fig 2, right).

The COSINE-100 collaboration (a merging of KIMS and DM-Ice groups) published their results on replicating the DAMA/LIBRA signal in December 2018 in journal Nature; their conclusion was that "this result rules out WIMP–nucleon interactions as the cause of the annual modulation observed by the DAMA collaboration".

The future of direct detection

The 2020s should see the emergence of several multi-tonne mass direct detection experiments, which will probe WIMP-nucleus cross sections orders of magnitude smaller than the current state-of-the-art sensitivity. Examples of such next-generation experiments are LUX-ZEPLIN (LZ) and XENONnT, which are multi-tonne liquid xenon experiments, followed by DARWIN, another proposed liquid xenon direct detection experiment of 50–100 tonnes.

Such multi-tonne experiments will also face a new background in the form of neutrinos, which will limit their ability to probe the WIMP parameter space beyond a certain point, known as the neutrino floor. However, although its name may imply a hard limit, the neutrino floor represents the region of parameter space beyond which experimental sensitivity can only improve at best as the square root of exposure (the product of detector mass and running time). For WIMP masses below 10 GeV the dominant source of neutrino background is from the Sun, while for higher masses the background contains contributions from atmospheric neutrinos and the diffuse supernova neutrino background.

In December 2021, results from PandaX have found no signal in their data, with a lowest excluded cross section of  pb at 40 GeV with 90% confidence level. In July 2022 the LZ experiment published its first limit excluding cross sections above  pb at 30 GeV with 90% confidence level.

Friday, August 5, 2022

Transcriptome

From Wikipedia, the free encyclopedia

The transcriptome is the set of all RNA transcripts, including coding and non-coding, in an individual or a population of cells. The term can also sometimes be used to refer to all RNAs, or just mRNA, depending on the particular experiment. The term transcriptome is a portmanteau of the words transcript and genome; it is associated with the process of transcript production during the biological process of transcription.

The early stages of transcriptome annotations began with cDNA libraries published in the 1980s. Subsequently, the advent of high-throughput technology led to faster and more efficient ways of obtaining data about the transcriptome. Two biological techniques are used to study the transcriptome, namely DNA microarray, a hybridization-based technique and RNA-seq, a sequence-based approach. RNA-seq is the preferred method and has been the dominant transcriptomics technique since the 2010s. Single-cell transcriptomics allows tracking of transcript changes over time within individual cells.

Data obtained from the transcriptome is used in research to gain insight into processes such as cellular differentiation, carcinogenesis, transcription regulation and biomarker discovery among others. Transcriptome-obtained data also finds applications in establishing phylogenetic relationships during the process of evolution and in in vitro fertilization. The transcriptome is closely related to other -ome based biological fields of study; it is complementary to the proteome and the metabolome and encompasses the translatome, exome, meiome and thanatotranscriptome which can be seen as ome fields studying specific types of RNA transcripts. The are quantifiable and conserved relationships between the Transcriptome and other -omes, and Transcriptomics data can be used effectively to predict other molecular species, such as metabolites. There are numerous publicly available transcriptome databases.

Etymology and history

The word transcriptome is a portmanteau of the words transcript and genome. It appeared along with other neologisms formed using the suffixes -ome and -omics to denote all studies conducted on a genome-wide scale in the fields of life sciences and technology. As such, transcriptome and transcriptomics were one of the first words to emerge along with genome and proteome. The first study to present a case of a collection of a cDNA library for silk moth mRNA was published in 1979. The first seminal study to mention and investigate the transcriptome of an organism was published in 1997 and it described 60,633 transcripts expressed in S. cerevisiae using serial analysis of gene expression (SAGE). With the rise of high-throughput technologies and bioinformatics and the subsequent increased computational power, it became increasingly efficient and easy to characterize and analyze enormous amount of data. Attempts to characterize the transcriptome became more prominent with the advent of automated DNA sequencing during the 1980s. During the 1990s, expressed sequence tag sequencing was used to identify genes and their fragments. This was followed by techniques such as serial analysis of gene expression (SAGE), cap analysis of gene expression (CAGE), and massively parallel signature sequencing (MPSS).

Transcription

The transcriptome encompasses all the ribonucleic acid (RNA) transcripts present in a given organism or experimental sample. RNA is the main carrier of genetic information that is responsible for the process of converting DNA into an organism's phenotype. A gene can give rise to a single-stranded messenger RNA (mRNA) through a molecular process known as transcription; this mRNA is complementary to the strand of DNA it originated from. The enzyme RNA polymerase II attaches to the template DNA strand and catalyzes the addition of ribonucleotides to the 3' end of the growing sequence of the mRNA transcript.

In order to initiate its function, RNA polymerase II needs to recognize a promoter sequence, located upstream (5') of the gene. In eukaryotes, this process is mediated by transcription factors, most notably Transcription factor II D (TFIID) which recognizes the TATA box and aids in the positioning of RNA polymerase at the appropriate start site. To finish the production of the RNA transcript, termination takes place usually several hundred nuclecotides away from the termination sequence and cleavage takes place. This process occurs in the nucleus of a cell along with RNA processing by which mRNA molecules are capped, spliced and polyadenylated to increase their stability before being subsequently taken to the cytoplasm. The mRNA gives rise to proteins through the process of translation that takes place in ribosomes.

Types of RNA transcripts

In accordance with the central dogma of molecular biology, the transcriptome initially encompassed only protein-coding mRNA transcripts. Nevertheless, several RNA subtypes with distinct functions exist. Many RNA transcripts do not code for protein or have different regulatory functions in the process of gene transcription and translation. RNA types which do not fall within the scope of the central dogma of molecular biology are non-coding RNAs which can be divided into two groups of long non-coding RNA and short non-coding RNA.

Long non-coding RNA includes all non-coding RNA transcripts that are more than 200 nucleotides long. Members of this group comprise the largest fraction of the non-coding transcriptome. Short non-coding RNA includes the following members:

Scope of study

In the human genome, about 5% of all genes get transcribed into RNA. The transcriptome consists of coding mRNA which comprise around 1-4% of its entirety and non-coding RNAs which comprise the rest of the genome and do not give rise to proteins. The number of non-protein-coding sequences increases in more complex organisms.

Several factors render the content of the transcriptome difficult to establish. These include alternative splicing, RNA editing and alternative transcription among others. Additionally, transcriptome techniques are capable of capturing transcription occurring in a sample at a specific time point, although the content of the transcriptome can change during differentiation. The main aims of transcriptomics are the following: "catalogue all species of transcript, including mRNAs, non-coding RNAs and small RNAs; to determine the transcriptional structure of genes, in terms of their start sites, 5′ and 3′ ends, splicing patterns and other post-transcriptional modifications; and to quantify the changing expression levels of each transcript during development and under different conditions".

The term can be applied to the total set of transcripts in a given organism, or to the specific subset of transcripts present in a particular cell type. Unlike the genome, which is roughly fixed for a given cell line (excluding mutations), the transcriptome can vary with external environmental conditions. Because it includes all mRNA transcripts in the cell, the transcriptome reflects the genes that are being actively expressed at any given time, with the exception of mRNA degradation phenomena such as transcriptional attenuation. The study of transcriptomics, (which includes expression profiling, splice variant analysis etc.), examines the expression level of RNAs in a given cell population, often focusing on mRNA, but sometimes including others such as tRNAs and sRNAs.

Methods of construction

Transcriptomics is the quantitative science that encompasses the assignment of a list of strings ("reads") to the object ("transcripts" in the genome). To calculate the expression strength, the density of reads corresponding to each object is counted. Initially, transcriptomes were analyzed and studied using expressed sequence tags libraries and serial and cap analysis of gene expression (SAGE).

Currently, the two main transcriptomics techniques include DNA microarrays and RNA-Seq. Both techniques require RNA isolation through RNA extraction techniques, followed by its separation from other cellular components and enrichment of mRNA.

There are two general methods of inferring transcriptome sequences. One approach maps sequence reads onto a reference genome, either of the organism itself (whose transcriptome is being studied) or of a closely related species. The other approach, de novo transcriptome assembly, uses software to infer transcripts directly from short sequence reads and is used in organisms with genomes that are not sequenced.

DNA microarrays

DNA microarray used to detect gene expression in human (left) and mouse (right) samples

The first transcriptome studies were based on microarray techniques (also known as DNA chips). Microarrays consist of thin glass layers with spots on which oligonucleotides, known as "probes" are arrayed; each spot contains a known DNA sequence.

When performing microarray analyses, mRNA is collected from a control and an experimental sample, the latter usually representative of a disease. The RNA of interest is converted to cDNA to increase its stability and marked with fluorophores of two colors, usually green and red, for the two groups. The cDNA is spread onto the surface of the microarray where it hybridizes with oligonucleotides on the chip and a laser is used to scan. The fluorescence intensity on each spot of the microarray corresponds to the level of gene expression and based on the color of the fluorophores selected, it can be determined which of the samples exhibits higher levels of the mRNA of interest.

One microarray usually contains enough oligonucleotides to represent all known genes; however, data obtained using microarrays does not provide information about unknown genes. During the 2010s, microarrays were almost completely replaced by next-generation techniques that are based on DNA sequencing.

RNA sequencing

RNA sequencing is a next-generation sequencing technology; as such it requires only a small amount of RNA and no previous knowledge of the genome. It allows for both qualitative and quantitative analysis of RNA transcripts, the former allowing discovery of new transcripts and the latter a measure of relative quantities for transcripts in a sample.

The three main steps of sequencing transcriptomes of any biological samples include RNA purification, the synthesis of an RNA or cDNA library and sequencing the library. The RNA purification process is different for short and long RNAs. This step is usually followed by an assessment of RNA quality, with the purpose of avoiding contaminants such as DNA or technical contaminants related to sample processing. RNA quality is measured using UV spectrometry with an absorbance peak of 260 nm. RNA integrity can also be analyzed quantitatively comparing the ratio and intensity of 28S RNA to 18S RNA reported in the RNA Integrity Number (RIN) score. Since mRNA is the species of interest and it represents only 3% of its total content, the RNA sample should be treated to remove rRNA and tRNA and tissue-specific RNA transcripts.

The step of library preparation with the aim of producing short cDNA fragments, begins with RNA fragmentation to transcripts in length between 50 and 300 base pairs. Fragmentation can be enzymatic (RNA endonucleases), chemical (trismagnesium salt buffer, chemical hydrolysis) or mechanical (sonication, nebulisation). Reverse transcription is used to convert the RNA templates into cDNA and three priming methods can be used to achieve it, including oligo-DT, using random primers or ligating special adaptor oligos.

Single-cell transcriptomics

Transcription can also be studied at the level of individual cells by single-cell transcriptomics. Single-cell RNA sequencing (scRNA-seq) is a recently developed technique that allows the analysis of the transcriptome of single cells. With single-cell transcriptomics, subpopulations of cell types that constitute the tissue of interest are also taken into consideration. This approach allows to identify whether changes in experimental samples are due to phenotypic cellular changes as opposed to proliferation, with which a specific cell type might be overexpressed in the sample. Additionally, when assessing cellular progression through differentiation, average expression profiles are only able to order cells by time rather than their stage of development and are consequently unable to show trends in gene expression levels specific to certain stages. Single-cell trarnscriptomic techniques have been used to characterize rare cell populations such as circulating tumor cells, cancer stem cells in solid tumors, and embryonic stem cells (ESCs) in mammalian blastocysts.

Although there are no standardized techniques for single-cell transcriptomics, several steps need to be undertaken. The first step includes cell isolation, which can be performed using low- and high-throughput techniques. This is followed by a qPCR step and then single-cell RNAseq where the RNA of interest is converted into cDNA. Newer developments in single-cell transcriptomics allow for tissue and sub-cellular localization preservation through cryo-sectioning thin slices of tissues and sequencing the transcriptome in each slice. Another technique allows the visualization of single transcripts under a microscope while preserving the spatial information of each individual cell where they are expressed.

Analysis

A number of organism-specific transcriptome databases have been constructed and annotated to aid in the identification of genes that are differentially expressed in distinct cell populations.

RNA-seq is emerging (2013) as the method of choice for measuring transcriptomes of organisms, though the older technique of DNA microarrays is still used. RNA-seq measures the transcription of a specific gene by converting long RNAs into a library of cDNA fragments. The cDNA fragments are then sequenced using high-throughput sequencing technology and aligned to a reference genome or transcriptome which is then used to create an expression profile of the genes.

Applications

Mammals

The transcriptomes of stem cells and cancer cells are of particular interest to researchers who seek to understand the processes of cellular differentiation and carcinogenesis. A pipeline using RNA-seq or gene array data can be used to track genetic changes occurring in stem and precursor cells and requires at least three independent gene expression data from the former cell type and mature cells.

Analysis of the transcriptomes of human oocytes and embryos is used to understand the molecular mechanisms and signaling pathways controlling early embryonic development, and could theoretically be a powerful tool in making proper embryo selection in in vitro fertilisation. Analyses of the transcriptome content of the placenta in the first-trimester of pregnancy in in vitro fertilization and embryo transfer (IVT-ET) revealed differences in genetic expression which are associated with higher frequency of adverse perinatal outcomes. Such insight can be used to optimize the practice. Transcriptome analyses can also be used to optimize cryopreservation of oocytes, by lowering injuries associated with the process.

Transcriptomics is an emerging and continually growing field in biomarker discovery for use in assessing the safety of drugs or chemical risk assessment.

Transcriptomes may also be used to infer phylogenetic relationships among individuals or to detect evolutionary patterns of transcriptome conservation.

Transcriptome analyses were used to discover the incidence of antisense transcription, their role in gene expression through interaction with surrounding genes and their abundance in different chromosomes. RNA-seq was also used to show how RNA isoforms, transcripts stemming from the same gene but with different structures, can produce complex phenotypes from limited genomes.

Plants

Transcriptome analysis have been used to study the evolution and diversification process of plant species. In 2014, the 1000 Plant Genomes Project was completed in which the transcriptomes of 1,124 plant species from the families viridiplantae, glaucophyta and rhodophyta were sequenced. The protein coding sequences were subsequently compared to infer phylogenetic relationships between plants and to characterize the time of their diversification in the process of evolution. Transcriptome studies have been used to characterize and quantify gene expression in mature pollen. Genes involved in cell wall metabolism and cytoskeleton were found to be overexpressed. Transcriptome approaches also allowed to track changes in gene expression through different developmental stages of pollen, ranging from microspore to mature pollen grains; additionally such stages could be compared across species of different plants including Arabidopsis, rice and tobacco.

Relation to other ome fields

General schema showing the relationships of the genome, transcriptome, proteome, and metabolome (lipidome).

Similar to other -ome based technologies, analysis of the transcriptome allows for an unbiased approach when validating hypotheses experimentally. This approach also allows for the discovery of novel mediators in signaling pathways. As with other -omics based technologies, the transcriptome can be analyzed within the scope of a multiomics approach. It is complementary to metabolomics but contrary to proteomics, a direct association between a transcript and metabolite cannot be established.

There are several -ome fields that can be seen as subcategories of the transcriptome. The exome differs from the transcriptome in that it includes only those RNA molecules found in a specified cell population, and usually includes the amount or concentration of each RNA molecule in addition to the molecular identities. Additionally, the transcritpome also differs from the translatome, which is the set of RNAs undergoing translation.

The term meiome is used in functional genomics to describe the meiotic transcriptome or the set of RNA transcripts produced during the process of meiosis. Meiosis is a key feature of sexually reproducing eukaryotes, and involves the pairing of homologous chromosome, synapse and recombination. Since meiosis in most organisms occurs in a short time period, meiotic transcript profiling is difficult due to the challenge of isolation (or enrichment) of meiotic cells (meiocytes). As with transcriptome analyses, the meiome can be studied at a whole-genome level using large-scale transcriptomic techniques. The meiome has been well-characterized in mammal and yeast systems and somewhat less extensively characterized in plants.

The thanatotranscriptome consists of all RNA transcripts that continue to be expressed or that start getting re-expressed in internal organs of a dead body 24–48 hours following death. Some genes include those that are inhibited after fetal development. If the thanatotranscriptome is related to the process of programmed cell death (apoptosis), it can be referred to as the apoptotic thanatotranscriptome. Analyses of the thanatotranscriptome are used in forensic medicine.

eQTL mapping can be used to complement genomics with transcriptomics; genetic variants at DNA level and gene expression measures at RNA level.

Relation to proteome

The transcriptome can be seen as a subset of the proteome, that is, the entire set of proteins expressed by a genome.

However, the analysis of relative mRNA expression levels can be complicated by the fact that relatively small changes in mRNA expression can produce large changes in the total amount of the corresponding protein present in the cell. One analysis method, known as gene set enrichment analysis, identifies coregulated gene networks rather than individual genes that are up- or down-regulated in different cell populations.

Although microarray studies can reveal the relative amounts of different mRNAs in the cell, levels of mRNA are not directly proportional to the expression level of the proteins they code for. The number of protein molecules synthesized using a given mRNA molecule as a template is highly dependent on translation-initiation features of the mRNA sequence; in particular, the ability of the translation initiation sequence is a key determinant in the recruiting of ribosomes for protein translation.

Lie point symmetry

From Wikipedia, the free encyclopedia https://en.wikipedia.org/wiki/Lie_point_symmetry     ...