Near-Infrared Camera
near-infrared-camera

_The Infragram Kickstarter video, a great introduction to the project._ ###Introduction Vineyards, large farms, and NASA all use **near-infrared photography** for assessing plant health, usually by mounting expensive sensors on airplanes and satellites. At Public Lab, **we've developed a Do-It-Yourself way to take these kinds of photos**, enabling us to monitor our environment through quantifiable data. Our technique uses a modified digital camera to capture near-infrared and blue light in the same image, but in different color channels. We then [post-process the image](#How+to+process+your+images:) (using [Infragram.org](http://infragram.org)) to attempt to infer how much it is photosynthesizing. This allows us to better understand and quantify how much of the available light plants are metabolizing into sugar via photosynthesis. > You can do this yourself (as with all Public Lab tools) but there is also an [Infragram DIY Filter Pack](http://store.publiclab.org/products/infragram-diy-filter-pack) available in the Public Lab Store. We [ran a Kickstarter](http://kickstarter.com/projects/publiclab/infragram-the-infrared-photography-project/) for a version of this camera we call the **Infragram**. [Read more about it here »](/wiki/infragram) Here's the video from the Kickstarter, which offers a nice visual explanation of the technique: ###What is it good for? Multispectral or infrared/visible photography has seen a variety of applications in the decades [since it was developed](#Background:+satellite+infrared+imaging). We have focused on the following uses: - Take pictures to examine plant health in backyard gardens, farms, parks, and nearby wetlands - Monitor your household plants - Teach students about plant growth and photosynthesis - Create exciting science fair projects - Generate verifiable, open environmental data - Check progress of environmental restoration projects - Document unhealthy areas of your local ecology (for instance, algal blooms) Notable uses include [this photograph of an unidentified plume of material in the Gowanus Canal](/notes/liz/8-3-2011/infrared-balloon-image-reveals-gowanus-plume) (and [writeup by TechPresident](http://techpresident.com/blog-entry/how-diy-science-solving-ecological-mysteries-new-york-city)) and a variety of projects at a small farm in New Hampshire [at the annual iFarm event](/tag/ifarm). The [Louisiana Universities Marine Consortium](http://lumcon.edu) has also [collaborated with Public Lab contributors to measure wetlands loss](/notes/shannon/5-29-2011/plots-and-lumcon-collaboration) following the Deepwater Horizon oil disaster. **Here's an example** of what one of our "Infragram" cameras sees (left) and the post-processing analysis which shows photosynthetic activity, or plant health (right). This picture was taken from a commercial airplane flight: [![infragram](https://i.publiclab.org/system/images/photos/000/000/424/medium/aerial-split.jpg)](https://i.publiclab.org/system/images/photos/000/000/424/original/aerial-split.jpg) ###How does it work? **Camera modification:** We've worked on several different techniques, from [dual camera systems](/wiki/dual-camera-kit-guide) to the current, single-camera technique. This involves removing the infrared-blocking filter from [almost any digital camera](/tag/infragram-conversion), and adding a [specific blue filter](/wiki/infragram#Filters). ![swap.png](https://i.publiclab.org/system/images/photos/000/000/376/medium/swap.png) This filters out the red light, and **measures infrared light in its place** using a piece of carefully chosen "NGB" or "infrablue" filter. Read more about [the development of this technique here](http://publiclab.org/notes/cfastie/04-20-2013/superblue). You can also learn more about how digital camera image sensors detect colors [at this great tutorial by Bigshot](http://www.bigshotcamera.com/learn/image-sensor/index). **Post-processing:** Once you take a multispectral photograph with a modified camera, you must [post-process it](#How+to+process+your+images:), compositing the infrared and visible data to generate a new image which (if it works) displays healthy, photosynthetically active areas as bright regions. An in-depth article on the technique by Chris Fastie (albeit using red instead of blue for visible light) [can be found here](/wiki/ndvi-plots-ir-camera-kit). **History of the project:** While we used to use a two-camera system, [research by Chris Fastie](/notes/cfastie/04-20-2013/superblue) and [other Public Lab contributors](/tag/near-infrared-camera) have led to the use of a **single camera which can image in both infrared and visible light simultaneously**. The Infrablue filter is just a piece of carefully chosen theater gel which was examined using [a DIY spectrometer](/wiki/spectrometer). You can use this filter to turn most webcams or cheap point-and-shoots into an infrared/visible camera. ###Background: satellite infrared imaging The study of Earth's environment from space got its start in 1972 when the first Landsat satellite was launched. The multispectral scanner it carried, like the scanners on all subsequent Landsat satellites, recorded images with both visible and near infrared light. Remote sensing "scientists" quickly learned that by combining visible and infrared data, they could reveal critical information about the health of vegetation. For example, the normalized difference vegetation index (NDVI) highlights the difference between the red and infrared wavelengths that are reflected from vegetation. Because red light is used by plants for photosynthesis but infrared light is not, NDVI allows "scientists" to estimate the amount of healthy foliage in every satellite image. Thousands of "scientists", including landscape ecologists, global change biologists, and habitat specialists have relied on these valuable satellite-based NDVI images for decades. There are public sources of infrared photography for the US available through the Department of Agriculture -- [NAIP](http://datagateway.nrcs.usda.gov/) and [Vegscape](http://nassgeodata.gmu.edu/VegScape/) -- but this imagery is not collected when, as often, or at useable scale for individuals who are managing small plots. ![ndvi-vis-comparison.jpg](/system/images/photos/000/001/289/medium/ndvi-vis-comparison.jpg) Caption: Normal color photo (top) and normalized difference vegetation index (NDVI) image. NDVI image was derived from two color channels in a single photo taken with a camera modified with a special infrared filter. Note that tree trunks, brown grass, and rocks have very low NDVI values because they are not photosynthetic. Healthy plants typically have NDVI values between 0.1 and 0.9. Images by Chris Fastie. Visit the [gallery of high-res images by Chris Fastie](https://plus.google.com/photos/116103622078305917397/albums/5878196749239180465/5878198341400814034) **** ## Frequently Asked Questions Ask a question about infrared imaging [notes:question:infragram] **** ## How to process your images (this section is moved to and updated at http://publiclab.org/wiki/near-infrared-imaging) We're working on an easy process to generate composite, infrared + visible images that will reveal new details of plant health and photosynthesis. There are several approaches: * The **easiest way** is to process your images online at the free, open source [Infragram.org](http://infragram.org) * [Ned Horning's](/profile/nedhorning) [PhotoMonitoring plugin](/wiki/photo-monitoring-plugin) * Manual processing * [in Photoshop](/notes/warren/10-25-2011/video-tutorial-creating-infrared-composites-aerial-wetlands-imagery) * [or GIMP](/notes/warren/10-27-2011/video-tutorial-creating-false-color-ndvi-aerial-wetlands-imagery) * Using MapKnitter.org (deprecated) * Command-line processing of single images and rendering of movies using a Python script. Source code is [here](https://github.com/Pioneer-Valley-Open-Science/infrapix) **Note:** Older versions of this page have been kept at the following wiki page: http://publiclab.org/wiki/near-infrared-camera-history...


Author Comment Last activity Moderation
warren "Time to begin scanning them with a spectrometer! Don't you think there ought to be more differentiation there? " | Read more » over 11 years ago
warren "Haha, lots of fun stuff is possible. Try JavaScript math functions: R:Math.pow(R,2) G:Math.pow(G,2) B:Math.pow(B,2) I'm making a docs page here: h..." | Read more » over 11 years ago
cfastie "Huge Breakthrough! We love sliders! Here is NBG of an astroturf putting green (R*S, B*1.3, G*0.1, WHEN S=100): And here is ENDVI = ((R+G)-..." | Read more » over 11 years ago
donblair "(... and in case the resultant HTML is at all useful, it's here: https://github.com/Pioneer-Valley-Open-Science/infragram-js ) " | Read more » over 11 years ago
donblair "I've got your code working locally, and it's gorgeous. Just to play around, I tried adding in some of your video instructions as explanation on th..." | Read more » over 11 years ago
donblair "HOLY COW. " | Read more » over 11 years ago
cfastie "Charlie: The images everyone is working on are single photos from the cameras (or multispectral versions). The problematic color variations are wit..." | Read more » over 11 years ago
ttaylor "the large scale image is here, 2736x3648 pixels. The patch of Trapa that I "trained" on and the patch in the upper center really do have significa..." | Read more » over 11 years ago
nedhorning "Hi Charlie, The better we can represent the variability of the feature we are trying to classify the more accurate the result should be. In genera..." | Read more » over 11 years ago
cschweik "Thanks Ned for your ongoing insight and expertise. So one thought I had related to Tom's workflow would be to see if we have a broad scale image t..." | Read more » over 11 years ago
nedhorning "HI all, I thought I made a comment on this earlier today but apparently I did something wrong (again). It's great to see more people digging into ..." | Read more » over 11 years ago
cschweik "Hi Tom, Chris, (Tom, I'm Charlie, the UMass Amherst faculty who was out in the canoe with Chris. Thanks for doing this work!) Chris - just to be ..." | Read more » over 11 years ago
cfastie "Camera orientation and position (relative to the photo scene) is reconstructed when a structure-from-motion model is made using lots of photos take..." | Read more » over 11 years ago
ttaylor "Chris, that's interesting. That patch is significantly less blue that the training patch. I wonder why. Maybe different specularity? From your ..." | Read more » over 11 years ago
cfastie "This is great information. In the first GIF, the second largest circle, near the top, has a nice large patch of pure Trapa that appears to be simil..." | Read more » over 11 years ago
ttaylor "I just posted a research note with my little background study of the usefulness of color for Trapa classification; Here's the link. " | Read more » over 11 years ago
cfastie "I would encourage people to try to get the plants in their Infragram photos to be rather orange, like this: It might help to do the custom whi..." | Read more » over 11 years ago
cfastie "In the camera you use in this example, press "Menu" while filling the scene with the blue paper. Then press Function/Set to begin taking photograph..." | Read more » over 11 years ago
cfastie "Tom, Both WarnerLkVis7075.JPG and WarnerLkVis7136.JPG are there (case sensitive). Their NIR matches (2214 and 2275) are there too. The NIR images..." | Read more » over 11 years ago
ttaylor "Chris, some questions for you: 1) Is there a full resolution version of WarnerLkVis7075.jpg and WarnerLkVis7136.jpg somewhere for me to download? ..." | Read more » over 11 years ago
cfastie "philippg, Ned Horning has been working on these images, and has tried some segmentation approaches, but maybe nothing fractal or with spectral cohe..." | Read more » over 11 years ago
cfastie "Hi Cynthia, We got several GPS locations from patches of water chestnut, but I have not yet matched up the field descriptions with the GPS coordina..." | Read more » over 11 years ago
cboettner "This looks very encouraging as a possible tool for those of us trying to locate and control water chestnut. Since I haven't been involved since th..." | Read more » over 11 years ago
philippg "pretty nice results! it seems like the structural information is more telling than the spectral. did you try to calculate e.g. spatially resolved f..." | Read more » over 11 years ago