SUPPORTING REMOTE SENSING RESEARCH WITH SMALL UNMANNED AERIAL SYSTEMS

Similar documents
The X100. Safe and fully automatic. Fast and with survey accuracy. revolutionary mapping. create your own orthophotos and DSMs

Greg Colley, Suave Aerial Photographers

3D Vision An enabling Technology for Advanced Driver Assistance and Autonomous Offroad Driving

INSPIRE 1 Release Notes Overview: What s New: Bug Fixes: Notice: 1. All-in-One Firmware version updated to: v

Matthew O. Anderson Scott G. Bauer James R. Hanneman. October 2005 INL/EXT

Nighthawk IV UAS. Versatility. User Friendly Design. Capability. Aerial Surveillance Simplified

Ground control station Users feedback

Development of Automatic shooting and telemetry system for UAV photogrammetry INTRODUCTION

Using Tactical Unmanned Aerial Systems to Monitor and Map Wildfires

How To Discuss Unmanned Aircraft System (Uas)

AeroVironment, Inc. Unmanned Aircraft Systems Overview Background

The Chillon Project: Aerial / Terrestrial and Indoor Integration

Photogrammetric Point Clouds

CHAPTER 1 INTRODUCTION

INSPIRE 1 Release Notes Overview: What s New: Bug Fixes: 1. All-in-One Firmware version updated to: v

Design Specifications of an UAV for Environmental Monitoring, Safety, Video Surveillance, and Urban Security

Onboard electronics of UAVs

sonobot autonomous hydrographic survey vehicle product information guide

U g CS for DJI Phantom 2 Vision+, Phantom 3 and Inspire 1

3D Vision An enabling Technology for Advanced Driver Assistance and Autonomous Offroad Driving

How To Make An Orthophoto

KILN GROUP AVIATION DIVISION UAS INSURANCE PROPOSAL FORM GLOSSARY

ZMART Technical Report The International Aerial Robotics Competition 2014

Application for Special Flight Operations Certificate File # ABC123. Company Name Address Phone

IP-S3 HD1. Compact, High-Density 3D Mobile Mapping System

Automatic Labeling of Lane Markings for Autonomous Vehicles

Control Design of Unmanned Aerial Vehicles (UAVs)

THE PERFORMANCE EVALUATION OF MULTI-IMAGE 3D RECONSTRUCTION SOFTWARE WITH DIFFERENT SENSORS

PERFORMANCE TEST ON UAV-BASED PHOTOGRAMMETRIC DATA COLLECTION

High Resolution Digital Surface Models and Orthoimages for Telecom Network Planning

A GUIDE ON HOW TO CHOOSE A DRONE / UAV OPERATOR FOR YOUR TV, FILM, VIDEO OR MEDIA PRODUCTION.

Robot Perception Continued

Current Challenges in UAS Research Intelligent Navigation and Sense & Avoid

HEXAGON GEOSPATIAL BENELUX Hexagon Geospatial Benelux 2015

Innovation Made in Germany SIRIUS UAS.

From Pixel to Info-Cloud News at Leica Geosystems JACIE Denver, 31 March 2011 Ruedi Wagner Hexagon Geosystems, Geospatial Solutions Division.

UAS deployment and data processing during the Balkans flooding with the support to Mine Action

The YellowScan system: A test on the oppida of Nages/Roque de Viou (Languedoc, France) Tristan Allouis, PhD Research and Development, L'Avion Jaune

A PHOTOGRAMMETRIC APPRAOCH FOR AUTOMATIC TRAFFIC ASSESSMENT USING CONVENTIONAL CCTV CAMERA

High speed 3D capture for Configuration Management DOE SBIR Phase II Paul Banks

How To Fuse A Point Cloud With A Laser And Image Data From A Pointcloud

Automatic georeferencing of imagery from high-resolution, low-altitude, low-cost aerial platforms

Micro and Mini UAV Airworthiness, European and NATO Activities

0 28 November 2011 N/A First Release of Manual 1 24 February 2012 Page 21 Error in FY31AP connection diagram corrected.

automatic road sign detection from survey video

RS platforms. Fabio Dell Acqua - Gruppo di Telerilevamento

An inertial haptic interface for robotic applications

MP2128 3X MicroPilot's. Triple Redundant UAV Autopilot

Bi-Directional DGPS for Range Safety Applications

UAV Road Surface Monitoring and Traffic Information

Control of a quadrotor UAV (slides prepared by M. Cognetti)

TAMU-CC UAS Program 1

AERIALPERSPECTIVE. Here we explore this aerial technology and understand what s involved to use these new tools

Aerospace Information Technology Topics for Internships and Bachelor s and Master s Theses

Saint Louis University Rogue Squadron

UAVNet: Prototype of a Highly Adaptive and Mobile Wireless Mesh Network using Unmanned Aerial Vehicles (UAVs) Simon Morgenthaler University of Bern

How To Make An Integrated System For An Unmanned Aircraft System

CORPORATE PRESENTATION December 2015

Visual Servoing using Fuzzy Controllers on an Unmanned Aerial Vehicle

AUTOPILOT QUICK START GUIDE

ACCURACY ASSESSMENT OF BUILDING POINT CLOUDS AUTOMATICALLY GENERATED FROM IPHONE IMAGES

Eigenständige Erkundung komplexer Umgebungen mit einem Hubschrauber UAV und dem Sampling basierten Missionsplaner MiPlEx

How To Control Gimbal

of large scale imagery without GPS

Author: By Siegfried Krainer and Michael Thomas, Infineon Technologies and Ronald Staerz, MCI Innsbruck

Files Used in this Tutorial

Mr. Jay Arnold, Magnolia River Mr. Richard Vincent, McKim & Creed

Synthetic Sensing: Proximity / Distance Sensors

Modeling and Autonomous Flight Simulation of a Small Unmanned Aerial Vehicle

Hardware In The Loop Simulator in UAV Rapid Development Life Cycle

Company Backgrounder 2015

VISIONMAP A3 - SUPER WIDE ANGLE MAPPING SYSTEM BASIC PRINCIPLES AND WORKFLOW

Opportunities for the generation of high resolution digital elevation models based on small format aerial photography

VEHICLE TRACKING USING ACOUSTIC AND VIDEO SENSORS

Learn How to Fly a Gatewing X-100 and Why

PHANTOM 2 Product Release Notes

Jean François Aumont (1), Bastien Mancini (2). (1) Image processing manager, Delair-Tech, (2) Managing director, Delair-Tech. Project : Authors:

IAI/Malat Solutions for the Maritime Arena

The University of Florida Unmanned Aircraft Systems Research Program: Lessons Learned and an Update on Selected Natural Resources Applications

IP-S2 HD. High Definition 3D Mobile Mapping System

Data Review and Analysis Program (DRAP) Flight Data Visualization Program for Enhancement of FOQA

See Appendix A for the petition submitted to the FAA describing the proposed operations and the regulations that the petitioner seeks an exemption.

Basic Principles of Inertial Navigation. Seminar on inertial navigation systems Tampere University of Technology

SIX DEGREE-OF-FREEDOM MODELING OF AN UNINHABITED AERIAL VEHICLE. A thesis presented to. the faculty of

Site Surveillance, Project Management and Communications System for Remote Locations in High-Risk Environments

Unmanned Airship Based High Resolution Images Acquisition and the Processing

Information Contents of High Resolution Satellite Images

Henry H. Perritt, Jr. Attorney at Law 1131 Carol Lane Glencoe, IL (312) August 2014

VTOL UAV. Design of the On-Board Flight Control Electronics of an Unmanned Aerial Vehicle. Árvai László, ZMNE. Tavaszi Szél 2012 ÁRVAI LÁSZLÓ, ZMNE

Using Multiple Unmanned Systems for a Site Security Task

RPV/UAV Surveillance for Transportation Management and Security

Small Unmanned Aircraft Systems (UAS) for Law Enforcement

SONOBOT AUTONOMOUS HYDROGRAPHIC SURVEY VEHICLE PRODUCT INFORMATION GUIDE

Tracking of Small Unmanned Aerial Vehicles

16 th IOCCG Committee annual meeting. Plymouth, UK February mission: Present status and near future

UAS OPERATORS INSURANCE PROPOSAL FORM

The Applanix SmartBase TM Software for Improved Robustness, Accuracy, and Productivity of Mobile Mapping and Positioning

CYCLOPS OSD USER MANUAL 5.0

LIDAR and Digital Elevation Data

Crater detection with segmentation-based image processing algorithm

Transcription:

SUPPORTING REMOTE SENSING RESEARCH WITH SMALL UNMANNED AERIAL SYSTEMS R. C. Anderson*, P. C. Shanks, L. A. Kritis, M. G. Trani National Geospatial-Intelligence Agency (NGA) Robert.C.Anderson@nga.mil, Philip.C.Shanks@nga.mil, Lawrence.A.Kritis@nga.mil, Matthew.G.Trani@nga.mil KEY WORDS: suas, remote sensing, autopilot, mission planner, ground control system, VisualSFM, PhotoScan, DEMs, orthomosaics ABSTRACT: We describe several remote sensing research projects supported with small Unmanned Aerial Systems (suas) operated by the NGA Basic and Applied Research Office. These suas collections provide data supporting Small Business Innovative Research (SBIR), NGA University Research Initiative (NURI), and Cooperative Research And Development Agreements (CRADA) efforts in addition to inhouse research. Some preliminary results related to 3D electro-optical point clouds are presented, and some research goals discussed. Additional details related to the autonomous operational mode of both our multi-rotor and fixed wing small Unmanned Aerial System (suas) platforms are presented. 1. INTRODUCTION The NGA Basic and Applied Research Office has been involved in a number of projects with several universities and companies to test various platforms and sensors with the aim of providing a low-cost, easy to use system in order to facilitate remote sensing research. In addition, the ability for fast photo collection and processing can generate quality products for other applications such as disaster response planning. Currently there are two active CRADAs. One is with St. Louis University (SLU) involving sensor integration to support suas and remote sensing research. SLU has graciously provided lab space for building and testing the systems along with technical expertise. The other is with Urban Robotics, Inc. that provides UAS support with novel approaches to 3-dimensional visualization and geopositioning. They will be exploring new methods of collecting and processing the sensor data to increase the development of 3D rendered maps from small Unmanned Aerial Vehicle (suav) imagery. Vision Systems, Inc. developed a 3-d uncertainty measure using error propagation under a SBIR for their voxel modeling algorithm to reconstruct a 3-d volumetric scene. A toolkit is under development and they have agreed to release some of its code to NGA. Ohio State University s Satellite Positioning and Inertial Navigation (SPIN) laboratory is working on the efficient compression of LiDAR wave forms and using suas for LiDAR data collection. Finally, NGA is gathering the data provided by the universities and companies to apply to their own in-house research. 2.1 suas Goals 2. ORGANIZATIONAL DETAILS The goal of the suas is to provide an inexpensive, safe, easy to use, repeatable method to place payloads in the sky for remote sensing purposes supporting the widest variety of collection modes. NGA has several suas platforms to meet requirements for payload capacity and collection geometry. There are two basic types of aerial platforms, fixed wing and multi-rotor. The fixed wing allows greater payload capacity with superior endurance. The trade-off is that the aircraft is always in forward flight. Also, our fixed wing (2.7m x1.8m) Rascal aircraft needs a relatively long area for takeoff and landing. Multi-rotors, however, offer nearly unlimited number of possibilities for collection geometry at the sacrifice of payload capacity and mission endurance. The NGA current inventory of multi-rotors consists of quadcopters, hexacopters, and an octocopter. Generally, the greater amount of engines corresponds to increased payload capacity and stability. The quadcopter airframe design NGA uses is the Team Black Sheep (TBS) Discovery. The right configuration of multi-rotors can fly in 25 knot winds easily to complete mapping missions. Those types of wind conditions will easily ground a manual aircraft or fixed-wing suas. 2.1 Autopilot The heart of any suas system is its autopilot, a system that enables control of the aircraft through manual or autonomous control. In the last 10 years, the size, weight, power, and cost of autopilots has significantly improved. What was once in the domain of military users now can be purchased online by hobbyists. In order to safely handle our payloads, a variety of proven, reliable autopilot systems were investigated. We currently field 3DRobotics Ardupilot family as it can be configured in a nearly unlimited amount of ways with over 500 parameters. Ardupilot and its ground control system (GCS), Mission Planner, is open source, yet another plus. Due to the large group of developers working on the source code, new functionality is constantly being built into Ardupilot/Pixhawk/Mission Planner systems. Both the Ardupilot and Pixhawk are capable of fully autonomous missions from takeoff to landing. This document contains domestic imagery, which has been included consistent with applicable laws, directives and policies. Handle in accordance with recipient s intelligence oversight and or information handling procedures. doi:10.5194/isprsarchives-xl-1-51-2014 51

2.2 Mission Planner Useful photogrammetric and remote sensing tools are built into Mission Planner as well. Mission Planner has three basic collections modes: racetrack covers large areas with few turnarounds, lawnmower collects over smaller areas using the same back and forth flight path as the racetrack, and circle with flights around a fixed area. Designing the typical lawnmower nadir collection mission consists of a few simple steps. First the user defines the target mapping region by drawing a polygon. With the polygon, a tool called Survey appears where the user can define altitude, camera parameters and cardinal headings (Figure 1). The proposed flight lines are calculated instantly showing the user important mission variables such as Ground Sample Distance, and anticipated mission duration. This lets the user know whether the area can be imaged on a single battery, or if multiple flights are required. Photogrammetric constraints such as focal length, overlap and sidelap can be factored into mission design too. Projected image footprints with overlap and sidelap are also available. After a few minutes, the mission is ready for upload to the suas. Ardupilot is capable of a flight mode called Follow Me which literally follows the user around at a constant altitude. Combined with a realtime vehicle telemetry laced video feed beamed to the user, the situational awareness increases considerably. For multirotors, an extremely useful flight mode features the ability to circle around a Region of Interest (ROI) at a defined radius and altitude while the suas points at the center independent of its flight path. In summary, with the combination of suas and 3DRobotics autopilots, we can place and orient a sensor payload anywhere in 3D space. 2.3 Payloads The TBS whose spider arm configuration eliminates rotor obstruction for our forward mounted GoPro payloads (approximately 400g including gimbal). Typical flight duration is about 15 minutes with a GoPro sized payload. The TBS Discovery is powered by a 5200mAH 4S Lithium-Polymer (Li- Po) battery pack. The TBS can fly approximately 5km downrange and land on a single battery charge. Should an emergency arise, such as loss of primary and secondary communications uplinks, battery failure, or GPS loss, the failsafe logic of the autopilot will safely return the suas and payload to the ground. By way of contrast, the modified Rascal with a gasoline engine has flight duration in excess of 3 hours with a payload as high as 11 kg. Using the same autopilot and mission planning tools. However, a runway is required and emergency landing situations can be problematic. 2.4 System Portability Figure 1. Mission Planner survey tool plots flight lines base on mission constraints using 2014 Google Earth image Camera trigger options can be programmed from the autopilot via electrical impulse, IR strobe or mechanical servo. The autopilot will log the time, position and orientation of the camera, which is very useful for 3D reconstruction. Although NGA is using Alexmos 2 axis gimbal boards to help stablize the sensor payloads, it is envisioned that the Pixhawk autopilot will be able to drive the gimbal motors directly in near future firmware releases. Mission Planner works on a fixed wing plane, multirotor, helicopter, ground rover and even a boat. This substantially decreases the learning curve required to operate a variety of autonomously navigated vehicles. As part of the mission planning, a user can define predetermined landing zones called rally points. The autopilot will seek the closest point in the event of an emergency landing. If the suas is closest to the launch point, it will automatically return and land there. The mapping mission planned earlier can be flown fully autonomously, only requiring an upward flick of the throttle control to begin the flight. The pilot can sit back and monitor battery and altitude levels in addition to mission progress. The suas will land automatically. The GCS also can be run from a smartphone and or tablet through several open-source apps (DroidPlanner v1, DroidPlanner v2, and Andropilot) which mirror the functionality of the laptop based Mission Planner. Utilizing the GPS on a smartphone or tablet, Portability is another strength of multi-rotor suas. Such a system can be packed away into a single Pelican case and transported easily. In a disaster relief scenario, these multi-rotors have the added advantage of being able to take off and land nearly anywhere. Regardless of platform, maps of the area of interest can be pre-cached and operated offline. 2.5 Small UAV flights The 3D reconstructions discussed later in this paper were produced from a modified GoPro flat lens (4mm) flown on a TBS Discovery quadcopter flown at 125m above ground at a restricted air space site. See Figure 2 for test areas. The GoPro was triggered to take still images via its timelapse function every 5 seconds. With the lower altitudes being flown, the sensor requirements decrease significantly for a desired GSD. The multirotor is unique in the sense that it is the only platform that could be flown to achieve sub-centimeter GSD with a payload being a smartphone. Geotagging the GoPro imagery consisted of syncing the clock on the GoPro through the smartphone app with the phone s UTC time then matching the images according to their timestamp metadata and position logs from the suas autopilot. Due to size, weight, power, and cost limitations, all commercial suas GPS receivers are L1 band only. The autopilot s IMU, gyroscopes, magnometer, and barometer are constantly correcting for positional errors. NGA s hexacopter and octocopters are designed to carry larger payloads. The gimbals for all payloads are designed in CAD and then 3D printed. NGA has been using an Afinia H480 model doi:10.5194/isprsarchives-xl-1-51-2014 52

printer to produce the gimbal and various spare parts such as landing gear from lightweight ABS fillament. Figure 2. Outline of surveyed test areas on 2014 Google Earth image Figure 3. Sparse point cloud over northern test area depicting racetrack mode photos looking down from nadir Five small test areas (Figure 2) containing small buildings and shipping containers arranged in various configurations were GPS surveyed for ground truth. 3.1 Data collection 3. DATA Several photographic test flights using the TBS Discovery quadcopter NGA with modified GoPro lens took place over a two day period. Products have been successfully built from these flights (see next section). The sparse point clouds below show the variety of collection modes used: racetrack (Figure 3), lawnmower (Figure 4) and circular (Figure 5). The camera was positioned looking down from nadir during the racetrack and lawnmower collections and at 45 degree off-nadir for the circular collections. These sparse clouds with camera positions were captured from Visual Structure from Motion (VisualSFM). The racetrack collection is best suited for covering large areas of ground for mapping missions while the lawnmower mode involves photographing a specific area. This mode can image an area for both mapping and point clouds in a short time. The circle mode also focuses on a very specific area but with the camera angled optimally can photographic the area for a dense point cloud. Products from all three modes will be tested to help vet the best collection practices. Figure 6 is a dense point cloud collected from photos from all three flights. This is a VisualSFM point cloud. Figure 4. Sparse point cloud over north-most test area depicting lawnmower mode photos looking down from nadir doi:10.5194/isprsarchives-xl-1-51-2014 53

4.1.1 Tools NGA is currently using two tools (VisualSFM and Agisoft PhotoScan) and will add at least two more (Urban Robotics TerraFlash 3D and Vision Systems, Inc. Probabilistic 3D Modeling a voxel approach) in the assessment. Figure 5. Sparse point cloud over machine shed in the north-most test area depicting several circular flights at 45 degrees from nadir The sparse and dense point clouds (Figures 3 7) seen here are from March flights over the test area. They were created using a structured pipeline in Visual Structure from Motion (VisualSFM)(Wu, 2011a, Wu, et al., 2011b), in conjunction with Clustering Views for Multi-view Stereo (CMVS) (Furukawa, et al., 2010) and Patch-based Multi-view Stereo version 2 (PMVS2)(Furukawa and Ponce 2010). VisualSFM includes the scale-invariant feature transform (SIFT) algorithm, feature detection, feature matching and bundle adjustment. CMVS and PMVS2 are also added into the chain to produce dense point clouds such as the ones you see in Figures 6 and 7. Figure 6 shows the results three collections combined into one point cloud. The tools in this pipeline are available as open source. Figure 7. Dense point cloud created by VisualSFM from single circular flight over machine storage area using the NGA quadcopter with a GoPro flat lens camera Figure 6. Dense point cloud created by VisualSFM, CMVS and PMVS from multiple flights over part of the northern test areas using the NGA quadcopter with a GoPro flat lens camera 4. DATA EXPLOITATION AND PRODUCTS 4.1 Assessments NGA plans to evaluate several open source and commercial tools and their products for ease of use, quality, and timeliness. Specifically, dense point clouds, 3D models, orthomosaics and digital elevation models (DEM) will be closely assessed. It is understood that not all software can make all products. Positional accuracy (absolute and relative), cameras (electro-optical and infrared) and collection techniques are part of the study. Figure 8. Dense point cloud created by Agisoft PhotoScan from single circular flight over machine storage area using the NGA quadcopter with a GoPro flat lens camera doi:10.5194/isprsarchives-xl-1-51-2014 54

Figure 9. Textured 3D model created by Agisoft PhotoScan from single circular flight over machine storage area using the NGA quadcopter with a GoPro flat lens camera Figure 7 is another dense point cloud built from 117 photos taken with the GoPro in a single circular orbit around a machine storage area. PhotoScan is a commercial 3D reconstruction program made by Agisoft. PhotoScan provides a simple step-through process for creating point clouds, polygon models, DEMs, orthophoto mosaics and georeferencing. Figure 8 is the PhotoScan dense point cloud using the same 117 photos. Figure 9 is the same scene as a textured 3D model. Figure 11. Orthophoto mosaic made by CMP-MVS from a NGA quadcopter with a flat lens GoPro camera Another program, CMP-MVS (Jancosek and Pajdla, 2011), is multi-view reconstruction software that works with VisualSFM outputting a textured mesh of a rigid scene. Additionally, CMP- MVS can output digital elevation models and orthophoto mosaics. See Figures 10 and 11. The other tools will be discussed in the Future Plans section below. 4.1.2 Products NGA will specifically look at dense point clouds, 3D model reconstructions, DEMs and orthophoto mosaics. Georeferencing is an important part of the assessment plan. Placing the product as close as possible to the right location on the earth is necessary to the users of the products. 4.1.3 Cameras and collection techniques Figure 10. Digital Elevation Model made by CMP-MVS from a NGA quadcopter with a flat lens GoPro camera The GoPro Hero 3 Black has both frame and video modes. Frame mode photos have higher resolution than the frames taken from the videos. Video frames are taken at a higher rate closer together and fit better into the multi-view stereo collections but their resolutions are lower. Preference of photo mode will be dependent on mission. Videos work well with circle mode collects with the camera off nadir but frames are more suitable to racetrack/lawnmower collects at nadir. doi:10.5194/isprsarchives-xl-1-51-2014 55

A mobile phone camera was used early in the flight testing and found to be suitable for use but did not have the resolution as the other low-cost commercial grade cameras used. Initial ground testing of one of these low-cost cameras provided good photos for these products but was found to not be suitable for flight on the copters due to limited photo triggering mechanism that could be only used by pushing down on the photo button. A mechanical device could be installed but it would impact the weight and electrical requirements. New commercial grade cameras and another GoPro camera were recently purchased but will be discussed in the Future plans section below. Lastly, collection techniques are important to the reconstructions, DEMs and orthos. With the first two tests the camera was pointed down from nadir and pointed 45 degrees from nadir for the circle test. It is important that collections be geared toward photographing not only horizontal (rooftops, roads, etc.) but also the vertical surfaces such as sides of features in order to have as much data as possible to create a 3D model. The dense point clouds over the machine sheds in the racetrack and lawnmower modes were less complete than the ones in the circle mode. So a collection strategy will be developed based on a Matlab simulation by Ekholm (2012) that aims the cameras at various degrees from nadir. 5.1 Fixed wing Rascal 5. FUTURE PLANS The Rascal autopilot and engine has been retrofitted to APM 2.6 and to gasoline.. This will increase the flight endurance from 45 minutes to 3 hours, cut operational costs (1 gallon of Model RC fuel costs approximately $20) and increase safety with the APM autopilot. The Rascal can easily reach altitudes in excess of 1500m where the multirotors will only spend seconds before draining their Li-Po batteries. The Rascal is capable of lifting payloads up to 11kg. Large area coverage with heavier payloads will be tested. 5.2 Updated positional accuracy system A newly implemented feature of the Pixhawk autopilot system is the implementation of a extended Kalman Filter algorithm to improve positional accuracy of the suas. The accuracy will be tested against ground truth. 2 can program the pulse-width modulation signals to cause the IR strobe to trigger different functionalities on the camera, i.e. start and stop video. Future payloads also include a FLIR Tau640 LWIR camera and an Ohio State University owned Velodyne LiDAR system. 5.5 New product software A CRADA between NGA and Urban Robotics will explore new methods of processing sensor data to increase the development of 3D renderings and maps from suav imagery. Probabilistic Volumetric Representation (PVR), a voxel approach, will be tested this fiscal year. Other DEM and orthomosaic generated software accuracy studies are planned. A software program for georeferencing and scaling of point clouds, SfM_georef, by Mike James of Lancaster University, supports VisualSFM. It will also be included in the study. REFERENCES Ekholm, J., 3-D Scene Reconstruction from Aerial Imagery, M.S. thesis, Air Force Institute of Technology Air University, Dayton, 168 p. Furukawa, Y., Curless, B., Seitz, S., Szeliski, R., 2010, Towards Internet-scale Multi-view Stereo, CVPR, 2010 IEEE Conference. Furukawa, Y. and Ponce, J., 2010, Accurate, Dense, and Robust Multi-View Stereopsis. IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 2, Issue 8. Jancosek, M. and Pajdla, T., 2011, Multi-View Reconstruction Preserving Weakly-Supported Surfaces, CVPR 2011 IEEE Conference on Computer Vision and Pattern Recognition 2011. Wu, C., 2011a. VisualSFM: A Visual Structure from Motion System, version 0.5, URL: http://ccwu.me/vsfm, (last accessed 14 August 2014) Wu, C., Agarwal, S., Curless, B., Seitz, S., 2011b. Multicore Bundle Adjustment, CVPR. 5.3 Refined collection techniques Alternative collection paradigms will be compared against survey data for a variety of GEOINT products based on Ekholm s thesis on airborne collection profiles. 5.4 New cameras A new commercial grade camera with 16-55mm lens has been ground tested but not flown. This camera is triggered via IR strobe through a device called Strato Snapper 2. The Strato Snapper 2 accepts pulse-width-modulations signals from the Pixhawk or APM and triggers the IR strobe. The Strato Snapper doi:10.5194/isprsarchives-xl-1-51-2014 56