Agisoft Metashape User Manual: Standard Edition, Version 1.7
Agisoft Metashape User Manual: Standard Edition, Version 1.7
Agisoft Metashape User Manual: Standard Edition, Version 1.7
iii
Agisoft Metashape User Manual
iv
Overview
Agisoft Metashape is a stand-alone software product that performs photogrammetric processing of digital
images (aerial and close-range photography) and generates 3D spatial data to be used in GIS applications,
cultural heritage documentation, and visual effects production as well as for indirect measurements of
objects of various scales.
The software allows to process images from RGB, thermal into the spatial information in the form of dense
point clouds, textured polygonal models. Wisely implemented digital photogrammetry technique enforced
with computer vision methods results in smart automated processing system that, on the one hand, can be
managed by a new-comer in the field of photogrammetry, yet, on the other hand, has a lot to offer to a
specialist who can benefit from advanced features like stereoscopic mode and have complete control over
the results accuracy, with detailed report being generated at the end of processing.
How it works
Typical tasks for a photogrammetry processing project in Metashape is to build a textured 3D model.
Imagery data processing procedure with Agisoft Metashape consists of two main steps.
1. The first step is called alignment. It includes aerial triangulation (AT) and bundle block adjustment
(BBA). At this stage Metashape searches for feature points on the images and matches them across
images into tie points. The program also finds the position of the camera for each image and
refines camera calibration parameters (estimates internal (IO) and external (EO) camera orientation
parameters).
The results of these procedures are visualized in the form of a sparse point cloud and a set of camera
positions. The sparse point cloud represents the results of image alignment and will not be directly used
in further processing (except for the sparse point cloud based surface reconstruction method, which
is suitable only for quick estimates, e.g., of completeness of a data set). But the sparse point cloud
is necessary for the determination of depth maps (based on the sparse cloud selected stereo pairs).
However it can be exported for further usage in external programs. For instance, a sparse point cloud
model can be used in a 3D editor as a reference. On the contrary, the set of camera positions is required
for further 3D surface reconstruction by Metashape.
2. The second step is generation of a surface in 3D (mesh) . Polygonal model (mesh) can be textured for
photorealistic digital representation of the object/scene and exported in numerous formats compatible
with post-processing software, both for CAD and 3D-modeling workflows.
Dense point cloud can be built by Metashape based on the estimated camera positions and images
themselves (dense stereo matching). Generated photogrammetric point cloud can be merged with
LIDAR data.
In some cases, however, additional actions may be required to get the desired results. Pictures taken using
uncommon lenses such as fisheye one may require preliminary calibration of optical system parameters
or usage of different calibration model specially implemented for ultra-wide angle lens. Metashape
v
Overview
enables to reestimate extrinsic and intrinsic camera parameters, optimizing them for a tie point pool
preliminary filtered by user. Chapter 4, Improving camera alignment results covers that part of the
software functionality. In some capturing scenarios masking of certain regions of the photos may be
required to exclude them from the calculations. Application of masks in Metashape processing workflow
as well as editing options available are described in Chapter 5, Editing. Chapter 6, Automation describes
opportunities to save up on manual intervention to the processing workflow.
It can take up quite a long time to reconstruct a 3D model. Metashape allows to export obtained results
and save intermediate data in a form of project files at any stage of the process. If you are not familiar with
the concept of projects, its brief description is given at the end of the Chapter 3, General workflow.
In the manual you can also find instructions on the Metashape installation and activation procedures and
basic rules for taking "good" photographs, i.e. pictures that provide most necessary information for 3D
reconstruction. For the information refer to Chapter 1, Installation and Activation and Chapter 2, Capturing
scenarios.
vi
Chapter 1. Installation and Activation
System requirements
Minimal configuration
• Windows 7 SP 1 or later (64 bit), Windows Server 2008 R2 or later (64 bit), macOS High Sierra or
later, Debian/Ubuntu with GLIBC 2.19+ (64 bit)
• 4 GB of RAM
Recommended configuration
• Windows 7 SP 1 or later (64 bit), Windows Server 2008 R2 or later (64 bit), macOS Mojave or later,
Debian/Ubuntu with GLIBC 2.19+ (64 bit)
• 32 GB of RAM
The number of photos that can be processed by Metashape depends on the available RAM and
reconstruction parameters used. Assuming that a single photo resolution is of the order of 10 MPix, 4 GB
RAM is sufficient to make a model based on 30 to 50 photos. 16 GB RAM will allow to process up to
300-400 photographs.
GPU recommendations
Metashape supports accelerated image matching; depth maps reconstruction; depth maps based mesh
model generation; texture blending; photoconsistent mesh refinement operation due to the graphics
hardware (GPU) exploiting.
NVIDIA
GeForce GTX 7xx series and later with CUDA support.
AMD
Radeon R9 series and later with OpenCL 1.2 support.
Metashape is likely to be able to utilize processing power of any CUDA enabled device with compute
capability 3.0 and higher or OpenCL 1.2 and higher enabled device with SPIR support for stages specified
above, provided that CUDA/OpenCL drivers for the device are properly installed. However, because of
the large number of various combinations of video chips, driver versions and operating systems, Agisoft
is unable to test and guarantee Metashape's compatibility with every device and on every platform.
The processing performance of the GPU device is mainly related to the number of CUDA cores for
NVIDIA video chips and the number of shader processor units for AMD and Intel video chips. Additionally
depth maps based mesh model reconstruction as well as photoconsistent mesh refinement operations and
texture blending would benefit from larger amount of VRAM available.
1
Installation and Activation
The table below lists currently supported devices (on Windows platform only). Agisoft will pay particular
attention to possible problems with Metashape running on these devices.
Metashape supports texture blending on GPU using Vulkan technology on Linux and Windows OS. GPU
accelerated texture blending is currently supported for frame and fisheye type cameras on NVIDIA cards
since GeForce GTX 8XX / Quadro M4000 and driver versions from 435.xx and on AMD cards since
Radeon R9 29x series / FirePro W9100 and 17.1.x drivers. Some older GPUs and older driver versions
could also support texture blending using Vulkan, however, it is not guaranteed.
Although Metashape is supposed to be able to utilize other compatible GPU models and being run under
a different operating system, Agisoft does not guarantee that it will work correctly. However, all GPU-
based processing issues should be reported to Agisoft support team for more detailed investigation.
Note
• Use CPU enable flag to allow calculations both on CPU and GPU for GPU-supported tasks.
However if at least one powerful discrete GPU is used it is recommended to disable CPU flag
for stable and rapid processing.
• Using GPU acceleration with mobile or integrated graphics video chips is not recommended
because of the low performance of such GPUs.
• CUDA supported devices for some older macOS versions may require to install CUDA drivers
from official web-site first: http://www.nvidia.com/object/mac-driver-archive.html.
Due to lack of CUDA support on certain macOS versions Metashape will automatically switch
to OpenCL implementation for GPU-based processing on NVIDIA graphic devices.
Installation procedure
Installing Metashape on Microsoft Windows
To install Metashape on Microsoft Windows simply run the downloaded msi file and follow the
instructions.
2
Installation and Activation
If you are not ready yet to start the trial period, you can opt for the Demo mode. The employment of
Metashape in the Demo mode is not time limited. Several functions, however, are not available in the
Demo mode. These functions are the following:
• all export features, including exporting reconstruction results (you can only view a 3D model on the
screen);
To use Metashape in the full function mode for various projects you have to purchase a license. On
purchasing you will get a license code to be entered into the activation dialog of Metashape. Once the
license code is entered you will get full access to all functions of the program and the activation dialog
will no longer appear upon program start, unless the license is deactivated.
Activation procedure
Metashape license activation
Metashape software requires license key (a digital code) to be activated. First of all, make sure that you
have a valid license key or a trial code at hand.
To activate Metashape
1. Launch Metashape software, previously installed on your machine, and go to Help menu for Activate
product... command.
2. In Activation dialog insert license key according to the suggested 5 digit blocks structure. Please note
that license codes does never include zero digit - only letter "O".
3. If the license code has been input correctly, then the OK button will become active. Click on it to
complete the activation procedure. If the button is still grayed out, please make sure that the key you
3
Installation and Activation
are using is meant for the product you are trying to activate: a license key for the Professional Edition,
for example, will not activate the Standard version of the software.
Note
• The node-locked license activation on Windows OS and macOS may require administrator
privileges. During the activation process additional confirmation dialog will appear to apply the
elevated privileges.
4
Chapter 2. Capturing scenarios
Photographs suitable for 3D model reconstruction in Metashape can be taken by any digital camera (both
metric and non-metric), as long as you follow some specific capturing guidelines. This section explains
general principles of taking and selecting pictures that provide the most appropriate data for 3D model
generation.
IMPORTANT! Make sure you have studied the following rules and read the list of restrictions before you
get out for shooting photographs.
Equipment
• Use a digital camera with reasonably high resolution (5 MPix or more).
• Avoid ultra-wide angle and fisheye lenses. The best choice is 50 mm focal length (35 mm film
equivalent) lenses. It is recommended to use focal length from 20 to 80 mm interval in 35mm equivalent.
If a data set was captured with fisheye lens, appropriate camera sensor type should be selected in
Metashape Camera Calibration dialog prior to processing.
• Fixed lenses are preferred. If zoom lenses are used - focal length should be set either to maximal or to
minimal value during the entire shooting session for more stable results, for intermediate focal lengths
separate camera calibration groups should be used.
Camera settings
• Using RAW data losslessly converted to the TIFF files is preferred, since JPG compression may induce
unwanted noise to the images.
• ISO should be set to the lowest value, otherwise high ISO values will induce additional noise to images.
• Aperture value should be high enough to result in sufficient focal depth: it is important to capture sharp,
not blurred photos.
• Shutter speed should not be too slow, otherwise blur can occur due to slight movements.
Object/scene requirements
• Avoid not textured, shiny, highly reflective or transparent objects.
Image preprocessing
• Metashape operates with the original images. So do not crop or geometrically transform, i.e. resize or
rotate, the images.
5
Capturing scenarios
Capturing scenarios
Generally, spending some time planning your shot session might be very useful.
• Number of "blind-zones" should be minimized since Metashape is able to reconstruct only geometry
visible from at least two cameras.
• Each photo should effectively use the frame size: object of interest should take up the maximum area.
In some cases portrait camera orientation should be used.
• Do not try to place full object in the image frame, if some parts are missing it is not a problem providing
that these parts appear on other images.
• Good lighting is required to achieve better quality of the results, yet blinks should be avoided. It is
recommended to remove sources of light from camera fields of view. Avoid using flash.
6
Capturing scenarios
Restrictions
In some cases it might be very difficult or even impossible to build a correct 3D model from a set of
pictures. A short list of typical reasons for photographs unsuitability is given below.
Modifications of photographs
Metashape can process only unmodified photos as they were taken by a digital photo camera. Processing
the photos which were manually cropped or geometrically warped is likely to fail or to produce highly
inaccurate results. Photometric modifications do not affect reconstruction results.
The details of necessary EXIF tags and instructions for manual setting of the calibration parameters are
given in the Camera calibration section.
Lens distortion
The distortion of the lenses used to capture the photos should be well simulated with the camera model
used in the software. Generally, Brown's distortion model implemented in Metashape works well for frame
cameras. However, since fisheye/ultra-wide angle lenses are poorly simulated by the mentioned distortion
model, it is crucial to choose proper camera type in Camera Calibration dialog prior to processing of such
data - the software will switch to the appropriate distortion model.
Lens calibration
It is possible to use Metashape for automatic lens calibration. Metashape uses LCD screen as a calibration
target (optionally it is possible to use a printed chessboard pattern, providing that it is flat and all its cells are
7
Capturing scenarios
squares). Lens calibration procedure supports estimation of the full camera calibration matrix, including
non-linear distortion coefficients. The details of camera models are given in the Appendix C, Camera
models section.
Note
• Lens calibration procedure can usually be skipped in common workflow, as Metashape calculates
the calibration parameters automatically during Align Photos process. However, if the alignment
results are unstable, for example, due to the lack of the tie points between the images, the lens
calibration may be useful.
f
Focal length measured in pixels.
cx, cy
Principal point coordinates, i.e. coordinates of lens optical axis interception with sensor plane in pixels.
b1, b2
Affinity and Skew (non-orthogonality) transformation coefficients.
p1, p2
Tangential distortion coefficients.
Before using lens calibration tool a set of photos of calibration pattern should be loaded in Metashape.
1. Select Show Chessboard... command from the Lens submenu in the Tools menu to display the
calibration pattern.
2. Use mouse scroll wheel to zoom in/out the calibration pattern. Scale the calibration pattern so that
the number of squares on each side of the screen would exceed 10.
3. Capture a series of photos of the displayed calibration pattern with your camera from slightly different
angles, according to the guidelines, outlined below. Minimum number of photos for a given focal
length is 3.
4. If you are calibrating zoom lens, change the focal length of your lens and repeat the previous step
for other focal length settings.
5. Click anywhere on the calibration pattern or press Escape button to return to the program.
When capturing photos of the calibration pattern, try to fulfill the following guidelines:
• Make sure that the focal length keeps constant throughout the session (in case of zoom lens).
• Avoid glare on the photos. Move the light sources away if required.
8
Capturing scenarios
• Preferably, the whole area of the photos should be covered by calibration pattern. Move the camera
closer to the LCD screen if required.
3. In the Open dialog box, browse to the folder, containing the photos, and select files to be processed.
Then click Open button.
Note
• You can open any photo by double clicking on its thumbnail in the Photos pane. To obtain good
calibration, the photos should be reasonably sharp, with crisp boundaries between cells.
• If you have loaded some unwanted photos, you can easily remove them at any time.
• Before calibrating fisheye lens you need to set the corresponding Camera Type in the Camera
Calibration... dialog available from the Tools menu. See information on other camera calibration
settings in Camera calibration section.
2. In the Calibrate Lens dialog box, select the desired calibration parameters. Click OK button when
done.
3. The progress dialog box will appear displaying the current processing status. To cancel processing
click the Cancel button.
4. The calibration results will appear on the Adjusted tab of the Camera Calibration... dialog available
from the Tools menu. The adjusted values can be saved to file by using Save button on the Adjusted
tab. The saved lens calibration data can later be used in another chunk or project, providing that the
same camera and lens is used.
Note
• After you have saved the calibration parameters for the lens, you may proceed with the workflow
steps in a separate chunk for the actual image set captured by the same camera and lens. To
protect the calibration data from being refined during Align Photos process one should check Fix
calibration box on the Initial tab for the chunk with the data to be processed. In this case initial
calibration values will not be changed during Align Photos process.
After calibration is finished, you will be presented with the following information:
Detected chessboard corners are displayed on each photo (the photo can be opened by double clicking on
its name in the Photos pane). It is preferable when the majority of the corners were detected correctly. For
9
Capturing scenarios
each detected corner the reprojection error between the detected corner position and estimated position
according to the calculated calibration is also displayed. The errors are scaled x20 times for display.
1. Align photos using entire dataset and build rough mesh model from the sparse cloud.
4. Click OK button.
5. The progress dialog box will appear displaying the current processing status. To cancel processing
click the Cancel button.
6. After the operation is finished all the cameras which are not included into optimal subset will get
disabled.
Focus on selection
To consider only selected triangles of the polygonal model as target for reconstruction. Cameras that
do not have any of the selected polygons in the field of view would be automatically disabled.
Surface coverage
Number of cameras observing each point from different angles.
10
Chapter 3. General workflow
Processing of images with Metashape includes the following main steps:
• aligning cameras;
• generating texture;
• exporting results.
If you are using Metashape in the full function (not the Demo) mode, intermediate results of the image
processing can be saved at any stage in the form of project files and can be used later. The concept of
projects and project files is briefly explained in the Saving intermediate results section.
The list above represents all the necessary steps involved in the construction of a textured 3D model from
your photos. Some additional tools, which you may find to be useful, are described in the successive
chapters.
Preferences settings
Before starting a project with Metashape it is recommended to adjust the program settings for your needs.
In Preferences dialog (General tab) available through the Tools menu you can indicate the path to the
Metashape log file to be shared with the Agisoft support team in case you face any problem during the
processing. Here you can also change GUI language to the one that is most convenient for you. The options
are: English, Chinese, French, German, Italian, Japanese, Portuguese, Russian, Spanish.
Switch Theme in case you have preferences between Dark or Light program GUI or leave it as Classic
for the simplest view. Shortcuts for the menu commands can be adjusted for your convenience on the
General tab as well.
On the GPU tab you need to make sure that all discrete GPU devices detected by the program are checked.
Metashape exploits GPU processing power that speeds up the process significantly. However, Agisoft
does not recommend to use integrated graphic card adapters due to their possible unstable work under
heavy load. If you have decided to switch on GPUs to boost the data processing with Metashape, it is
recommended to uncheck Use CPU when performing GPU accelerated processing option, providing that
at least one discrete GPU is enabled for processing.
Advanced tab allows to switch on such advanced features like rich Python console, for example.
Furthermore, you can enable loading of extra camera data from XMP (camera calibration).
Keep depth maps option can be beneficial in terms of saving up the processing time, in case there might
be a need to rebuild dense point cloud, once generated, for a smaller part of the scene, or if both mesh and
dense cloud are based on the same quality depth maps.
Fine-level task subdivision option is useful in cases when large datasets are to be processed. Enabled option
provides internal splitting of some tasks to the sub-jobs, thus allowing to reduce the memory consumption
11
General workflow
during the processing. The tasks that are supported for fine-level distribution are the following: Match
Photos, Align Cameras, Build Depth Maps, Build Dense Cloud.
Metashape allows for incremental image alignment, which may be useful in case of some data missing in
the initially aligned project. If this may be the case, you should switch on the Keep key points option on
the Advanced tab of the Preferences dialog before you start the processing of the data.
Loading images
Before starting any operation it is necessary to point out which images will be used as a source for
photogrammetric processing. In fact, images themselves are not loaded into Metashape until they are
needed. So, when Add photos command is used, only the links to the image files are added to the project
contents to indicate the images that will be used for further processing.
Metashape uses the full color range for image matching operation and is not downsampling the color
information to 8 bit. The point cloud points and texture would also have the original bit depth, providing
that they are exported in the formats that support non 8-bit colors.
1.
Select Add Photos... command from the Workflow menu or click Add Photos toolbar button on
the Workspace pane.
2. In the Add Photos dialog box browse to the folder containing the images and select files to be
processed. Then click Open button.
Note
• Metashape accepts the following image formats: JPEG, JPEG 2000, TIFF, DNG, PNG,
OpenEXR, BMP, TARGA, PPM, PGM, SEQ, ARA (thermal images) and JPEG Multi-Picture
Format (MPO). Image files in any other format will not be displayed in the Add Photos dialog
box. To work with such images is it necessary to convert them to one of the supported formats.
If some unwanted images have been added to the project, they can be easily removed at any moment.
2. Right-click on the selected cameras and choose Remove Items command from the opened context
menu, or click Remove Items toolbar button on the Workspace pane. The related images will be
removed from the working set.
Camera groups
If all the photos or a subset of photos were captured from one camera position - camera station, for
Metashape to process them correctly it is obligatory to move those photos to a camera group and mark the
group as Camera Station. It is important that for all the photos in a Camera Station group distances between
camera centers were negligibly small compared to the camera-object minimal distance. Photogrammetric
12
General workflow
processing will require at least two camera stations with overlapping photos to be present in a chunk.
However, it is possible to export panoramic picture for the data captured from only one camera station.
Refer to Exporting results section for guidance on panorama export.
Alternatively, camera group structure can be used to manipulate the image data in a chunk easily, e.g. to
apply disable/enable functions to all the cameras in a group at once.
1. On the Workspace pane (or Photos pane) select the photos to be moved.
2. Right-click on the selected photos and choose Move Cameras - New Camera Group command from
the opened context menu.
3. A new group will be added to the active chunk structure and selected photos will be moved to that
group.
4. Alternatively selected photos can be moved to a camera group created earlier using Move Cameras
- Camera Group - Group_name command from the context menu.
To mark group as camera station, right click on the camera group name and select Set Group Type
command from the context menu.
NC (Not calibrated)
Notifies that the EXIF data available is not sufficient to estimate the camera focal length. In this case
Metashape assumes that the corresponding photo was taken using 50mm lens (35mm film equivalent).
If the actual focal length differs significantly from this value, manual calibration may be required.
More details on manual camera calibration can be found in the Camera calibration section.
NA (Not aligned)
Notifies that external camera orientation parameters have not been estimated for the current image yet.
Images loaded to Metashape will not be aligned until you perform the next step - photos alignment.
Video Data
Metashape allows for video data processing as well, which can be beneficial for quick inspection scenarios,
for example. The video is to be divided into frames which will be further used as source images for 3D
reconstruction.
13
General workflow
2. In the Import Video dialog you can inspect the video and set the output folder for the frames.
3. Set the filename pattern for the frames and indicate the frame extraction rate.
4. You can import part of the video, specify the parameters: Start from and End at.
5. Click OK button for the frames to be automatically extracted and saved to the designated folder. The
images extracted from the video will be automatically added to the active chunk.
Note
• In Metashape you can choose the automatic frame step (Small, Medium, Large) which may be
helpful to skip similar sequential frames or set manually via Custom option. Once the parameter
value is set, the program calculates the shift for the images to be captured. For Small value, the
shift of about 3% of the image width will be taken into account. For Medium, it corresponds to
7% and for Large - 14% of the image width.
After the frames have been extracted you can follow standard processing workflow for the images.
Aligning photos
The camera position at the time of image capture is defined by the interior and exterior orientation
parameters.
Interior orientation parameters include camera focal length, coordinates of the image principal point and
lens distortion coefficients. Before starting processing in Metashape the following configuration steps
should be performed:
• Separate calibration groups should be created for each physical camera used in the project. It is also
recommended to create a separate calibration group for each flight or survey. For details see Camera
groups subsection of Loading images.
• For each calibration group initial approximation of interior orientation parameters should be specified.
In most cases this is done automatically based on EXIF meta data. When EXIF meta data is not available,
initial interior orientation parameters needs to be configured according to the camera certificate.
Exterior orientation parameters define the position and orientation of the camera. They are estimated during
image alignment and consist of 3 translation components and 3 Euler rotation angles.
Exterior and interior image orientation parameters are calculated using aerotriangulation with bundle block
adjustment based on collinearity equations.
The result of this processing step consists of estimated exterior (translation and rotation) and interior
camera orientation parameters together with a sparse point cloud containing triangulated positions of
matched image points.
2. In the Align Photos dialog box select the desired alignment options. Click OK button when done.
3. The progress dialog box will appear displaying the current processing status. To cancel processing
click Cancel button.
14
General workflow
Alignment having been completed, computed camera positions and a sparse point cloud will be displayed.
You can inspect alignment results and remove incorrectly positioned photos, if any. To see the matches
between any two photos use View Matches... command from a photo context menu in the Photos pane.
1. Reset alignment for incorrectly positioned cameras using Reset Camera Alignment command from
the photo context menu.
2. Select photos to be realigned and use Align Selected Cameras command from the photo context menu.
3. The progress dialog box will appear displaying the current processing status. To cancel processing
click Cancel button.
When the alignment step is completed, the point cloud and estimated camera positions can be exported
for processing with another software if needed.
Image quality
Poor input, e. g. vague photos, can influence alignment results badly. To help you to exclude poorly
focused images from processing Metashape suggests automatic image quality estimation feature. Images
with quality value of less than 0.5 units are recommended to be disabled and thus excluded from
photogrammetric processing, providing that the rest of the photos cover the whole scene to be
reconstructed. To disable a photo use Disable button from the Photos pane toolbar.
Metashape estimates image quality for each input image. The value of the parameter is calculated based
on the sharpness level of the most focused part of the picture.
1.
Switch to the detailed view in the Photos pane using Details command from the Change menu
on the Photos pane toolbar.
3. Right button click on the selected photo(s) and choose Estimate Image Quality command from the
context menu.
4. Once the analysis procedure is over, a figure indicating estimated image quality value will be
displayed in the Quality column on the Photos pane.
Alignment parameters
15
General workflow
The following parameters control the photo alignment procedure and can be modified in the Align Photos
dialog box:
Accuracy
Higher accuracy settings help to obtain more accurate camera position estimates. Lower accuracy
settings can be used to get the rough camera positions in a shorter period of time.
While at High accuracy setting the software works with the photos of the original size, Medium setting
causes image downscaling by factor of 4 (2 times by each side), at Low accuracy source files are
downscaled by factor of 16, and Lowest value means further downscaling by 4 times more. Highest
accuracy setting upscales the image by factor of 4. Since tie point positions are estimated on the
basis of feature spots found on the source images, it may be meaningful to upscale a source photo
to accurately localize a tie point. However, Highest accuracy setting is recommended only for very
sharp image data and mostly for research purposes due to the corresponding processing being quite
time consuming.
Generic preselection
The alignment process of large photo sets can take a long time. A significant portion of this time
period is spent on matching of detected features across the photos. Image pair preselection option may
speed up this process due to selection of a subset of image pairs to be matched.
In the Generic preselection mode the overlapping pairs of photos are selected by matching photos
using lower accuracy setting first.
Reference preselection
The Estimated preselection mode takes into account the calculated exterior orientation parameters
for the aligned cameras. That is, if the alignment operation has been already completed for the project,
the estimated camera locations will be considered when the Align Photos procedure is run again with
the Estimated preselection selected.
16
General workflow
When using Sequential preselection mode the correspondence between the images is determined
according to the sequence of photos (the sequence number of the image) it is worth noting that with
this adjustment, the first with the last images in the sequence will also be compared.
Apply mask to
If apply mask to key points option is selected, areas previously masked on the photos are excluded
from feature detection procedure. Apply mask to tie points option means that certain tie points are
excluded from alignment procedure. Effectively this implies that if some area is masked at least on a
single photo, relevant key points on the rest of the photos picturing the same area will be also ignored
during alignment procedure (a tie point is a set of key points which have been matched as projections
of the same 3D point on different images). This can be useful to be able to suppress background in
turntable shooting scenario with only one mask. For additional information on the usage of masks
please refer to the Using masks section.
Note
• Tie point limit parameter allows to optimize performance for the task and does not generally
effect the quality of the further model. Recommended value is 4000. Too high or too low tie point
17
General workflow
limit value may cause some parts of the dense point cloud model to be missed. The reason is that
Metashape generates depth maps only for pairs of photos for which number of matching points is
above certain limit. This limit equals to 100 matching points, unless moved up by the figure "10%
of the maximum number of matching points between the photo in question and other photos,
only matching points corresponding to the area within the bounding box being considered."
• The number of tie points can be reduced after the alignment process with Tie Points - Thin Point
Cloud command available from Tools menu. As a results sparse point cloud will be thinned, yet
the alignment will be kept unchanged.
Components
Some image subsets may be not aligned as the result of the Align Photos operation, if the sufficient amount
of tie points was not detected between such subsets. Such subsets which are not aligned with the main
subset will be grouped into individually aligned parts - Components.
The components into which the camera alignment have been split after the Align Photos operation will be
displayed in the chunk's contents of the Workspace pane inside the Components folder.
To subalign some extra images added to the chunk with already aligned set of
images
1. Add extra photos to the active chunk using Add photos command from the Workflow menu.
3. Set alignment parameters for the newly added photos. IMPORTANT! Uncheck Reset alignment
option.
4. Click OK. Metashape will consider existing key points and try to match them with key points detected
on the newly added images.
4. The data will be loaded into the software. Camera calibration data can be inspected in the Camera
Calibration dialog, Adjusted tab, available from Tools menu.
18
General workflow
Camera data can be loaded in one of the following formats: Agisoft (*.xml), BINGO (*.dat), Inpho Project
File (*.prj), Blocks Exchange (*.xml), Bundler (*.out), Autodesk FBX (*.fbx), VisionMap Detailed Report
(*.txt), Realviz RZML (*.rzml), Alembic (*.abc).
Once the data is loaded, Metashape will offer to build point cloud. This step involves feature points
detection and matching procedures. As a result, a sparse point cloud - 3D representation of the tie-points
data, will be generated. Build Point Cloud command is available from Tools - Tie Points menu. Parameters
controlling Build Point Cloud procedure are the same as the ones used at Align Photos step (see above).
Dense point cloud generation is based on depth maps calculated using dense stereo matching. Depth maps
are calculated for the overlapping image pairs considering their relative exterior and interior orientation
parameters estimated with bundle adjustment. Multiple pairwise depth maps generated for each camera
are merged together into combined depth map, using excessive information in the overlapping regions to
filter wrong depth measurements.
Combined depth maps generated for each camera are transformed into the partial dense point clouds,
which are then merged into a final dense point cloud with additional noise filtering step applied in the
overlapping regions. The normals in the partial dense point clouds are calculated using plane fitting to the
pixel neighborhood in the combined depth maps, and the colors are sampled from the images.
For every point in the final dense point cloud the number of contributing combined depth maps is recorded
and stored as a confidence value. This confidence value can be used later to perform additional filtering of
low confidence points using the Filter by Confidence... command from the Tools > Dense Cloud menu.
Metashape tends to produce extra dense point clouds, which are of almost the same density, if not denser,
as LIDAR point clouds. A dense point cloud can be edited within Metashape environment and used as
a basis for such processing stages as Build Mesh. Alternatively, the point cloud can be exported to an
external tool for further analysis.
2. Select the Build Dense Cloud... command from the Workflow menu.
3. In the Build Dense Cloud dialog box select the desired reconstruction parameters. Click OK button
when done.
4. The progress dialog box will appear displaying the current processing status. To cancel processing
click Cancel button.
Note
• More than one instance of Dense cloud can be stored in one chunk. In case you want to save
current Dense cloud instance and build new one in current chunk, right-click on Dense cloud and
19
General workflow
uncheck Set as default option. In case you want to save current Dense cloud instance and edit its
copy, right-click on Dense cloud and choose Duplicate option.
Quality
Specifies the desired quality of the depth maps genration. Higher quality settings can be used to obtain
more detailed and accurate geometry, but they require longer time for processing. Interpretation of the
quality parameters here is similar to that of accuracy settings given in Photo Alignment section. The
only difference is that in this case Ultra High quality setting means processing of original photos, while
each following step implies preliminary image size downscaling by factor of 4 (2 times by each side).
If there are important small details which are spatially distinguished in the scene to be
reconstructed, then it is recommended to set Mild depth filtering mode, for important features
not to be sorted out as outliers. This value of the parameter may also be useful for aerial projects
in case the area contains poorly textured roofs, for example. Mild depth filtering mode is also
required for the depth maps based mesh reconstruction.
If the area to be reconstructed does not contain meaningful small details, then it is reasonable
to choose Aggressive depth filtering mode to sort out most of the outliers. This value of the
parameter normally recommended for aerial data processing, however, mild filtering may be
useful in some projects as well (see poorly textured roofs comment in the mild parameter value
description above).
Moderate depth filtering mode brings results that are in between the Mild and Aggressive
approaches. You can experiment with the setting in case you have doubts which mode to choose.
Additionally depth filtering can be Disabled. But this option is not recommended as the resulting
dense cloud could be extremely noisy.
20
General workflow
The filtering modes control noise filtering in the raw depth maps. This is done using a connected
component filter which operates on segmented depth maps based on the pixel depth values. The
filtering preset control a maximum size of connected components that are discarded by the filter.
Note
• Stronger filter presets remove more noise, but also may remove useful information in case there
are small and thin structures in the scene.
Dense point cloud can be imported in one of the following formats: Wavefront OBJ, Stanford PLY, ASPRS
LAS, LAZ, ASTM E57, ASCII PTS.
Building mesh
Mesh based on point cloud or depth maps data
Metashape can reconstruct polygonal mesh model based on the point cloud information (Dense Cloud,
Sparse Cloud, Point Cloud imported from external source) or based on the depth maps data.
To build a mesh
1. Check the reconstruction volume bounding box. If the model has already been referenced, the
bounding box will be properly positioned automatically. Otherwise, it is important to control its
position manually.
To adjust the bounding box manually, use the Resize Region, Move Region and Rotate
Region toolbar buttons. Rotate the bounding box and then drag corners of the box to the desired
positions - only part of the scene inside the bounding box will be reconstructed. If the Height field
reconstruction method is to be applied, it is important to control the position of the red side of the
bounding box: it defines reconstruction plane. In this case make sure that the bounding box is correctly
oriented.
21
General workflow
3. In the Build Mesh dialog box select the desired reconstruction parameters. Click OK button when
done.
4. The progress dialog box will appear displaying the current processing status. To cancel processing
click Cancel button.
Note
• More than one instance of Mesh can be stored in one chunk. In case you want to save current
Mesh instance and build new one in current chunk, right-click on Mesh and uncheck Set as
default option. In case you want to save current Mesh instance and edit its copy, right-click on
Mesh and choose Duplicate option.
Metashape supports several reconstruction methods and settings, which help to produce optimal
reconstructions for a given data set.
Source data
Specifies the source for the mesh generation procedure.
• Sparse cloud can be used for fast 3D model generation based solely on the sparse point cloud.
• Dense cloud setting will result in longer processing time but will generate high quality output based
on the previously reconstructed dense point cloud.
• Depth maps setting allows to use all the information from the input images more effectively
and is less resource demanding compared to the dense cloud based reconstruction. The option
is recommended to be used for Arbitrary surface type reconstruction, unless the workflow used
assumes dense cloud editing prior to the mesh reconstruction.
22
General workflow
Surface type
• Arbitrary surface type can be used for modeling of any kind of object. It should be selected for
closed objects, such as statues, buildings, etc. It doesn't make any assumptions on the type of the
object being modeled, which comes at a cost of higher memory consumption.
• Height field surface type is optimized for modeling of planar surfaces, such as terrains or
basereliefs. It should be selected for aerial photography processing as it requires lower amount of
memory and allows for larger data sets processing.
Quality
Specifies the desired reconstruction quality of the depth maps, providing that they are selected as a
source option. Higher quality settings can be used to obtain more detailed and accurate geometry, but
they require longer time for the processing.
Interpretation of the quality parameters here is similar to that of accuracy settings given in Photo
Alignment section. The only difference is that in this case Ultra High quality setting means processing
of original photos, while each following step implies preliminary image size downscaling by factor
of 4 (2 times by each side). For depth maps based mesh generation Mild filtering option is used by
default, unless Reuse depth maps option is enabled. Aggressive filtering can be used if the excessive
geometry (such as isolated mesh components around the reconstructed object) is observed, however,
some fine level thin elements may be lost due to this depth filtering mode selection.
Face count
Specifies the maximum number of polygons in the final mesh. Suggested values (High, Medium, Low)
present optimal number of polygons for a mesh of a corresponding level of detail. For the dense cloud
based reconstruction they are calculated based on the number of points in the source dense point cloud:
the ratio is 1/5, 1/15, and 1/45 respectively. It is still possible for a user to indicate the target number
of polygons in the final mesh through the Custom value of the Face count parameter. Please note that
while too small number of polygons is likely to result in too rough mesh, too huge custom number
(over 10 million polygons) is likely to cause model visualization problems in external software.
Interpolation
If interpolation mode is Disabled it leads to accurate reconstruction results since only areas
corresponding to dense point cloud points are reconstructed. Manual hole filling is usually required
at the post processing step.With Enabled (default) interpolation mode Metashape will interpolate
some surface areas within a circle of a certain radius around every dense cloud point. As a result some
holes can be automatically covered. Yet some holes can still be present on the model and are to be
filled at the post processing step. In Extrapolated mode the program generates holeless model with
extrapolated geometry. Large areas of extra geometry might be generated with this method, but they
could be easily removed later using selection and cropping tools.
23
General workflow
masking out background from single camera. For additional information on the usage of masks please
refer to the Using masks section. The option is only applicable for the Depth Maps source option
selected.
Note
• Metashape tends to produce 3D models with excessive geometry resolution, so it may be
reasonable to perform mesh decimation after geometry computation. More information on mesh
decimation and other 3D model geometry editing tools is given in the Editing model geometry
section.
To calibrate colors
1. Select Calibrate colors... command from the Tools menu.
2. Select the desired colors calibration parameters in the Calibrate colors dialog box. Click OK button
when done.
3. The progress dialog box will appear displaying the current processing status. To cancel processing
click Cancel button.
Source data
Defines what data should be taken as the basis for overlapping areas estimation.
Sparse cloud - the quickest yet the roughest estimation available.
24
General workflow
Model - gives more precise results, but only on condition that the surface is detailed enough. This
parameter value is the recommended one if the aim is to calibrate colors to improve the quality of
the model texture.
In Metashape it is possible to manually set color level adjustments for one or more images using the Adjust
color levels tool. The Adjust Color Levels tool is available from the selected images context menu in the
Photos pane. If the image is very dark or overexposed, you can manually set adjust the levels for each
channel of the image in the Color levels dialog of the tool or set the Average value for all channels.
Build Texture
The texture feature allows to build different types of textures for a model.
2. Select the desired texture generation parameters in the Build Texture dialog box. Click OK button
when done.
3. The progress dialog box will appear displaying the current processing status. To cancel processing
click Cancel button.
Texture type
Diffuse map - the basic texture that stores the colors of the model surface.
25
General workflow
Normal map - texture map that allows to calculate the illumination of the model parts from different
light sources in the post-processing workflow.
Occlusion map - a texture that contains pre-calculated shading information from background lighting.
Source data
Images - allow to build a color texture map (diffuse map) based on the aligned images of the model
or transferred from another model with already generated color texture map.
3D model - the normal and occlusion texture maps can be built based on another model and the relief
of the model specified in the Source data field will be transferred to the current model. Usually it
is recommended to use the more detailed model as a source data for the Normal and Occlusion map
generation for the 3D model with the lower polygon count. Normal map option will not be available
in the Source data field, if less than two 3D model instances are present in the active chunk.
Mapping mode
The texture mapping mode determines how the object texture will be packed in the texture atlas.
Proper texture mapping mode selection helps to obtain optimal texture packing and, consequently,
better visual quality of the final model.
The default mode is the Generic mapping mode; it allows to parametrize texture atlas for arbitrary
geometry. No assumptions regarding the type of the scene to be processed are made; program tries
to create as uniform texture as possible.
In the Orthophoto mapping mode the whole object surface is textured in the orthographic projection.
The Orthophoto mapping mode produces even more compact texture representation than the
Adaptive orthophoto mode at the expense of texture quality in vertical regions.
In the Adaptive orthophoto mapping mode the object surface is split into the flat part and vertical
regions. The flat part of the surface is textured using the orthographic projection, while vertical
regions are textured separately to maintain accurate texture representation in such regions. When in the
Adaptive orthophoto mapping mode, program tends to produce more compact texture representation
for nearly planar scenes, while maintaining good texture quality for vertical surfaces, such as walls
of the buildings.
Spherical mapping mode is appropriate only to a certain class of objects that have a ball-like form.
It allows for continuous texture atlas being exported for this type of objects, so that it is much easier
to edit it later. When generating texture in Spherical mapping mode it is crucial to set the Bounding
box properly. The whole model should be within the Bounding box. The red side of the Bounding
box should be under the model; it defines the axis of the spherical projection. The marks on the front
side determine the 0 meridian.
The Single camera mapping mode allows to generate texture from a single image. The image to be
used for texturing can be selected from 'Texture from' list.
The Keep uv mapping mode generates texture atlas using current texture parametrization. It can
be used to rebuild texture atlas using different resolution or to generate the atlas for the model
parametrized in the external software.
Mosaic - implies two-step approach: it does blending of low frequency component for overlapping
images to avoid seamline problem (weighted average, weight being dependent on a number of
parameters including proximity of the pixel in question to the center of the image), while high
frequency component, that is in charge of picture details, is taken from a single image - the one that
26
General workflow
presents good resolution for the area of interest while the camera view is almost along the normal to
the reconstructed surface in that point.
Average - uses the weighted average value of all pixels from individual photos, the weight being
dependent on the same parameters that are considered for high frequency component in mosaic mode.
Max Intensity - the image which has maximum intensity of the corresponding pixel is selected.
Min Intensity - the image which has minimum intensity of the corresponding pixel is selected.
Disabled - the image to take the color value for the pixel from is chosen like the one for the high
frequency component in mosaic mode.
Multi-page texture atlas generation is supported for Generic mapping mode only and Keep UV option,
if the imported model contains proper texture layout.
Note
• HDR texture generation requires HDR photos on input.
Metashape estimates image quality as a relative sharpness of the image with respect to other images in
the data set. The value of the parameter is calculated based on the sharpness level of the most focused
part of the picture.
27
General workflow
3. Right button click on the selected image(s) and choose Estimate Image Quality command from the
context menu.
4. Once the analysis procedure is over, a figure indicating estimated image quality value will be
displayed in the Quality column on the Photos pane.
• Photo alignment data such as information on camera positions, sparse point cloud model and set of
refined camera calibration parameters for each calibration group.
• Reconstructed 3D polygonal model with any changes made by user. This includes mesh and texture
if it was built.
• Structure of the project, i.e. number of chunks in the project and their content.
Note that since Metashape tends to generate extra dense point clouds and highly detailed polygonal models,
project saving procedure can take up quite a long time.
Project files use relative paths to reference original photos. Thus, when moving or copying the project file
to another location do not forget to move or copy photographs with all the folder structure involved as
well. Otherwise, Metashape will fail to run any operation requiring source images, although the project
file including the reconstructed model will be loaded up correctly. Alternatively, you can enable Store
absolute image paths option on the Advanced tab of the Preferences dialog available from Tools menu.
Exporting results
Metashape supports export of processing results in various representations: sparse and dense point clouds,
camera calibration and camera orientation data, mesh, etc.
Point cloud and camera calibration data can be exported right after photo alignment is completed. All other
export options are available after the corresponding processing step.
28
General workflow
If you are going to export the results (point cloud / mesh) for the chunk that is not referenced, please note
that the resulting file will be oriented according to a default coordinate system (see axes in the bottom
right corner of the Model view), i. e. the model can be shown differently from what you see in Metashape
window.
To align the model orientation with the default coordinate system use Rotate object button from the
Toolbar. Move object and Scale object instruments can be used to adjust the size and location of
the unreferenced model.
In some cases editing model geometry in the external software may be required. Metashape supports model
export for editing in external software and then allows to import it back as it is described in the Editing
model geometry section of the manual.
2. Browse the destination folder, choose the file type, and print in the file name. Click Save button.
3. In the Export Points dialog box select desired type of point cloud: Sparse point cloud or Dense point
cloud.
6. The progress dialog box will appear displaying the current processing status. To cancel processing
click Cancel button.
In some cases it may be reasonable to edit point cloud before exporting it. To read about point cloud editing
refer to the Editing point cloud section of the manual.
• LAZ (*.laz)
• U3D (*.u3d)
29
General workflow
• potree (*.zip)
Note
• Saving color information of the point cloud is not supported by the OBJ and DXF formats.
• Saving point normals information is not supported by the LAS, LAZ, PTS, CL3 and DXF
formats.
Metashape supports direct uploading of the point clouds to the following resources: 4DMapper, PointBox,
PointScene and Sketchfab. To publish your point cloud online use Upload Data... command from the File
menu.
• Alembic (*.abc)
30
General workflow
Note
• Camera data export in Bundler and Boujou file formats will save sparse point cloud data in the
same file.
• Camera data export in Bundler file format would not save distortion coefficients k3, k4.
To export tie points data one should choose one of the following export formats in the Export Cameras
dialog: BINGO, ORIMA, PATB, Summit Evolution or Blocks exchange. Tie points can be exported only
along with interior and exterion orientation parameters of the cameras.
example.pro
Project file PATB. It contains information on the units of measurement (microns/millimeters). This
eliminates the ambiguity in interpreting the data on import.
example.im
File with coordinates of tie points projections on the images.
example.ori
File with the calculated parameters of the exterior orientation for the cameras.
example.at
File with the calculated coordinates of tie points (adjusted terrain coordinates).
Note
• *.im and *.con formats are input files for PATB and *.ori. and *.at are output formats.
As a result of export in BINGO format, the following files will be written to the destination folder: itera.dat
(ITERA file); image.dat (IMAGE COORDINATE file); geoin.dat (GEO INPUT file); gps-imu.dat (GPS/
IMU data).
As a result of export in Summit Evolution format two files will be created (*.cam, *.smtxml). File in
*.cam format includes information about camera calibration parameters. File in *.smtxml format is Summit
Evolution project file.
To export / import camera calibration data only select Camera Calibration... command from the Tools
menu. Using / buttons it is possible to load / save camera calibration data in the following formats:
31
General workflow
Panorama export
Metashape is capable of panorama stitching for images taken from the same camera position - camera
station. To indicate for the software that loaded images have been taken from one camera station, one
should move those photos to a camera group and assign Camera Station type to it. For information on
camera groups refer to Loading images section.
To export panorama
1. Select Export Panorama... command from the File menu.
3. Choose panorama orientation in the file with the help of navigation buttons to the right of the preview
window in the Export Panorama dialog.
4. Set exporting parameters: select camera groups which panorama should be exported for and indicate
export file name mask.
5. Click OK button
Additionally, you can set boundaries for the region of panorama to be exported using Setup boundaries
section of the Export Panorama dialog. Text boxes in the first line (X) allow to indicate the angle in the
horizontal plane and the second line (Y) serves for angle in the vertical plane limits. Image size (pix) option
enables to control the dimensions of the exported image file.
• PNG (*.png)
• BMP (*.bmp)
• OpenEXR (*.exr)
• TARGA (*.tga)
3D model export
To export 3D model
1. Select Export Model... command from the File menu.
32
General workflow
2. Browse the destination folder, choose the file type, and print in the file name. Click Save button.
3. In the Export Model dialog indicate export parameters applicable to the selected file type.
5. The progress dialog box will appear displaying the current processing status. To cancel processing
click Cancel button.
• COLLADA (*.dae)
• Alembic (*.abc)
Some file formats (OBJ, 3DS, VRML, COLLADA, PLY, FBX) save texture image in a separate file. The
texture file should be kept in the same directory as the main file describing the geometry. If the texture
atlas was not built only the model geometry is exported.
Metashape supports direct uploading of the models to Sketchfab resource. To publish your model online
use Upload Data... command from the File menu.
- Convert Images... command is available from Export submenu of the File menu and allows to export free
of lens distortions and geometrically corrected images. It also allows to convert large images to a pyramid-
optimized TIFF format for prompt navigation in the Photo view mode.
In the Parameters section of the Convert Images dialog window the user can check the following options:
Correct distortions, Center principal point, Square pixels, Apply color correction. Filename template for
converted images should also be specified in the dialog window.
33
General workflow
In the Compression section of the Convert Images dialog window the user can set one of the suggested
TIFF compression parameters: LSW, JPEG, Packbits, Deflate; set JPEG quality and specify if Tiled TIFF,
BigTIFF file or TIFF overviews should also be saved. The corrections can be applied to all cameras, entire
workspace, selected cameras or even to a current photo.
- Depth map for any image (Export Depth... command available from photo context menu). You can export
diffuse map, depth map and normal map.
- High resolution image of the model as it is shown in the Model view mode. Capture View command
available from the context menu shown on right button click in the Model view.
Note
• You need to have mesh model generated in the chunk in order to export diffuse map, depth map
and normal map.
2. Browse the destination folder, choose the file type, and print in the file name. Click Save button.
3. The progress dialog box will appear displaying the current processing status. To cancel processing
click Cancel button.
• Survey data including coverage area, flying altitude, GSD, general camera(s) info, as well as overlap
statistics.
• Camera calibration results: figures and an illustration for every sensor involved in the project.
Note
• Processing report can be exported after alignment step.
Survey Data
Fig. Camera locations and image overlap - shows the model and position the cameras is shown and the
number of overlaps is displayed in color.
34
General workflow
Tie points - total number of valid tie points (equals to the number of points in the sparse cloud).
Ground resolution - effective ground resolution averaged over all aligned images.
Reprojection error - root mean square reprojection error averaged over all tie points on all images.
Table. Cameras - table with the parameters: camera model, resolution, focal length, pixel size,
precalibrated.
Reprojection error is the distance between the point on the image where a reconstructed 3D point can be
projected and the original projection of that 3D point detected on the photo and used as a basis for the
3D point reconstruction procedure.
Camera Calibration
For precalibrated cameras internal parameters input by the user are shown on the report page. If a camera
was not precalibrated, internal camera parameters estimated by Metashape are presented.
Fig. Image residuals for camera - displays the reprojection errors for the tie points detected on the source
images, averaged across all the images of the calibration group and inside the certain "cells" on the images.
Camera name (focal length) - camera model name and number of images.
Table. Calibration coefficients and correlation matrix - table with the calibration coefficients and
parameters of the covariance matrix (F, Cx, Cy, B1, B2, K1, K2, K3, K4, P1, P2).
Resolution - effective resolution of the exported DEM. The value depends on the Quality parameter value
used at Build point cloud step, providing that DEM has been generated from dense point cloud.
Point Density - average number of dense cloud points per square meter.
Processing Parameters
Processing report contains processing parameters information, which is also available form Chunk context
menu. Along with the values of the parameters used at various processing stages, this page of the report
presents information on processing time.
Metashape matches images on different scales to improve robustness with blurred or difficult to match
images. The accuracy of tie point projections depends on the scale at which they were located. Metashape
35
General workflow
uses information about scale to weight tie point reprojection errors. Key point size is the Sigma of the
Gaussian blur at the pyramid level of scales at which the key point was found.
On the processing parameters page of the report (as well as in chunk information dialog) two reprojection
errors are provided: the reprojection error in the units of key point scale (this is the quantity that is
minimized during bundle adjustment), and the reprojection error in pixels (for convenience). The Mean
key point size value is averaged Key point size value over all key points. Average tie point multiplicity
value is the ratio of the total number of projections to the number of tie points.
To manually create a track and add a viewpoint, click Append button on the Animation pane. Append
option will add the current viewpoint to the active camera track.
3. Choose Horizontal preset parameter in Create Track dialog window and set the Keyframe count.
4. Click OK button.
5. Adjust the position of the viewpoint in the camera track in Model view by dragging the left mouse
button to display it visually.
3. Choose Vertical preset parameter in Create Track dialog window and set the Keyframe count.
4. Click OK button.
5. Adjust the position of the viewpoint in the camera track in Model view by dragging the left mouse
button to display it visually.
Note
• To display the camera track path, select Show animation command from the Show/Hide items
submenu on the Model menu.
Camera track parameters can be adjusted in the Animation Settings dialog window. The following
parameters can be changed: Camera Track Label, Duration (in seconds), Field of view (deg), Smooth
camera track option, Loop camera track option, Rotation angles convention. Click Settings button on the
Animation pane to open Animation Settings dialog.
36
General workflow
2. Select the desired export parameters in the Capture Video dialog, such as export video resolution,
compression type and frame rate.
3. To save the track path to external file click Save button on the Animation pane.
• Camera Path
• KML
Note
• That KML format is only supported for projects georeferenced in geographic/projected systems
convertible to WGS84 system.
37
Chapter 4. Improving camera alignment
results
Camera calibration
Calibration groups
While carrying out photo alignment Metashape estimates both internal and external camera orientation
parameters, including nonlinear radial distortions. For the estimation to be successful it is crucial to apply
the estimation procedure separately to photos taken with different cameras. Once photos have been loaded
in the program, Metashape automatically divides them into calibration groups according to the image
resolution and/or EXIF meta data like camera type and focal length. All the actions described below could
and should be applied (or not applied) to each calibration group individually.
2. In the Camera Calibration dialog box, select photos to be arranged in a new group.
4. A new group will be created and depicted on the left-hand part of the Camera Calibration dialog box.
2. In the Camera Calibration dialog box choose the source group on the left-hand part of the dialog.
3. Select photos to be moved and drag them to the target group on the left-hand part of the Camera
Calibration dialog box.
To place each photo into a separate group you can use Split Groups command available at the right button
click on a calibration group name in the left-hand part of the Camera Calibration dialog
Camera types
Metashape supports two major types of camera: frame camera and fisheye camera. Camera type can be
set in Camera Calibration dialog box available from Tools menu.
Frame camera. If the source data within a calibration group was shot with a frame camera, for successful
estimation of camera orientation parameters the information on approximate focal length (pix) is required.
Obviously, to calculate focal length value in pixel it is enough to know focal length in mm along with the
sensor pixel size in mm. Normally this data is extracted automatically from the EXIF meta data.
Frame camera with Fisheye lens. If extra wide lenses were used to get the source data, standard Metashape
camera model will not allow to estimate camera parameters successfully. Fisheye camera type setting will
initialize implementation of a different camera model to fit ultra-wide lens distortions.
Optionally rolling shutter compensation can be enabled by checking the Enable rolling shutter
compensation box in the Camera Calibration dialog available from Tools menu.
38
Improving camera alignment results
In case source images lack EXIF data or the EXIF data is insufficient to calculate focal length in pixels,
Metashape will assume that focal length equals to 50 mm (35 mm film equivalent). However, if the initial
guess values differ significantly from the actual focal length, it is likely to lead to failure of the alignment
process. So, if photos do not contain EXIF meta data, it is preferable to specify focal length (mm) and
sensor pixel size (mm) manually. It can be done in Camera Calibration dialog box available from Tools
menu. Generally, this data is indicated in camera specification or can be received from some online source.
To indicate to the program that camera orientation parameters should be estimated based on the focal
length and pixel size information, it is necessary to set the Type parameter on the Initial tab to Auto value.
2. Select calibration group, which requires re-estimation of camera orientation parameters on the left
side of the Camera Calibration dialog box.
f
Focal length measured in pixels.
cx, cy
Principal point coordinates, i.e. coordinates of lens optical axis interception with sensor plane in pixels.
b1, b2
Affinity and Skew (non-orthogonality) transformation coefficients.
p1, p2
Tangential distortion coefficients.
Note
• Alternatively, initial calibration data can be imported from file using Load button on the Initial
tab of the Camera Calibration dialog box. In addition to Agisoft calibration file format it is
possible to import data from Australis, PhotoModeler, 3DM CalibCam, CalCam, Inpho camera
calibration, USGS camera calibration, OpenCV and Z/I Distortion Grid formats.
Initial calibration data will be adjusted during the Align Photos processing step. Once Align Photos
processing step is finished adjusted calibration data will be displayed on the Adjusted tab of the Camera
39
Improving camera alignment results
Calibration dialog box. The details about distortion models used in Metashape are given in the Appendix C,
Camera models section.
If very precise calibration data is available, click Select... button next to Fixed parameters: item. In Fixed
parameters dialog choose calibration parameters that are to be fixed and click OK button. The initial values
for the chosen parameters will not be changed during Align Photos or Optimize Cameras operations.
In some cases it may be required to estimate some sub-set of parameters individually for each camera in the
calibration group. In this case such parameters should be selected in the section Image-variant parameters.
Click Select button and choose these parameters in Image-variant parameters dialog box.
Adjusted camera calibration data can be saved to file using Save button on the Adjusted tab of the Camera
Calibration dialog box.
Distortion
Distortion tab presents estimated camera distortion plot. Total, Radial, Decentering, Corrections and
Residual options are available in the tab. Distortion graph represent the distortion values and direction
according to the adjusted calibration coefficient values. It shows the plot in the discrete vectors mode
for a central point in the corresponding image cell. Residuals presents residuals graph which allows
to evaluate how adequately the camera is described with the applied mathematical model. Note that
residuals are averaged per cell of an image and then across all the images in a camera group. Scale
reference under the plot indicates the scale of the distortions/residuals.
Profile
Profile tab presents increase of the corresponding radial and decentering distortions with distance
from the center of the photo. Profiles can be saved as image.
Correlation
Correlation tab presents:
• Correlation values for internal camera orientation parameters - reflect the degree of correlation
between the corresponding parameters.
Vignetting
Vignetting tab presents radiometric distortion of the lens. Metashape uses radiometric distortion
parameters of the lens from photos metadata, in case there is no information about radiometric
distortion of the lens in photo metadata Vignetting tab is inactive. You can calculate radiometric
distortion manually using Calibrate Colors command available from the Tools menu, read more in
Building model texture section.
Optimization
Optimization of camera alignment
During photo alignment step Metashape automatically finds tie points and estimates intrinsic and extrinsic
camera parameters. However, the accuracy of the estimates depends on many factors, like overlap between
40
Improving camera alignment results
the neighboring photos, as well as on the shape of the object surface. Thus, it is recommended to inspect
alignment results in order to delete tie points with too large reprojection error if any. Please refer to Editing
point cloud section for information on point cloud editing. Once the set of tie points has been edited, it is
necessary to run optimization procedure to reestimate intrinsic and extrinsic camera parameters.
Optimization procedure calculates intrinsic and extrinsic camera parameters based on the tie points left
after editing procedure. Providing that outliers have been removed, the estimates will be more accurate.
In addition, this step involves estimation of a number of intrinsic camera parameters which are fixed at
the alignment step: b1 (aspect), b2 (skew) and higher order radial distortion coefficient k4. Optionally you
may turn on Adaptive camera model fitting read more in Aligning photos section.
2. In Optimize Camera Alignment dialog box check camera parameters to be optimized. Click OK button
to start optimization.
3. After optimization is complete, estimated intrinsic camera parameters can be inspected on the
Adjusted tab of the Camera Calibration dialog available from the Tools menu.
Note
• The model data (if any) is cleared by the optimization procedure. You will have to rebuild the
model geometry after optimization.
41
Chapter 5. Editing
Using masks
Overview
Masks are used in Metashape to specify the areas on the photos which can otherwise be confusing to
the program or lead to incorrect reconstruction results. Masks can be applied at the following stages of
processing:
• Building Mesh from the depth maps source with the strict volumetric masks option applied
• Building Texture
Masking may be also useful when the object of interest occupies only a small part of the photo. In
this case a small number of useful matches can be filtered out mistakenly as a noise among a much
greater number of matches between background objects.
42
Editing
Masked areas are always excluded from processing during dense point cloud and texture generation
stages, including Tiled Model generation process.
Let's take for instance a set of photos of some object. Along with an object itself on each photo some
background areas are present. These areas may be useful for more precise camera positioning, so it
is better to use them while aligning the photos. However, impact of these areas at the building dense
point cloud is exactly opposite: the resulting model will contain object of interest and its background.
Background geometry will "consume" some part of mesh polygons that could be otherwise used for
modeling the main object.
Setting the masks for such background areas allows to avoid this problem and increases the precision
and quality of geometry reconstruction.
Loading masks
Masks can be loaded from external sources, as well as generated automatically from background images
if such data is available. Metashape supports loading masks from the following sources:
To import masks
2. In the Import Mask dialog select suitable parameters. Click OK button when done.
3. When generating masks from separate or background images, the folder selection dialog will appear.
Browse to the folder containing corresponding images and select it.
4. The progress dialog box will appear displaying the current processing status. To cancel processing
click Cancel button.
43
Editing
Method
Specifies the source of the mask data.
From Alpha - load masks from alpha channel of the source photos.
Operation
Specifies the action to be done in case a second mask is imported for the photo.
Replacement - new mask will be loaded and stored instead of the original one.
Intersection - the intersection of the two masks will be stored as a new mask for the photo.
Difference - only the difference between two masks will be stored as a new mask for the photo.
44
Editing
For example, {filename}_mask.png template can be used if masks are available in PNG format and
have a _mask suffix.
Apply to
Specifies whether masks should be imported for the currently opened photo, active chunk or entire
Workspace.
Selected cameras - load mask for the currently checked cameras (if any).
Current photo - load mask for the currently opened photo (if any).
Editing masks
Modification of the current mask is performed by adding or subtracting selections. A selection is created
with one of the supported selection tools and is not incorporated in the current mask until it is merged with
a mask using Add Selection or Subtract Selection operations.
3.
Click on Add Selection toolbar button to add current selection to the mask, or Subtract
Selection to subtract the selection from the mask. Invert Selection button allows to invert current
selection prior to adding or subtracting it from the mask.
45
Editing
The range of pixel colors selected by Magic Wand is controlled by the tolerance value. At lower
tolerance values the tool selects fewer colors similar to the pixel you click with the Magic Wand tool.
Higher value broadens the range of colors selected.
Note
• To add new area to the current selection hold the Ctrl key during selection of additional area.
• To subtract a part from the current selection hold the Shift key during selection of the area to
be subtracted.
A mask can be inverted using Invert Mask command from the Edit menu. The command is active in Photo
view only. Alternatively, you can invert masks either for selected cameras or for all cameras in a chunk
using Invert Masks... command from a photo context menu on the Photos pane.
The masks are generated individually for each image. If some object should be masked out, it should be
masked out on all photos, where that object appears.
Saving masks
Created masks can be also saved for external editing or storage.
To export masks
1. Select Export Masks... command from the File menu.
2. In the Export Mask dialog select suitable parameters. Click OK button when done.
3. Browse to the folder where the masks should be saved and select it.
4. The progress dialog box will appear displaying the current processing status. To cancel processing
click Cancel button.
Current photo - save mask for the currently opened photo (if any).
46
Editing
File type
Specifies the type of generated files.
Single channel mask image - generates single channel black and white mask images.
Image with alpha channel - generates color images from source photos combined with mask data
in alpha channel.
For example, {filename}_mask.png template can be used to export masks in PNG format with _mask
suffix.
Note
• When importing/exporting mask for the current photo only, Metashape will prompt for the actual
image instead of image folder. Mask file names parameter will not be used in this case.
• Reducing number of points in point cloud by setting tie point per photo limit (sparse cloud only)
• Sample points operation to create dense point cloud from the mesh or tiled model
Note
• Point cloud editing operation can be undone/redone using Undo/Redo command from the Edit
menu.
47
Editing
Reprojection error
Maximum reprojection error in normalized units across all images where tie-point was measured.
where
x'i - point projection according to adjusted orientation parameters on the i-th image in pixels,
si - image scale at which corresponding projection was measured on the i-th image
High reprojection error usually indicates poor localization accuracy of the corresponding point
projections at the point matching step. It is also typical for false matches. Removing such points can
improve accuracy of the subsequent optimization step.
Reconstruction uncertainty
Ratio of the largest semi-axis to the smallest semi-axis of the error ellipse of the triangulated 3D point
coordinates. The error ellipse corresponds to the uncertainty of the point triangulation alone without
taking into account propagation of uncertainties from interior and exterior orientation parameters.
sqrt(k1 / k3)
where
High reconstruction uncertainty is typical for points, reconstructed from nearby photos with small
baseline. Such points can noticeably deviate from the object surface, introducing noise in the point
cloud. While removal of such points should not affect the accuracy of optimization, it may be useful
to remove them before building geometry in Point Cloud mode or for better visual appearance of the
point cloud.
Image count
Number of images where the tie point was measured.
Metashape reconstruct all the points that are visible at least on two photos. However, points that are
visible only on two photos are likely to be located with poor accuracy. Image count filtering enables
to remove such unreliable points from the cloud.
Projection Accuracy
Average image scale at which image coordinates of the tie-point were measured.
sumi si / n
where
si - image scale at which corresponding projection was measured on the i-th image,
48
Editing
This criterion allows to filter out points which projections were relatively poorer localized due to their
bigger size.
3. In the Gradual Selection dialog box specify the criterion to be used for filtering. Adjust the threshold
level using the slider. You can observe how the selection changes while dragging the slider. Click
OK button to finalize the selection.
4.
To remove selected points use Delete Selection command from the Edit menu or click Delete
Selection toolbar button (or simply press Del button on the keyboard).
2. Choose Select Masked Points... command from the Dense Cloud submenu of the Tools menu.
3. In the Select Masked Points dialog box indicate the photos whose masks to be taken into account.
Adjust the edge softness level using the slider. Click OK button to run the selection procedure.
4.
To remove selected points use Delete Selection command from the Edit menu or click Delete
Selection toolbar button (or simply press Del button on the keyboard).
2. Choose Select Points by Color... command from the Dense Cloud submenu of the Tools menu.
3. In the Select Points by Color dialog box the color to be used as the criterion. Adjust the tolerance
level using the slider. Click OK button to run the selection procedure.
4.
To remove selected points use Delete Selection command from the Edit menu or click Delete
Selection toolbar button (or simply press Del button on the keyboard).
49
Editing
2.
Select points to be filter Rectangle Selection, Circle Selection or Free-Form Selection tools.
3. Choose Filter By Selection... command from the Dense Cloud submenu of the Tools menu.
5. Select Reset filter command from the Dense Cloud submenu of the Tools menu to cancel filtering
by classes
Filter by confidence
1.
Switch to Dense Cloud view mode using Dense Cloud Confidence toolbar button.
2. Choose Filter by Confidence... command from the Dense Cloud submenu of the Tools menu.
3. In the Select Confidence Range dialog box choose min and max value to be used as the filter. The
values define the number of depth maps involved in the point generation.
6. Select Reset filter command from the Dense Cloud submenu of the Tools menu to cancel filtering
by confidence.
The number of tie points can also be reduced after the alignment process with Tie Points - Thin Point Cloud
command available from Tools menu. As a results sparse point cloud will be thinned, yet the alignment
will be kept unchanged.
2. In the Filter Dense Cloud dialog box specify Point spacing in meters to be used.
4. The progress dialog box will appear displaying the current processing status. To cancel processing
click Cancel button.
50
Editing
Note
• In the confirmation dialog box click Yes button if a new dense cloud instance with result
calculation should be created in the active chunk. If No button is pressed the dense cloud filtering
result will overwrite the active dense point cloud instance.
2. Choose Colorize Dense Cloud... command from the Dense Cloud submenu of the Tools menu.
3. In the Colorize Dense Cloud dialog box select Source data from Images option.
4. Click on the OK button to start the procedure. The progress dialog box will appear displaying the
current processing status.
Sample points
Sample points tool creates a dense point cloud based on a Mesh or Tiled model.
2. In the Sample Points dialog select the source polygonal surface type (mesh or tiled model) and desired
point spacing value. Target point count based on the input parameters will be displayed in Estimated
points field.
3. Click on the OK button to start the procedure. The progress dialog box will appear displaying the
current processing status. To cancel processing click Cancel button.
Source data
Specifies the source polygonal surface for the dense point cloud generation (Mesh or Tiled model).
Points spacing
Specifies the desired distance between points of the dense cloud to be generated.
Estimated points
Defines the expected number of points to be generated.
51
Editing
2.
Select points to be removed using Rectangle Selection, Circle Selection or Free-Form
Selection tools. To add new points to the current selection hold the Ctrl key during selection of
additional points. To remove some points from the current selection hold the Shift key during
selection of points to be removed.
3.
To delete selected points click the Delete Selection toolbar button or select Delete Selection
command from the Edit menu. To crop selection to the selected points click the Crop Selection
toolbar button or select Crop Selection command from the Edit menu.
• Decimation tool
• Smooth tool
• Colorize vertices
More complex editing can be done in the external 3D editing tools. Metashape allows to export mesh and
then import it back for this purpose.
Note
• For polygon removal operations such as manual removal and connected component filtering it
is possible to undo the last mesh editing operation. There are Undo/Redo commands in the Edit
menu.
• Please note that Undo/Redo commands are not supported for mesh decimation and this operation
cannot be undone.
Decimation tool
Decimation is a tool used to decrease the geometric resolution of the model by replacing high resolution
mesh with a lower resolution one, which is still capable of representing the object geometry with high
accuracy. Metashape tends to produce 3D models with excessive geometry resolution, so mesh decimation
is usually a desirable step after geometry computation.
Highly detailed models may contain hundreds of thousands of polygons. While it is acceptable to work
with such a complex models in 3D editor tools, in most conventional tools like Adobe Reader or Google
Earth high complexity of 3D models may noticeably decrease application performance. High complexity
also results in longer time required to build texture and to export model in PDF file format.
In some cases it is desirable to keep as much geometry details as possible like it is needed for scientific and
archive purposes. However, if there are no special requirements it is recommended to decimate the model
52
Editing
down to 100 000 - 200 000 polygons for exporting in PDF, and to 100 000 or even less for displaying in
Google Earth and alike tools.
To decimate 3D model
1. Select Decimate Mesh... command from the Tools menu.
2. In the Decimate Mesh dialog box specify the target number of polygons, which should remain in the
final model. Click on the OK button to start decimation.
3. The progress dialog box will appear displaying the current processing status. To cancel processing
click on the Cancel button.
Note
• Texture atlas is discarded during decimation process. You will have to rebuild texture atlas after
decimation is complete.
To refine mesh
1. Check the model - it will be refined with respect to camera photos. You can duplicate it with right
click on the model and selecting Duplicate... to refine the copy of model.
3. In the Refine Mesh... dialog box specify the target quality of refinement, number of iterations and
smoothness. Click on the OK button to start refinement.
4. The progress dialog box will appear displaying the current processing status. To cancel processing
click on the Cancel button.
Refinement parameters
Quality
Specifies the desired refinement quality. Higher quality settings can be used to obtain more detailed
and accurate geometry, but they require longer time for processing. Interpretation of the quality
parameters here is similar to that of accuracy settings given in Building dense point cloud section.
Iterations
Number of refinement iterations. In some cases additional iterations are able to recover more details,
but it will lead to proportional slow down.
Smoothness
Smaller smoothness parameter leads to better features recovering but also it can increase noise. Bigger
smoothness parameter leads to better noise suppression, but can smooth out features too. Changing
smoothness value can help to balance between noise suppression and features recovering.
53
Editing
2. In the Colorize Model dialog box select source data for the operation.
4. The progress dialog box will appear displaying the current processing status. To cancel processing
click on the Cancel button.
Close holes tool enables to close void areas on the model substituting photogrammetric reconstruction
with extrapolation data. It is possible to control an acceptable level of accuracy indicating the maximum
size of a hole to be covered with extrapolated data.
2. In the Close Holes dialog box indicate the maximum size of a hole to be covered with the slider.
4. The progress dialog box will appear displaying the current processing status. To cancel processing
click on the Cancel button.
Note
• The slider allows to set the size of a hole in relation to the size of the whole model surface.
Smooth tool
The Smooth tool allows you to make the model smoother and remove irregularities on the surface. Some
tasks require a sleek surface disregarding of details or real object more smooth than mesh in Metashape.
You can apply the tool to the entire mesh or only to the selected area.
Note
• To apply smoothing to a specific area, you must first select it and then apply the tool.
To smooth mesh
1. Select Smooth mesh... command in Mesh submenu of Tools menu.
2. In the Smooth mesh dialog box set the Strength parameter using slider. You can also enable check
box Apply to selected faces.
54
Editing
Note
• Fix borders option can be applied to the models with the open edges, it allows to preserve the
position of the mesh vertices along the open edges when using smoothing.
Polygon size
This filtering criteria allows to select polygons up to a certain size. The size of the polygons to be
selected is set with a slider and is indicated in relation to the size of the whole model. This function
can be useful, for example, in case the geometry was reconstructed with the extrapolation and there
is a need to remove extra polygons automatically added by Metashape to fill the gaps; these polygons
are often of a larger size than the rest.
2. In the Gradual Selection dialog box select Connected component size criterion.
3. Select the size of isolated components to be removed using the slider. Size of the largest component
is taken for 100%. You can observe how the selection changes while dragging the slider. Click OK
button to finalize the selection.
4. To remove the selected components use Delete Selection command from the Edit menu or click
Delete Selection toolbar button (or simply press Del button on the keyboard).
3. Select the size of polygons to be removed using the slider. Size of the largest polygon is taken for
100%. You can observe how the selection changes while dragging the slider. Click OK button to
finalize the selection.
4. To remove the selected components use Delete Selection command from the Edit menu or click
Delete Selection toolbar button (or simply press Del button on the keyboard).
Note that Metashape always selects the fragments starting from the smallest ones. If the model contains
only one component the selection will be empty.
55
Editing
2. Make the selection using the mouse. To add new polygons to the current selection hold the Ctrl key
during selection of additional polygons. To remove some polygons from the current selection hold
the Shift key during selection of polygons to be excluded.
3.
To delete selected polygons click the Delete Selection toolbar button or use Delete Selection
command from the Edit menu. To crop selection to the selected polygons click the Crop Selection
toolbar button or use Crop Selection command from the Edit menu.
2. To shrink current selection press PageDown key in the selection mode. To shrink selection by even
a larger amount, press PageDown while holding Shift key pressed.
2. In the Mesh Statistics dialog box you can inspect mesh parameters. If there are any topological
problems, Fix Topology button will be active and can be clicked to solve the problems.
3. The progress dialog box will appear displaying the current processing status. To cancel processing
click on the Cancel button.
2. In the Save As dialog box, specify the desired mesh format in the Save as type combo box. Select the
file name to be used for the model and click Save button.
3. In the opened dialog box specify additional parameters specific to the selected file format. Click OK
button when done.
56
Editing
2. In the Open dialog box, browse to the file with the edited model and click Open.
Note
• Metashape supports loading models in Wavefront OBJ, 3DS, STL, COLLADA, Stanford PLY,
Autodesk FBX, Autodesk DXF, OpenCTM and U3D file formats only. Please make sure to select
one of these file formats when exporting model from the external 3D editor.
57
Chapter 6. Automation
Using chunks
When working with typical data sets, automation of general processing workflow allows to perform routine
operations efficiently. Metashape allows to assign several processing steps to be run one by one without
user intervention thanks to Batch Processing feature. Manual user intervention can be minimized even
further due to 'multiple chunk project' concept, each chunk to include one typical data set. For a project with
several chunks of the same nature, common operations available in Batch Processing dialog are applied to
each selected chunk individually, thus allowing to set several data sets for automatic processing following
predefined workflow pattern.
In addition, multiple chunk project could be useful when it turns out to be hard or even impossible to
generate a 3D model of the whole scene in one go. This could happen, for instance, if the total amount
of photographs is too large to be processed at a time. To overcome this difficulty Metashape offers a
possibility to split the set of photos into several separate chunks within the same project. Alignment of
photos, building dense point cloud, building mesh, and forming texture atlas operations can be performed
for each chunk separately and then resulting 3D models can be combined together.
Working with chunks is not more difficult than using Metashape following the general workflow. In fact,
in Metashape always exists at least one active chunk and all the 3D model processing workflow operations
are applied to this chunk.
To work with several chunks you need to know how to create chunks and how to combine resulting 3D
models from separate chunks into one model.
Creating a chunk
To create new chunk click on the Add Chunk toolbar button on the Workspace pane or select Add
Chunk command from the Workspace context menu (available by right-clicking on the root element on
the Workspace pane).
After the chunk is created you may load photos in it, align them, generate dense point cloud, generate mesh
surface model, build texture atlas, export the models at any stage and so on. The models in the chunks
are not linked with each other.
The list of all the chunks created in the current project is displayed in the Workspace pane along with flags
reflecting their status.
R (Referenced)
Will appear when two or more chunks are aligned with each other.
To move photos from one chunk to another simply select them in the list of photos on the Workspace pane,
and then drag and drop to the target chunk.
Note
• Chunk can contain several instances of the same object (point clouds, 3D model, etc).
58
Automation
Note that all these operations are applied to the active chunk. When a new chunk is created it is activated
automatically. Save project operation saves the content of all chunks. To save selected chunks as a separate
project use Save Chunks command from the chunk context menu.
To remove chunk
To rearrange the order of chunks in the Workspace pane simply drag and drop the chunks in the pane.
Aligning chunks
After the "partial" 3D models are built in several chunks they can be merged together. Before merging
chunks they need to be aligned.
2. In the Align Chunks dialog box select chunks to be aligned, indicate reference chunk with a double-
click. Set desired alignment options. Click OK button when done.
3. The progress dialog box will appear displaying the current processing status. To cancel processing
click the Cancel button.
59
Automation
The following parameters control the chunks alignment procedure and can be modified in the Align Chunks
dialog box:
Method
Defines the chunks alignment method. Point based method aligns chunks by matching photos across
all the chunks. Camera based method is used to align chunks based on estimated camera locations.
Corresponding cameras should have the same label.
Fix scale
Option is to be enabled in case the scales of the models in different chunks were set precisely and
should be left unchanged during chunks alignment process.
60
Automation
single photo, relevant key points on the rest of the photos picturing the same area will be also ignored
during alignment procedure (a tie point is a set of key points which have been matched as projections
of the same 3D point on different images). This can be useful to be able to suppress background in
turntable shooting scenario with only one mask. For additional information on the usage of masks
please refer to the Using masks section.
Note
• Chunk alignment can be performed only for chunks containing aligned photos.
Merging chunks
After alignment is complete the separate chunks can be merged into a single chunk.
Note
• We recommend processing all the images in a single chunk to provide better accuracy and quality
of the final results than the chunking method (separate processing of the sub-sets of the complete
photos set and merging them).
To merge chunks
1. Select Merge Chunks command from the Workflow menu.
2. In the Merge Chunks dialog box select chunks to be merged and the desired merging options. Click
OK button when done.
3. Metashape will merge the separate chunks into one. The merged chunk will be displayed in the project
content list on Workspace pane.
The following parameters control the chunks merging procedure and can be modified in the Merge Chunks
dialog box:
61
Automation
Merge models
Defines if models from the selected chunks are combined.
Chunks merging result (i.e. photos, point clouds and geometry) will be stored in the new chunk and it may
be treated as an ordinary chunk (e.g. the model can be textured and/or exported).
Note
• Dense cloud and model merging operations will be applied only to the active items. Disabled
(inactive) items will not be transferred to the merged chunk.
Batch processing
Metashape allows to perform general workflow operations with multiple chunks automatically. It is useful
when dealing with a large number of chunks to be processed.
Batch processing can be applied to all chunks in the Workspace, to unprocessed chunks only, or to the
chunks selected by the user. Each operation chosen in the Batch processing dialog will be applied to every
selected chunk before processing will move on to the next step.
3. In the Add Job dialog select the kind of operation to be performed, the list of chunks it should be
applied to, and desired processing parameters. Click OK button when done.
5. Arrange jobs by clicking Up and Down arrows at the right of the Batch Process... dialog box.
7. The progress dialog box will appear displaying the list and status of batch jobs and current operation
progress. To cancel processing click the Cancel button.
62
Automation
When the batch process includes import/export features that are applied to multiple chunks it is reasonable
to use the following templates in the Path field of the import/export jobs:
The list of tasks for batch processing can be exported to XML structured file using Save button in the
Batch processing dialog and imported in a different project using Open button.
63
Appendix A. Graphical user interface
Application window
General view
Model view
Model view tab is used for displaying 3D data as well as for mesh and point cloud editing. The view of
the model depends on the current processing stage and is also controlled by mode selection buttons on
the Metashape toolbar.
64
Graphical user interface
Model can be shown as a dense cloud or as a mesh in shaded, solid, wireframe or textured mode. Along
with the model the results of photo alignment can be displayed. These include sparse point cloud and
camera positions visualized data.
All navigation tools are accessible in the navigation mode only. To enter the navigation mode click the
Note
• Zooming into the model can be also controlled by the mouse wheel.
Photo view
Photo view tab is used for displaying individual images as well as corresponding depth maps in semi-
transparent representation and masks on them.
In the Photo view tab it is allowed to draw masks on the images. Additionally in Photo view tab it is
possible to display the residuals for the tie points providing that the corresponding camera is aligned.
Photo view is visible only if any image is opened. To open an image in Photo view mode double-click on
the label of the corresponding camera on the Workspace or Photos pane.
Switching to Photo view mode changes the contents of the Toolbar, presenting related instruments and
hiding irrelevant buttons.
Workspace pane
On the Workspace pane all elements comprising the current project are displayed. These elements can
include:
• Add chunk
65
Graphical user interface
• Add photos
• Remove items
Each element in the list is linked with the context menu providing quick access to some common
operations.
Photos pane
Photos pane displays the list of photos / masks / depth maps in the active chunk in the form of thumbnails.
• Remove cameras
• Increase / decrease icons' size or display detailed information on photos including EXIF data
Console pane
Console pane is used for:
• Clear log
Jobs pane
Jobs pane is designed for:
66
Graphical user interface
Note
• You can switch between projects by right-clicking on the project name.
Animation pane
Animation pane is designed for:
• Creating a new camera track from pre-sets (Horizontal, Vertical) with the given number of the keyframes
• Loading camera track from external file in supported formats (Autodesk FBX, Camera Path)
• Rendering the frame sequence to the form of the separate images for the keyframes
• Settinging the camera track parameters (label, duration, field of view, loop camera track)
Note
• To open any pane select a corresponding command from the View menu.
• To display the track path, select Show animation in Show/Hide items submenu from the Model
menu.
67
Graphical user interface
• You can adjust the position of the viewpoint in Model view by dragging the left mouse button
to display it visually.
Menu commands
File Menu
Creates an empty Metashape project.
New
Opens Metashape project file.
Open...
Appends existing Metashape project file to the
Append...
current one.
Saves Metashape project file.
Save
Saves Metashape project file with a new name.
Save As...
Saves sparse / dense point cloud.
Export Points...
Saves 3D model.
Export Model...
Exports camera positions, orientation data and tie
Export Cameras...
points measurements.
Exports masks.
Export Masks...
Exports model texture.
Export Texture...
Exports spherical panorama for camera stations.
Export Panorama...
Exports the images corresponding to the original
Convert Images...
camera with distortion or color correction applied.
Generates lenticular images for the current scene.
Render Photos...
Imports camera positions and orientation data.
Import Cameras...
Imports reference data.
Import Reference
Imports masks or creates mask from model or
Import Masks...
background.
Imports points in a form of the dense point cloud.
Import Points...
Imports polygonal mesh model.
Import Model...
Imports texture and applies it to the current model.
Import Texture...
Imports laser scans data.
Import Laser Scans
Imports video in a form of frame sequence and
Import Video...
saves the extracted frames as images.
Uploads generated products (points, textured mesh
Upload Data...
models ) to one of the supported web-sites.
Closes the application window.
Exit
68
Graphical user interface
Edit Menu
Undo the last editing operation.
Undo
Redo the previously undone editing operation.
Redo
Enable or disable snap type (Axis, Vertex, Edge,
Snap type
2D snap).
Adds current selection to the mask.
Add selection
Subtracts current selection from the mask.
Subtract selection
Inverts current selection for the photo.
Invert Selection
Inverts current selection of faces / points / cameras.
Invert Selection
Grows current selection of mesh faces.
Grow Selection
Shrinks current selection of mesh faces.
Shrink Selection
Removes selected faces from the mesh or selected
Delete Selection
points from the point cloud.
Crops selected faces / points.
Crop Selection
Inverts mask for the current photo.
Invert Mask
Resets mask for the current photo.
Reset mask
Rotates the photo clockwise.
Rotate right
Rotates the photo counterclockwise.
Rotate left
View Menu
Increases magnification in the active view mode.
Zoom In
Decreases magnification in the active view mode.
Zoom Out
Resets the viewport to display the complete model
Reset View
or photo.
Saves screenshot of current view of the project
Capture view
(Model/Photo)
Shows or hides Workspace pane.
Workspace
Shows or hides Animation pane.
Animation
Shows or hides shape Properties pane.
Properties
Shows or hides Photos pane.
Photos
Shows or hides Console pane.
Console
Shows or hides Jobs pane.
Jobs
Shows or hides Toolbar.
Toolbar
69
Graphical user interface
View Menu
Switches to or from Full Screen mode.
Full Screen
Workflow Menu
Loads additional photos to be processed by
Add Photos...
Metashape.
Loads additional photos from folders to be
Add Folder...
processed by Metashape.
Estimates camera positions and sparse point cloud.
Align Photos...
Generates dense point cloud.
Build Dense Cloud...
Generates polygonal mesh model.
Build Mesh...
Generates texture map for the mesh model.
Build Texture...
Aligns multiple chunks.
Align Chunks...
Merges multiple chunks into the single chunk.
Merge Chunks...
Opens Batch Process dialog box.
Batch Process...
Model Menu
Switches to navigation mode.
Navigation
Rectangle selection tool for the elements of the
Rectangle Selection
Model view.
Circle selection tool for the elements of the Model
Circle Selection
view.
Free-form selection tool for the elements of the
Free-form Selection
Model view.
Selects faces / points based on the specified
Gradual selection...
criterion.
Switches to object movement tool.
Move Object
Switches to object rotation tool.
Rotate Object
Switches to object scaling tool.
Scale Object
Resets the transformations applied to the object.
Reset Transform
Switches to volume movement tool.
Move Region
Switches to volume rotation tool.
Rotate Region
Switches to volume resize tool.
Resize Region
Resets the transformations applied to the volume to
Reset Region
default.
Displays or hides Basemap.
Show Basemap
70
Graphical user interface
Model Menu
Displays or hides camera positions estimated
Show Cameras
during image alignment.
Displays or hides image thumbnails in the camera
Show Thumbnails
placeholders.
Displays or hides item labels.
Show Labels
Displays or hides region selector.
Show Region
Displays or hides animation track.
Show Animation
Displays or hides the trackball.
Show Trackball
Displays or hides the auxiliary on-screen
Show Info
information.
Displays or hides the grid.
Show Grid
Displays all enabled aligned chunks in project
Show Aligned Chunks
workspace.
Displays all elements at the same time.
Show All
Hides all elements.
Hide All
Displays sparse point cloud reconstructed during
Point Cloud
photo alignment.
Displays sparse point cloud colored by variance.
Point Cloud Variance
Displays dense point cloud.
Dense Cloud
Displays dense point cloud colored according to
Dense Cloud Confidence
the confidence values.
Displays 3D model in the shaded mode with
Model Shaded
vertices colored with interpolated colors.
Displays 3D model in the solid mode.
Model Solid
Displays 3D model in the wireframe mode.
Model Wireframe
Displays 3D model in the textured mode.
Model Textured
Displays 3D model with vertices colored according
Model Confidence
to the confidence values.
Switches visualization view between Perspective
Perspective/Orthographic
and Orthographic.
Enables or disables stereo view mode according to
Stereo Mode
the parameters specified in Preferences dialog.
Switches viewport to one of the predefined views.
Predefined views
Photo Menu
Switches to navigation mode.
Navigation
71
Graphical user interface
Photo Menu
Rectangle selection tool.
Rectangle Selection
Intelligent Scissors selection tool.
Intelligent Scissors
Intelligent Paint selection tool.
Intelligent Paint
Magic Wand selection tool.
Magic Wand
Opens next photo from the list in the Photos pane.
Open Next Photo
Opens previous photo from the list in the Photos
Open Previous Photo
pane.
Turns mask shading on or off.
Show Masks
Displays or hides the depth maps overlay.
Show Depth Maps
Displays or hides item labels.
Show Labels
Displays or hides tie point projections used for the
Show Points
camera alignment.
Displays or hides the residuals of each point.
Show Point Residuals
Displays all elements at the same time.
Show All
Hides all elements.
Hide All
Tools Menu
Builds sparse point cloud based on the estimated
Build Point Cloud...
camera parameters available.
Thins sparse point cloud by reducing the number
Thin Point Cloud...
of projections on the individual photos to the given
limit.
Displays View Matches dialog for visual
View Matches...
representation of the common tie points between
the image pairs.
Inverts normals for the selected points of the dense
Invert Point Normals...
cloud.
Selects dense cloud points according to the masks
Select Points by Masks...
of the selected images.
Selects dense cloud points according to the color
Select Points by Color...
and tolerance.
Filters the points in the dense cloud according to
Filter By Confidence
the calculated confidence value.
Filters the points in the dense cloud according to
Filter By Selection
the selected points.
Resets all applied dense cloud filters.
Reset Filter
Permanently removes all deleted points from the
Compact Dense Cloud...
dense cloud.
72
Graphical user interface
Tools Menu
Restores all deleted points of the dense cloud that
Restore Dense Cloud...
were once marked as removed.
Updates statistics of the dense cloud, including
Update Dense Cloud...
point numbers and assigned classes.
Starts photoconsistent mesh refinement operation.
Refine Mesh...
Decimates mesh to the target face count.
Decimate Mesh...
Smooths mesh.
Smooth Mesh...
Closes holes on the model surface.
Close Holes...
Applies colors to the mesh vertices basing on
Colorize Vertices...
source images or point cloud.
Starts resize texture operation for mesh.
Resize texture...
Collects and displays mesh statistics.
View Mesh Statistics...
Displays mesh UV mapping.
View Mesh UVs...
Filters the faces of the polygonal model according
Filter By Selection
to the selected points.
Resets all applied mesh filters.
Reset Filter
Displays lens calibration dialog.
Calibrate Lens...
Displays the calibration board on screen.
Show Chessboard...
Displays camera calibration dialog box.
Camera Calibration...
Starts the optimization of exterior and interior
Optimize Cameras...
parameters of the aligned cameras.
Displays the color correction dialog for the
Calibrate Colors...
brightness and white balance compensation of the
images.
Displays the primary channel selection dialog.
Set Primary Channel...
Adjusts image brightness and contrast for more
Set Brightness...
convenient display.
Displays preferences dialog box.
Preferences...
Help Menu
Displays help contents.
Contents
Checks if Metashape update is available for
Check for Updates...
download.
Displays the activation dialog for activation /
Activate Product...
deactivation of the product using the activation
key.
73
Graphical user interface
Help Menu
Displays program information, version number and
About Metashape...
copyright.
Toolbar buttons
General commands
Creates a new Metashape project file.
New
Opens a Metashape project file.
Open
Saves a Metashape project file.
Save
3D view commands
Undo the last editing operation.
Undo
Redo the previously undone editing operation.
Redo
Navigation tool.
Navigation
Rectangle selection tool.
Rectangle Selection
Circle selection tool.
Circle Selection
Free-form selection tool.
Free-Form Selection
Resets current selection.
Reset Selection
Volume translation tool.
Move region
Volume resize tool.
Resize region
Volume rotation tool.
Rotate Region
Resets region according to the actual point cloud.
Reset Region
Model translation tool.
Move Object
Model rotation tool.
Rotate Object
Model resize tool.
Scale Object
Resets all transformations applied to the model.
Reset Transform
Removes selected faces / points.
Delete Selection
Crops selected faces / points.
Crop Selection
3D view settings
Increases magnification.
Zoom In
Decreases magnification.
Zoom Out
74
Graphical user interface
3D view settings
Resets model view.
Reset View
Displays sparse point cloud reconstructed during
Point Cloud
image alignment.
Displays sparse point cloud colored by variance.
Point Cloud Variance
Displays dense point cloud model.
Dense Cloud
Displays dense point cloud colored according to
Dense Cloud Confidence
the confidence values.
Displays 3D model in the shaded mode with
Model Shaded
vertices colored with interpolated colors.
Displays 3D model in the solid mode.
Model Solid
Displays 3D model in the wireframe mode.
Model Wireframe
Displays 3D model with vertices colored according
Model Confidence
to the confidence values.
Displays 3D model in the textured mode.
Model Textured
Displays diffuse map for the textured mode.
Diffuse Map
Displays normal map for the textured mode.
Normal Map
Displays occlusion map the textured mode.
Occlusion Map
Displays or hides basemap.
Show Basemap
Displays or hides camera positions, reconstructed
Show Cameras
during image alignment.
Displays or hides image thumbnails in the camera
Show Thumbnails
placeholders.
Displays or hides enabled aligned chunks.
Show Aligned Chunks
75
Graphical user interface
Hot keys
For convenient operation in the Metashape it is possible to use Hot keys. Below is a list of default hot keys
that can be customized in the Preferences dialog. Select the Customize button at the Shortcuts menu item.
General
Create new project Ctrl + N
Save project Ctrl + S
Open project Ctrl + O
Run Script Ctrl + R
Full Screen F11
Model View
Undo (only for Delete, Masking and Close Holes Ctrl + Z
operations)
Redo (only for Delete, Masking and Close Holes Ctrl + Y
operations)
Switch between navigation and any other Space
previously selected mode
76
Graphical user interface
Zoom In Ctrl + +
Zoom Out Ctrl + -
Reset view 0
Switch to stereoview mode 9
Switch between orthographic and perspective view 5
modes
Change the angle for perspective view Ctrl + mouse wheel
Predefined Views
Top 7
Bottom Ctrl + 7
Right 3
Left Ctrl + 3
Front 1
Back Ctrl + 1
Rotate View
Rotate Up 8
Rotate Down 2
Rotate Left 4
Rotate Right 6
Photo View
Next photo (according to Photos pane order) Page Up
Previous photo (according to Photos pane order) Page Down
Navigation mode V
Selection Tools
Rectangle selection M
Intelligent scissors L
Intelligent paint P
Magic wand W
Add selection Ctrl + Shift + A
Subtract selection Ctrl + Shift + S
Invert selection Ctrl + Shift + I
77
Appendix B. Supported formats
Images
Input formats Undistort formats
JPG JPG
TIFF TIFF
PNG PNG
BMP BMP
OpenEXR OpenEXR
JPEG 2000 JPEG 2000
TARGA
Digital Negative (DNG)
Portable Bit Map (PGM, PPM)
Multi-Picture Object (MPO)
Norpix Sequence (SEQ)
AscTec Thermal Images (ARA)
Camera calibration
Import formats Export formats
Agisoft Camera Calibration (*.xml) Agisoft Camera Calibration (*.xml)
Australis Camera Parameters (*.txt) Australis Camera Parameters (*.txt)
Australis v.7 Camera Parameters (*.txt) Australis v.7 Camera Parameters (*.txt)
PhotoModeler Camera Calibration (*.ini) PhotoModeler Camera Calibration (*.ini)
3DM CalibCam Camera Parameters (*.txt) 3DM CalibCam Camera Parameters (*.txt)
CalCam Camera Calibration (*.cal) CalCam Camera Calibration (*.cal)
Inpho Camera Calibration (*.txt) Inpho Camera Calibration (*.txt)
USGS Camera Calibration (*.txt) USGS Camera Calibration (*.txt)
Pix4D Camera Calibration (*.cam) Pix4D Camera Calibration (*.cam)
OpenCV Camera Calibration (*.xml) OpenCV Camera Calibration (*.xml)
Photomod Camera Calibration (*.x-cam) Photomod Camera Calibration (*.x-cam)
Z/I Distortion Grid (*.dat) Z/I Distortion Grid (*.dat)
78
Supported formats
Tie points
Import tie points Export tie points
Bundler (*.out) Bundler (*.out)
BINGO (*.dat)
ORIMA (*.txt)
PATB (*.ptb)
Summit Evolution Project (*.smtxml)
Blocks Exchange (*.xml)
Universal 3D (*.u3d)
Autodesk DXF (*.dxf)
79
Supported formats
Mesh model
Import mesh Export mesh
Wavefront OBJ (*.obj) Wavefront OBJ (*.obj)
3DS models (*.3ds) 3DS models (*.3ds)
COLLADA (*.dae) COLLADA (*.dae)
Stanford PLY (*.ply) Stanford PLY (*.ply)
Alembic (*.abc) Alembic (*.abc)
STL models (*.stl) STL models (*.stl)
OpenCTM models (*.ctm) VRML models (*.wrl)
Universal 3D models (*.u3d) Universal 3D models (*.u3d)
Autodesk FBX (*.fbx) Autodesk FBX (*.fbx)
Autodesk DXF (*.dxf) Autodesk DXF Polyline (*.dxf)
Autodesk DXF 3DFace (*.dxf)
Binary glTF (*.glb)
X3D models (*.x3d)
OpenSceneGraph (*.osgb)
Adobe PDF (*.pdf)
Texture maps
Import texture Export texture
JPG JPG
TIFF TIFF
PNG PNG
BMP BMP
TARGA TARGA
JPEG 2000 JPEG 2000
OpenEXR OpenEXR
Portable Bit Map
Digital negative
Multi-Picture Object
Norpix Sequence File
AscTec Thermal Images
80
Appendix C. Camera models
A camera model specifies the transformation from point coordinates in the local camera coordinate system
to the pixel coordinates in the image frame.
The local camera coordinate system has origin at the camera projection center. The Z axis points towards
the viewing direction, X axis points to the right, Y axis points down.
The image coordinate system has origin in the middle of the top-left pixel (with coordinates (0.5, 0.5)).
The X axis in the image coordinate system points to the right, Y axis points down. Image coordinates are
measured in pixels.
Equations used to project a points in the local camera coordinate system to the image plane are provided
below for each supported camera model.
(u, v) - projected point coordinates in the image coordinate system (in pixels),
81
Camera models
Frame cameras
x=X/Z
y=Y/Z
r = sqrt(x2 + y2)
v = h * 0.5 + cy + y'f
Fisheye cameras
x0 = X / Z
y0 = Y / Z
r0 = sqrt(x02 + y02)
x = x0 * tan-1r0 / r0
y = y0 * tan-1r0 / r0
r = sqrt(x2 + y2)
v = h * 0.5 + cy + y'f
82