Photogrammetry As An Engineering Design Tool
Photogrammetry As An Engineering Design Tool
Photogrammetry as an
Engineering Design Tool
Ana Pilar Valerga Puerta, Rocio Aletheia Jimenez-Rodriguez,
Sergio Fernandez-Vidal and Severo Raul Fernandez-Vidal
Abstract
1. Reverse engineering
2. Photogrammetry
1
Product Design
Figure 1.
Classification of 3D scanning technologies.
Figure 2.
Structure of the motion algorithm [9].
2
Photogrammetry as an Engineering Design Tool
DOI: http://dx.doi.org/10.5772/intechopen.92998
3. Fundamentals of photogrammetry
where λ = scaling factor; M = rotation matrix; Xo, Yo, and Zo = the position of the
perspective center in the object’s space; and pn = (xn, yn)T and Pn = (Xn, Yn,
Zn)T = target n coordinates at the image plane and the space of the object, respec-
tively. The above equation manipulated algebraically produces the well-known
collinearity equations that relate the location of destination nth in the space of
objects with the corresponding point in the plane of the image:
where mij (i, j = 1, 2, 3) = elements of the rotation matrix M which are functions
of the Euler orientation angles (ω, ф, к), which are essentially the angles of tilt,
rotation and rotation of the camera in the object space (Eq. (4)–(11))
The plane of the image can be transformed analytically into its X, Y, and Z
coordinates in global space. Photogrammetry is effective and computationally sim-
ple. It should be noted that its algorithm is based on definitions of both interior and
exterior orientations. In a photographic system, if the internal parameters of a
3
Product Design
camera are known, any spatial point can be fixed by the intersection of two beams
of light that are projected.
There are two main factors that induce photogrammetry measurement errors:
System error due to lens distortion and random error due to human factors.
1. System error due to lens distortion. It causes a point in the image in the plane
to move from its true position (x, y) to a disturbed position. The coordinates of
any point in the image can be compensated with Eqs. (13)–(14):
x´n ¼ xn þ dx (13)
y´n ¼ yn þ dy (14)
In the lens, the largest error occurs at the point of the projected image. There-
fore, dx, dy can be broken down by Eqs. (15)–(16):
The main camera and photography parameters are focal length, focal point, bias,
distortion, and pixel error; they will allow more accurate calibration [18] and are
shown in Figure 3.
4
Photogrammetry as an Engineering Design Tool
DOI: http://dx.doi.org/10.5772/intechopen.92998
Figure 3.
Scheme of operation of a camera objective.
Included in the optical part of the camera, it is in charge in projecting the image
that crosses it on the same plane and in outstanding conditions of sharpness.
Therefore, it is a matter of focusing on the objects that are at equal distance on the
focal plane. From certain distance, all the objects will be projected on the same
plane. The light points are transmitted to an element that composes the scenario. As
a result of diffraction, this is shown as a circular point with a halo around it and
concentric rings, named Airy discs. Suppressing them is unfeasible because it is a
physical light effect. Even so, it would be desirable for such rings to be as diffuse
and thin as possible [17, 19].
Its resolving capacity depends on two parameters: aberrations and diffraction.
One of the main functions of the objective is to suppress aberrations. When the
diaphragm is closed, the aberrations are placated, and the only limiting factor is
diffraction. When the diaphragm is opened, diffraction diminishes its significance
in the wake of aberrations, which add up to force [20].
This parameter is measured from the optical center of the lens to the focal plane,
when the camera focused toward the infinity [5, 21]. Normal lenses are those which
have a distance close to the diagonal of the cliché. The representation of the focal
length is shown in Figure 4.
Relative aperture (Ab) is the connection of the lens diameter (D) and its focal
length (f) (Eq. (17)).
It is shown by the denominator, known as brightness or “f-number.” In a
different way, the aperture is the span through which light enters to be captured by
the sensor. The more spacious the opening will be, the more light will enter the
sensor as the number becomes smaller [4, 7]:
5
Product Design
D
Ab ¼ (17)
f
This is the viewing angle of the camera and is closely related to the focal length
and dimension of the sensor [8, 22]. A schematic representation is proposed in
Figure 5.
5.5 Shutter
It is a mechanism that keeps the light passing through the lens into the closed
camera. At certain intervals of time, it has the ability to open, allowing the passage
of light so that the film can be impressed. The opening time can be set [21].
Figure 4.
Representation and focal length types on a camera.
Figure 5.
Focal distances and corresponding angles.
6
Photogrammetry as an Engineering Design Tool
DOI: http://dx.doi.org/10.5772/intechopen.92998
aperture, and under some pretexts, large aperture systems have more pronounced
depths of focus than low aperture systems, even if the depth of field is small [19].
Depth of field is the area of sharp reproduction seen in the photograph. In this
one, there are some objects observed which are located at a certain distance, as well
as others more distant or adjacent to them [20].
5.8 Sensor
Its function is to modify the light received in order to obtain a digital systemati-
zation. The sensor is called pixel in its minimum element. A digital image consists of
a set of pixels. The technology based on complementary metal oxide semiconductor
(CMOS) sensors is the most applied. The sensors consist of a semiconductor and
sensitive material in the visible spectrum, between 300 and 1000 nm [10]. Charge-
coupled device (CCD) sensors are becoming obsolete due to the cost and speed of
processing images.
The comparison reading of the information in the CMOS sensors has the advan-
tage of obtaining enough captures, obtaining readings using less time and with
greater flexibility. Using a high dynamic range of work, high contrasts and a correct
display of objects are achieved. In terms of quality, the physical size of the sensor is
more significant than the number of cells or resolution. A large unit may allow
higher-quality photographs to be taken than another sensor with a higher resolution
but with a smaller surface [23].
As far as color is concerned, it must be seen that color is just a human visual
perception. In order to be able to glimpse the color of an object, it is necessary to
have a light source and something that reflects this light. A color is represented in
digital format by applying a system of representation. The most commonly used is
the RGB system. To represent a color, the exact percentages of primary red, pri-
mary green, and primary blue (RGB, red, green, blue) must be available. By this
way, the color is displayed through the implementation of three numbers [24].
5.9 Diaphragm
5.10.1 Focus
The first step in taking a picture is focusing. The most commonly used types of
automatic focusing are [25]:
7
Product Design
• Dual pixel. This method uses more focus points along the sensor than the
PDAF. This system uses two photodiodes at each pixel to compare minimal
dissimilarities. This is the most effective focusing technology.
• Contrast detection. It is the oldest of the three systems exposed. Its operation
theoretically bases that the contrast of an image is greater, and its edges are
appreciated in a clearer way, when it is focused correctly. The disadvantage is
its slowness.
5.10.2 Perspective
x X
¼ (18)
X Z
Figure 6.
Solution to (a) different openings, (b) shutter speeds, and (c) ISO.
8
Photogrammetry as an Engineering Design Tool
DOI: http://dx.doi.org/10.5772/intechopen.92998
Figure 7.
Diagram of the projection of a camera.
y Y
¼ (19)
f Z
where f is the focal length, (X, Y, Z) are the actual coordinates of the point, and P
(x,y) are the coordinates on the projection plane of the image or photograph.
It would be in front of more complex expressions if the planes that contain the
points are not parallel to the one of projection, being indispensable to know the
inclination of the plane having as reference the plane of projection. In practice, in
order to avoid complications in the calculation of coordinates, photographs are
usually taken in a way that the planes are parallel.
5.10.3 Exposure
1. ISO Sensitivity: it indicates the amount of light required to take a picture. The
higher the light, the lower the ISO.
2. Diaphragm opening: it inspects the light reaching the focal plane, along with
the shutter speed, and regulates the depth of field of the photograph.
3. Shutter speed: shutter opening time allows light to reach the sensor. The higher
the shutter speed, the lower the percentage of light reaching the sensor.
When a sensor has the ability to capture as many tones (dynamic range) and
information (light) as its ability allows, the picture is perfectly exposed.
It measures the amount of light and dark tones that a camera has the ability to
capture in the same picture. It shows the amount of tonal nuances that a camera is
capable of capturing, measurable by contrast and sharpness.
Contrast and sharpness are based on the differentiation of tonality with which a
pair of white and black lines are obtained, captured, or reproduced. It is measurable
of the degree of detail, being 100% when both lines can be perfectly differentiated
as pure whites and blacks. Resolution and contrast are closely related concepts.
9
Product Design
Figure 8.
Contrast sensitivity change as a function of the spatial frequency of the target.
If the contrast falls below 5%, it is difficult to observe any detail, which is shown
more clearly and distinctly the higher it is. Frequency and modulation are shown in
the way they are altered when light passes through the different optical components
of the lens of the photographed image, thanks to contrast transfer functions. As the
viewer moves away, a substantial loss of contrast begins to be noticed [12].
By performing a contrast correction, different filters are applied to the central
zones instead of the peripheral zones. An example of contrast and resolution is
shown in Figure 8.
5.10.5 Aberrations
2. Shape aberrations: the point is shown as a point but with a different position to
the one arranged by means of paraxial approximation. This is a systematic
error and can be of two types: field curvature and distortions.
• Field curvature: defect when creating the image, being curved instead of
flat. It is difficult to correct the aberration, but it can be mitigated in a low
percentage.
• Distortion: only affects the shape of the image. It occurs due to the
difference in the scale of reproduction of the image off-axis. If an object
with straight lines is photographed, such as a square, the center lines will
appear straight, and the edge lines will curve inward or outward producing
the so-called barrel or cushion distortions. This aberration is not corrected
by closing the diaphragm. This error affects the tone of the image and needs
to be corrected.
10
Photogrammetry as an Engineering Design Tool
DOI: http://dx.doi.org/10.5772/intechopen.92998
3. Illumination: sufficient light bulb. In most cases, natural light is not sufficient,
and it is necessary to use spotlights or other artificial elements.
Image quality is a prerequisite for working with it properly. There are two main
characteristics that define it:
5.10.8 Histogram
This is a visual tool very useful for the study of digital images. With the naked
eye, it is possible to study the contrast or the distribution of intensities, because it
follows the following discrete function of Eq. (20):
f ðxi Þ ¼ ni (20)
where x is the level of gray or color and n is the number of pixels in the image
with this value. The histogram is normalized in values ranging from 0 to 1. In
Figure 9 it is possible to see their different zones [18, 31].
The most common errors in the image, which prevent good image quality, can
be identified in the histogram and are muted tones, black areas, overexposure or
burned areas, and backlight. In order to know that a good image is acquired, the
best thing is to have a histogram that has the shape of a Gauss bell, that is to say,
that has the most information in the central part and less in the extremes. Another
important point is that the histogram must embrace and reach both ends, so as to
ensure that there are blacks and whites in the photograph.
11
Product Design
Figure 9.
Histogram areas.
5.10.9 Binarization
where 0/1 represents the black/white values and f is the value of the gray tone of
the coordinates (x,y) [32]. Figure 10 shows a grayscale image versus a binary.
To obtain an image with sufficient quality, the binarization must correspond
with white pixels to the objects of interest, being the blacks of the environment. If
the object of interest turns out to be darker than the environment, a reversal is
applied after the binarization. The most important point in the process is the
calculation of the threshold. There are different methods for this: histogram, clus-
tering, entropy, similarity, spatial, global, and local.
The setting of the threshold value is latent, due to its difficulty, in all methods.
The techniques are supported by statistics applied to the histogram. They are as
follows: carry error method, Otsu method, and Saulova’s pixel deviation method.
Figure 10.
Grayscale (left) and binary (right).
12
Photogrammetry as an Engineering Design Tool
DOI: http://dx.doi.org/10.5772/intechopen.92998
Figure 11.
Types of lens distortion.
the value of a pixel p into the position (x,y), always taking into account the values of
the adjacent pixels. For this operation, a weighted sum of the values of the neigh-
boring points of this point p is required. A mask (h), behaving like a filter, is in
charge of exposing the values of the weighting. The size of the mask varies
according to the pixels used.
These operations modify the spatial coordinates of the image. There are several
operations that are easy to understand and apply, such as interpolation, rotation,
rectification, and distortion correction.
Due to the geometry of the lens, it reproduces a square object with variations in
its parallel lines. There are three types of distortion: barrel, pincushion, and mus-
tache (combination of the first two) (Figure 11) [25, 33]. This error is negligible in a
photograph of a natural scene, but to take engineering measurements and obtain a
virtual object, it is necessary to compensate for the distortion. There is a mathe-
matical model for the treatment of distortion.
The barrel distortion is centered and symmetrical. Therefore, to correct the
distortion of a certain point, a radial transformation is performed, expressed math-
ematically in Eq. (22):
qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
^ xd
x 2 2 x xd
¼ L ð x x d Þ þ y yd (22)
^y yd y yd
where ðx^, ^yÞ represents the result of the distortion correction at point (x, y), (xd,
yd) represents the center of the distortion which is usually a point near the center of
the image, and finally the radial function L(r) determines the magnitude of the
distortion correction as a function of the distance from the point to the center of
distortion [34].
The radial function L(r) is performed by applying two strategies. The first one
gives rise to the so-called polynomial models (Eq. (23)):
1
LðrÞ ¼ (24)
1 þ k1 r2 þ k2 r4 þ … þ kn r2n
13
Product Design
Figure 12.
Visual example of photo rectification.
The values k1–kn are called distortion model parameters. These values, together
with the distortion center coordinates (xd, yd), completely represent the distortion
model. The distortion of the lens is represented by the ki coefficients. They are
obtained from a known calibration image.
Image correction is necessary because either it is difficult to keep the optical axis
vertical at all points of the shot or the axis is tilted toward the vertical. Vertical
images are obtained free of displacement because of the inclination of the shot but
still have inclinations, product of the depth of the workpiece. Displacements can be
suppressed by applying differential grinding or orthorectification process. In the
original digital image or a scan, the technique is applied pixel by pixel. In a scanned
image, the initial data are the coordinates of the control points. The procedure is
divided into two steps:
After this process, it is necessary to know that all the pixels of the resulting
orthophotography have their level of gray, performing a digital resampling [17, 34].
Figure 12 shows an unrectified (left) and rectified (right) photograph.
Several resamples are made on the initial image. Three resampling methods are
regularly used: bilinear interpolation, nearest neighbor, and bicubic convolution.
The transformations to be applied to the images are [19] Helmert transformation;
affine transformation; polynomial transformation; and two-dimensional projective
transformation.
14
Photogrammetry as an Engineering Design Tool
DOI: http://dx.doi.org/10.5772/intechopen.92998
The efficiency with which the techniques are used will be linked to the final quality
of the reconstruction.
The stereoscopic scene analysis system presented by Koch uses image matching,
object segmentation, interpolation, and triangulation techniques to obtain the 3D
point density map. The system is divided into three modules: sensor processing,
image pair processing, and model-based sequence processing.
Pollefeys features a 3D reconstruction process based on well-defined stages. The
input is an image sequence, and the output of the process is a 3D surface model. The
stages are the following: image ratio, structure and motion recovery, dense
matching, and model construction.
Another proposal is expressed by Remondino. He presents a 3D reconstruction
system following these steps: image sequence acquisition and analysis, image
calibration and orientation, matching process and the generation of points, and 3D
modeling [18].
where
dpðsÞ
αn ðsÞ ¼ pðsÞx (26)
ds
15
Product Design
Figure 13.
Harmonic homology of the figure and its transformation to orthogonal projection [35].
Figure 14.
The matching results of the line segments in four main planes [36].
16
Photogrammetry as an Engineering Design Tool
DOI: http://dx.doi.org/10.5772/intechopen.92998
The tetrahedron grid can only be used to represent the geometrical structure of
geological objects. The natural characteristics of geological objects are reflected in
their different attributes, such as different rock formations, different contents of
mineral bodies, etc. It is defined that the attribute value of the internal point can be
linearly interpolated from the attribute values in four vertices in a tetrahedron. But
the attributes could change suddenly between different formations and different
mineral bodies. To cope with sudden changes, interpolation of the tetrahedron is
needed that can only be applied to six sides of a tetrahedron. Those interpolated
points are only used as time data for the following processing [37].
This section presents a robust and precise system for the 3D reconstruction of
real objects with shapes and textures in high resolution. The reconstruction method
is passive, and the only information required is 2D images obtained with a camera
calibrated from different viewing angles as the object rotates on a rotating plate.
The triangle surface model is obtained through a scheme that combines the octree
construction and the walking cube algorithm. A texture mapping strategy based on
surface particles is developed to adequately address photographic-related problems
such as inhomogeneous lighting, lights, and occlusion [38]. To conclude, the results
of the reconstruction are included to demonstrate the quality obtained (Figure 15).
The scheme combining octree construction and isolevel extraction through
marching cubes is presented for the problem concerning the shape of the silhouette.
The use of octree representation allows to reach very high resolutions, while the
method of fast walking cubes is adapted through a properly defined isolevel func-
tion to work with binary silhouettes, resulting in a mesh of triangles with vertices
precisely located in the visual object.
Calibration is performed on the camera and rotary table. One of the problems
found is the discontinuity of the texture due to the nonhomogeneous lighting in
different parts of the element due to shadows.
Next, the octree is represented. An octree is a hierarchical tree structure that can
be used to represent volumetric data in terms of cubes of different sizes. Each octree
node corresponds to a cube in the octree space that is entirely within the object. This
opens up different possibilities: voxels, particles, triangles, and more complicated
Figure 15.
Flowchart to the reconstruction of objects.
17
Product Design
Figure 16.
From cube to triangulation, adapted from [38].
2. All the pixels of an image are made to coincide with those of the neighboring
images so that the system can reconstruct these points.
18
Photogrammetry as an Engineering Design Tool
DOI: http://dx.doi.org/10.5772/intechopen.92998
Figure 17.
Steps to obtain the 3D model, adapted from [39].
approached with a triangular grid, to reduce geometric complexity and adapt the
model to the requirements of the computer graphic display system. Then construct
a corresponding 3D mesh by placing the triangle vertices in 3D space according to
the values found in the corresponding depth map. To reconstruct more complex
shapes, the system must combine multiple depth maps. Finally, it is provided with
texture.
It is used for medical purposes in many cases, as a base for implants, splints, etc.
The process consists of the following parts: acquisition and analysis of the image
sequence; calibration and orientation of the images; matching process on the sur-
face of the human body; and generation and modeling of the point cloud. Once the
necessary images have been obtained from different points of view, the calibration
and orientation of the images are carried out.
The choice of the camera model is often related to the final application and the
required accuracy. The correct calibration of the sensor used is one of the main
objectives. Another important point is image matching [40].
To evaluate the quality of the matching results, different indicators are used: an
ex post standard deviation of the least squares adjustment, the standard deviation of
the change in the x-y directions, and the shift from the initial position in the x-y
directions. The performance of the process, in the case of uncalibrated images, can
only be improved with a local contrast enhancement of the images.
Finally, 3D reconstruction and modeling of the human body shape is performed.
The 3D coordinates of each matching triplet are calculated through a forward
intersection. Using collinearity and the results of the orientation process, the 3D
paired points are determined with a solution of least squares. For each triplet of
images, a point cloud is calculated, and then all the points are joined together to
create a unique point cloud. A spatial filter is applied to reduce noise and obtain a
more uniform point cloud density. Figure 18 shows the results before and after
filtering (approximately 20,000 points, left); a view of the recovered point cloud
with pixel intensity (center); and a 3D human model (right).
The system is composed of two main modules. The first one is in charge of image
processing, to determine the depth map in a pair of views, where each pair of
successive views follows a sequence of phases: detection of points of interest,
correspondence of points, and reconstruction of these. In this last phase, the
19
Product Design
Figure 18.
3D reconstruction of a human body, adapted from [40].
parameters that describe the movement (rotation matrix R and translation vector T)
between the two views are determined. This sequence of steps is repeated for all
successive pairs of views of the set.
The second module is responsible for creating the 3D model, for which it must
determine the total 3D points map generated. In each iteration of the previous
module, the 3D mesh is generated by applying Delaunay’s triangulation method.
The results obtained from the process are modeled in a virtual environment to
obtain a more realistic visualization of the object [16].
The number of detected minutiae is related to the number of reconstructed 3D
points and the quality of that reconstruction (higher number of details). Therefore,
the higher the number of points on the map, the more detailed areas are obtained. In
some cases this does not apply, due to the geometry of the object, for example, in a
cube, more points can result in a distorted object.
7. Conclusion
Acknowledgements
The authors would like to thank the call for Innovation and Teaching
Improvement Projects of the University of Cadiz and AIRBUS-UCA Innovation
Unit (UIC) for the Development of Advanced Manufacturing Technologies in the
Aeronautical Industry.
20
Photogrammetry as an Engineering Design Tool
DOI: http://dx.doi.org/10.5772/intechopen.92998
Conflict of interest
Author details
© 2020 The Author(s). Licensee IntechOpen. This chapter is distributed under the terms
of the Creative Commons Attribution License (http://creativecommons.org/licenses/
by/3.0), which permits unrestricted use, distribution, and reproduction in any medium,
provided the original work is properly cited.
21
Product Design
References
[1] Rolin R, Antaluca E, Batoz JL, [8] Katz D, Friess M. Technical note: 3D
Lamarque F, Lejeune M. From point from standard digital photography of
cloud data to structural analysis through human crania - A preliminary
a geometrical hBIM-oriented model. assessment. American Journal of
Journal of Cultural Heritage. 2019;12: Physical Anthropology. 2014;154:
1-26. DOI: 10.1145/3242901 152-158. DOI: 10.1002/ajpa.22468
22
Photogrammetry as an Engineering Design Tool
DOI: http://dx.doi.org/10.5772/intechopen.92998
23
Product Design
24