22.10.2012 Views

From digital image series to 3D models - Arctron

From digital image series to 3D models - Arctron

From digital image series to 3D models - Arctron

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

Martin Schaich<br />

<strong>From</strong> <strong>digital</strong> <strong>image</strong> <strong>series</strong> <strong>to</strong> <strong>3D</strong> <strong>models</strong><br />

Possibilities for software application<br />

<strong>From</strong> <strong>digital</strong> <strong>image</strong> <strong>series</strong> <strong>to</strong> <strong>3D</strong> <strong>models</strong><br />

Three-dimensional computer <strong>models</strong> have a lot of advantages for use in as-built<br />

documentations and damage mapping. This software generates precise <strong>3D</strong> <strong>models</strong><br />

based on a <strong>series</strong> of <strong>digital</strong> <strong>image</strong>s and offers an alternative solution <strong>to</strong> expensive <strong>3D</strong><br />

scanning.<br />

<strong>3D</strong> <strong>models</strong> in a res<strong>to</strong>rational context<br />

<strong>3D</strong> <strong>models</strong> present us with new opportunities that can help us in the complex everyday work<br />

of res<strong>to</strong>ration and cultural heritage preservation. On one hand, they assist in <strong>digital</strong><br />

documentation. Findings and damage areas can be reliably mapped. This is especially<br />

interesting for complex <strong>3D</strong> objects. On the other hand, this technology is capable of virtual<br />

reproduction of works of art. The dissemination of these <strong>models</strong> using internet and<br />

multimedia solutions (including publication and presentation, animations and elaborate<br />

computer films) offers exciting and interesting possibilities.<br />

There are different ways <strong>to</strong> create a <strong>3D</strong> model. The most common one is the <strong>3D</strong> scan, during<br />

which the object is scanned without contact. This can be done using laser scanners,<br />

triangulation scanners or structured-light scanners (for high resolution). If necessary, the<br />

scanning data can be combined with <strong>to</strong>tal station or GPS data for referencing.<br />

In practice, a <strong>digital</strong> <strong>3D</strong> scan (requiring special equipment and knowledge) is way <strong>to</strong>o<br />

laborious and expensive for res<strong>to</strong>rers. To these professionals, the software aSPECT <strong>3D</strong> offers<br />

an alternative. This software generates <strong>3D</strong> <strong>models</strong> from <strong>digital</strong> <strong>image</strong> sequences, which can<br />

be recorded using commercially available <strong>digital</strong> cameras.<br />

Pho<strong>to</strong>graphic recording and <strong>image</strong> quality<br />

At the beginning, a suitable <strong>image</strong> <strong>series</strong> of the object has <strong>to</strong> be recorded. For that, basic<br />

pho<strong>to</strong>graphic knowledge about the correlation between exposure time and aperture, depth of<br />

field and the use of different lenses is – of course – required. Dis<strong>to</strong>rtions, for example<br />

produced by using so-called “fish-eye”-lenses and extreme wide angle lenses, cannot be<br />

processed by the algorithm.<br />

In general, no calibrated cameras are needed. Even small, low-cost compact cameras or<br />

system cameras deliver usable results. Nonetheless, for high quality survey results, semiprofessional<br />

or professional SLR systems are recommended.<br />

The software requires well-focussed <strong>image</strong>s with a sufficient depth of field. Moving shadows<br />

or light points affect the computation process as well as reflections (e.g. from direct flash<br />

light). While homogenous lighting is easy <strong>to</strong> realize in a studio setting, it is not always<br />

possible during outdoor projects. However, experience has shown that usable results can<br />

even be derived from data with difficult content like hard shadows. During the pho<strong>to</strong>graphic<br />

recording, it is important <strong>to</strong> circle the object as “hemispherically” as possible with different


<strong>From</strong> <strong>digital</strong> <strong>image</strong> <strong>series</strong> <strong>to</strong> <strong>3D</strong> <strong>models</strong><br />

vertical angles (one per circle) (Fig. 1). A single panorama recording, during which the<br />

pho<strong>to</strong>grapher stays in one position, does not produce <strong>3D</strong> results.<br />

Fig. 1: aSPECT <strong>3D</strong> res<strong>to</strong>ration project “Old Fritz”. Left: Pictures taken with a commercially available camera. Result after the<br />

bundling with the cameras positioned and oriented in space. The coloured prisms around the <strong>image</strong>s visualize the neighbouring<br />

qualities of the <strong>image</strong>s and the lenses’ focal lengths. Right: First “basic <strong>3D</strong> result” as a dense, coloured point cloud. © ArcTron <strong>3D</strong><br />

GmbH, 2012.<br />

Preselection of <strong>image</strong>s<br />

Before the software aSPECT <strong>3D</strong> can start computing the model, the <strong>image</strong>s have <strong>to</strong> be<br />

checked and sorted with relation <strong>to</strong> the criteria described above. Large projects often<br />

produce <strong>image</strong>s in various stages with very different perspectives and lighting conditions. In<br />

these cases, it is useful <strong>to</strong> sort the <strong>image</strong>s: e.g. all aerial <strong>image</strong>s of the building, all pictures<br />

from the ground, by time of day or room location etc. A similar approach applies <strong>to</strong><br />

archaeological or res<strong>to</strong>ration projects which document stratigraphic layers. For that, the<br />

software offers an <strong>image</strong> database, in which the <strong>image</strong>s can be tagged. Furthermore, all<br />

internal camera information (EXIF-data) concerning sensor, lenses, focal length, shutter<br />

speed, ISO-value etc. are s<strong>to</strong>red and later on used for the computation.<br />

<strong>From</strong> <strong>digital</strong> <strong>image</strong> <strong>series</strong> <strong>to</strong> <strong>3D</strong> point cloud<br />

After checking and preselecting the <strong>image</strong>s, the software computes a <strong>3D</strong> point cloud from the<br />

selected pictures. It utilizes a technology, which has established itself in computer based<br />

<strong>image</strong> processing: SFM (“Structure <strong>From</strong> Motion” 1 ).<br />

Different algorithms (e.g. Sift, ASift, Hesian-affine, MSER etc. 2 ) are used for computing <strong>3D</strong><br />

object information. They identify distinctive points (“features”) in several pictures from<br />

1 Different providers allow working with such solution in the „cloud“. For that, the <strong>image</strong>s have <strong>to</strong> be uploaded<br />

<strong>to</strong> the respective providers, which is hard <strong>to</strong> realize and not really sensible for large <strong>image</strong> <strong>series</strong>, sensitive data<br />

or when operating in remote regions. Generally compare: Remondino, F., El-Hakim, S.F., Gruen, A. & Zhang, L.<br />

(2008): Turning<strong>image</strong>sin<strong>to</strong> <strong>3D</strong>-<strong>models</strong>. IEEE Signal Processing Magazine 25 (4), S. 55-65; Kersten, T.,<br />

Lindstaedt, M., Mechelke, K., & Zobel, K. (2012): Au<strong>to</strong>matische <strong>3D</strong>-Objektrekonstruktion aus unstrukturierten<br />

<strong>digital</strong>en Bilddaten für Anwendungen in Architektur, Denkmalpflege und Archäologie. 32. Wiss.-techn.<br />

Jahrestagung der DGPF (März 2012), Potsdam, Germany.


<strong>From</strong> <strong>digital</strong> <strong>image</strong> <strong>series</strong> <strong>to</strong> <strong>3D</strong> <strong>models</strong><br />

different perspectives. The algorithm also includes the EXIF-data of each picture, e.g. which<br />

camera was used with what kind of sensor and lenses. Pho<strong>to</strong>grammetric procedures are<br />

used in order <strong>to</strong> reconstruct the camera position and orientation. All is done by correlating the<br />

<strong>image</strong>s <strong>to</strong> each other.<br />

In the first step, the <strong>image</strong>s are put in<strong>to</strong> spatial correlation using the so-called bundle block<br />

adjustment 3 . The result is a <strong>3D</strong> point cloud of all identified features. In a subsequent<br />

computational step, this basic data can be transformed in<strong>to</strong> an already very detailed, socalled<br />

“dense”, coloured point cloud 4 . (Fig. 1)<br />

SFM-technology is suitable for objects with amorphous geometries and / or objects with<br />

structured / multi edged surfaces. Artifacts, which display an abundance of corresponding<br />

<strong>image</strong> points and shades lend themselves <strong>to</strong> a successful result. Among the objects<br />

rather unsuited for SFM are unstructured, monochrome, translucent, reflective, evenly<br />

patterned or so-called self-similar surfaces, in which structures and characteristics are<br />

repeated. Such objects produce rather bad or incomplete results.<br />

Fully au<strong>to</strong>mated processes<br />

The computational process for generating <strong>3D</strong> point clouds from <strong>image</strong>s is fully au<strong>to</strong>mated.<br />

That means that the user does not have <strong>to</strong> carry out any more steps after selecting the<br />

<strong>image</strong>s. Processing time is dependant on the number and size of the <strong>image</strong>s. Data banks<br />

containing several hundred or thousand pictures, require high quality and very powerful<br />

computers. Most importantly, these computers need a multiple processor architecture with<br />

64bit processors, a lot of main memory and they need <strong>to</strong> support parallel computational<br />

processes on the graphic board (GPU). Usually, current “workstations” meet these<br />

requirements.<br />

Scaling and georeferencing<br />

Initially, the resulting point cloud is neither scaled nor georeferenced. This means, that the<br />

ratio of the point cloud <strong>to</strong> reality remains unknown. The point cloud needs <strong>to</strong> be allocated <strong>to</strong><br />

definite coordinates in a superordinate coordinate system. Afterwards, the <strong>3D</strong> object is<br />

scaled, meaning the size of the object is within a defined, <strong>to</strong>lerable deviation.<br />

aSPECT <strong>3D</strong> offers different approaches for this procedure. The simplest way is scaling the<br />

object via a scale that was pho<strong>to</strong>graphed along with the object. The GPS and <strong>to</strong>tal station<br />

2 Lowe, D. (2004), Distinctive <strong>image</strong> features from scale-invariant keypoints. International Journal of Computer<br />

Vision 60,2, S. 91-110.<br />

3 th<br />

Wu, C., Agarwal, S., Curless, B. & Seitz, S. (2011): Multicore Bundle Adjustment. 24 IEEE Conference on<br />

Computer Vision and Pattern Recognition. Colorado Springs, USA.<br />

4<br />

Furukawa, Y. & Ponce, J. (2007): Accurate, dense and robust multiview stereopsis. In: Proc. International<br />

Conference on Computer Vision and Pattern Recognition. S. 1-8.


<strong>From</strong> <strong>digital</strong> <strong>image</strong> <strong>series</strong> <strong>to</strong> <strong>3D</strong> <strong>models</strong><br />

interface allows correlating reference points within the point cloud <strong>to</strong> independently<br />

measured control points. By deploying a so-called multi-point transformation, the point clouds<br />

are transformed in<strong>to</strong> the target coordinate system, whereupon the errors are checked and<br />

can be optimized in general (Fig. 2).<br />

Fig. 2: Scaling and georeferencing in aSPECT <strong>3D</strong> . Left: Pho<strong>to</strong>graphs of a house façade as a SFM <strong>image</strong> <strong>series</strong>. Middle: Single<br />

reference points are recorded with a <strong>to</strong>tal station and processed in aSPECT <strong>3D</strong> . Right:: Using a multi-point transformation, the<br />

scaling and georeferencing is carried out with the computation of the errors. © ArcTron <strong>3D</strong> GmbH, 2012.<br />

Larger objects (e.g. roof landscapes, façades) sometimes consist of several point clouds<br />

merged <strong>to</strong>gether, because they were recorded from very different points of view – e.g. via<br />

airborne pho<strong>to</strong>graphy or from the ground. For this purpose, aSPECT <strong>3D</strong> includes an ICPalgorithm,<br />

which can fuse overlapping point clouds <strong>to</strong>gether. This is especially suited for<br />

<strong>image</strong> <strong>series</strong> with very little overlap. Once this step is done, the <strong>3D</strong> point model is finished.<br />

How precise are <strong>3D</strong> <strong>models</strong>?<br />

<strong>3D</strong> <strong>models</strong> have <strong>to</strong> depict the original object as precisely as possible for res<strong>to</strong>ration,<br />

preservative or archaeological purposes. Size and shape need <strong>to</strong> have the smallest possible<br />

deviation. Depending on the deployed survey procedure (<strong>to</strong>tal station, laser scanner etc.),<br />

the accuracies can vary.<br />

The crea<strong>to</strong>rs of the software, which generates <strong>models</strong> based on SFM-<strong>image</strong>s, asked<br />

themselves if aSPECT <strong>3D</strong> could compete with other <strong>3D</strong> methods. In order <strong>to</strong> answer this<br />

question, ArcTron <strong>3D</strong> carried out accuracy analyses during different projects 5 . The results<br />

(below) show, that for both small and large objects, significantly precise <strong>3D</strong> <strong>models</strong> can be<br />

generated.<br />

5<br />

Publications of the author concerning the <strong>to</strong>pic are currently in print or preparation. Compare also: Kersten, T.<br />

&Mechelke, K., Fort Al Zubarah in Katar – <strong>3D</strong>-Modell aus Scanner- und Bilddaten im Vergleich. In:<br />

Luhmann/Müller (Ed.), Pho<strong>to</strong>grammetrie, Laserscanning, Optische <strong>3D</strong>-Messtechnik. Beitr. Oldenburger <strong>3D</strong>-<br />

Tage 2012, S. 89-98.


<strong>From</strong> <strong>digital</strong> <strong>image</strong> <strong>series</strong> <strong>to</strong> <strong>3D</strong> <strong>models</strong><br />

Fig. 3: Project “wall documentation”. Veitsberg, Bad Neustadt a.d. Saale (in cooperation with the University of Jena,<br />

Professorship for Prehis<strong>to</strong>ric and Pro<strong>to</strong>his<strong>to</strong>ric Archaeology). Left: Pho<strong>to</strong>graphic documentation airborne and from the ground.<br />

Middle: Reference surveys with <strong>to</strong>tal station and laser scanner. Right: Deviation result after a “best fit” registration of the SFM<br />

point cloud on<strong>to</strong> the terrestrial laser scan with standard deviation of below 1 cm. © ArcTron <strong>3D</strong> GmbH, 2012.<br />

Example 1: In cooperation with the University of Jena, ArcTron <strong>3D</strong> surveyed a small<br />

representative area of wall fragments in Veitsberg showing early medieval Carolingian<br />

remains (Fig. 3). For the accuracy analysis, the model generated by aSPECT <strong>3D</strong> was<br />

compared <strong>to</strong> <strong>models</strong> generated in a different way. In one test various control points placed in<br />

the area were recorded via <strong>to</strong>tal station. In a second test the reference area was scanned<br />

with a terrestrial <strong>3D</strong> laser scanner which produces accuracies of about 3 mm.<br />

The aSPECT <strong>3D</strong> model was computed from airborne pictures taken by a so-called ‘Oc<strong>to</strong>copter’<br />

drone and pictures taken from the ground. The comparison of the reference data from SFM<br />

point cloud and laser scan showed a very small standard deviation of about 1 cm. If only a<br />

<strong>to</strong>tal station is employed, the standard deviation still remains smaller than 3 cm. When a<br />

reference difference (e.g. a scale pho<strong>to</strong>graphed with the object) is the only device for scaling<br />

the point cloud, excavation findings of the respective size provide accuracies of around 5 cm.<br />

Fig. 4: Project “Enigma <strong>3D</strong>” (in cooperation with the “celtic + roman museum” Manching). Left: Angular “bread loaf idol” in<br />

pho<strong>to</strong>realistic and shaded depiction. Right: The highly precise structured light scanner PT-M 1280 serves as reference<br />

measuring device. For the au<strong>to</strong>mated, pho<strong>to</strong>graphic SFM-documentation, two Nikon cameras with macro-lenses and a turn<br />

table were set up in a homogenous white light room. The deviation analysis proved accuracies better than 0.2 mm. © ArcTron <strong>3D</strong><br />

GmbH, 2012.<br />

Example 2: Another accuracy test showed that the software is also suited for small objects. It<br />

was carried out in cooperation with the “celtic + roman museum” Manching 6 and focussed on<br />

6 M. Schaich, „Aenigma <strong>3D</strong>“. Zur Generierung hochaufgelöster <strong>3D</strong>-Modelle mit Fo<strong>to</strong>s einer handelsüblichen<br />

Spiegelreflexkamera – eine Fallstudie mit ArcTron’s Software aSPECT <strong>3D</strong> am Beispiel der „Brotlaibidole“. In W.<br />

David (Ed.), Aenigma. Katalog zur Ausstellung (2012) – in preparation.


<strong>From</strong> <strong>digital</strong> <strong>image</strong> <strong>series</strong> <strong>to</strong> <strong>3D</strong> <strong>models</strong><br />

more than a hundred of so-called “bread loaf idols” and other small archaeological objects.<br />

Professional camera systems with macro lenses produced accuracies of 0.2 mm for the<br />

resulting SFM-point cloud. A high-resolution structured light scan with a resolution of 0.05<br />

mm served as a reference.<br />

Processing pho<strong>to</strong>realistic <strong>3D</strong> <strong>models</strong><br />

Using one additional step, point clouds can be processed in<strong>to</strong> polygonally meshed <strong>3D</strong><br />

<strong>models</strong>. First of all, the coloured point clouds should be filtered and rid of interfering objects<br />

(e.g. scaffold elements). aSPECT <strong>3D</strong> offers different kinds of <strong>to</strong>ols for reducing noise in <strong>3D</strong><br />

data or deleting dispersive points. Afterwards, a closed <strong>3D</strong> mesh is laid over the point cloud<br />

using a so-called “Poisson triangulation” – a probability distribution. This <strong>3D</strong> mesh consists of<br />

numerous polygons, which connect at least three different points each. Simultaneously, the<br />

software generates a so-called “vertex texture”. For this, a colour gradient is computed from<br />

the respective RGB colours of each triangle. That is why the coloured surface of an object<br />

with a very dense point cloud is already realistically depicted.<br />

However, it takes another texturing for generating the pho<strong>to</strong>realistic <strong>3D</strong> model. During this<br />

step, the optimized pho<strong>to</strong>graphs of the SFM-process are mapped on<strong>to</strong> the geometries of the<br />

model surface. An Apsis-fresco may serve as an illustrating example (Fig. 5).<br />

Fig. 5: Project “Ninfa <strong>3D</strong>” (in cooperation with the German His<strong>to</strong>rical Institute, Rome). Left: SFM-documentation of a fresco.<br />

Middle: <strong>3D</strong> model with vertex texture. Right: Texture with high-resolution pho<strong>to</strong>graphs on the <strong>3D</strong> model. © ArcTron <strong>3D</strong> GmbH,<br />

2012.<br />

Combined airborne and terrestrial surveys<br />

During the past few years at ArcTron <strong>3D</strong> , a combination of airborne and terrestrial<br />

documentation technologies has proven very suitable for comprehensive <strong>3D</strong> surveys. In<br />

targeted survey flights anultra-light paraglider trike is deployed by the company.<br />

Up <strong>to</strong> 90 % of all required information can be gathered by aerial surveys. That is why since<br />

2010 ArcTron has also been using a remote controlled “camera drone” (Fig. 6). This aircraft<br />

robot can be completely au<strong>to</strong>nomously controlled via a ground station and is especially<br />

programmed for taking software-specific, pho<strong>to</strong>grammetrical survey <strong>image</strong>s. The system is<br />

able <strong>to</strong> carry cameras of over 2 kg and is equipped with a stabilized and separately<br />

controllable camera platform. This technology is suited for replacing bulky lifting platforms<br />

and can even reach remote objects very precisely (e.g. the gothic embellishment of dome<br />

spires).


<strong>From</strong> <strong>digital</strong> <strong>image</strong> <strong>series</strong> <strong>to</strong> <strong>3D</strong> <strong>models</strong><br />

Fig. 6: Project “Ninfa <strong>3D</strong>” (in cooperation with the German His<strong>to</strong>rical Institute, Rome). Left: Oc<strong>to</strong>copter, camera drone with a fully<br />

stabilized camera platform. Right: Technology in practice. Deployment of the system for the <strong>3D</strong> documentation of medieval<br />

remains of the city Ninfa (Italy). © ArcTron <strong>3D</strong> GmbH, 2012.<br />

Data management and exploitation<br />

The finished pho<strong>to</strong>realistic model can now be further edited and processed using the<br />

software’s inherent database.<br />

• Data management<br />

All object geometries can be individually ordered and described in the database. Different<br />

drafts of database structures for archaeological, architectural or res<strong>to</strong>ration purposes are<br />

available. They can be adjusted <strong>to</strong> any complex project. The database provides a<br />

hierarchical object system (“part of relation system”) as well as an additional classification<br />

tree structure. If necessary, a completely new global structure can be created.<br />

• Further processing<br />

The software allows <strong>to</strong> allocate additional information – e.g. for damage mapping – <strong>to</strong> the <strong>3D</strong><br />

model. Damage mapping can be applied directly on<strong>to</strong> pho<strong>to</strong>realistic <strong>models</strong> (Fig. 7). Mapping<br />

can also be carried out on location just beside the original using a notebook or a tablet PC<br />

with the respective <strong>3D</strong> software. This capability facilitates precise and reliable documentation<br />

in the field (essential <strong>to</strong> complex <strong>3D</strong> projects). Furthermore, the <strong>3D</strong> <strong>models</strong> can be linked with<br />

additional meta data such as texts, CAD-files, <strong>image</strong>s, films, PDFs etc. Using the included<br />

<strong>to</strong>ols complex documentation is a logical, step by step process.<br />

Fig. 7: Damage mapping of the Bronze Statue of general Yorck of Wartenberg (1759-1830), Berlin. The complex <strong>3D</strong> mapping of<br />

all damage areas with the integrated database was carried out by the res<strong>to</strong>ration office Helmich (Berlin) with the <strong>3D</strong> GIS system<br />

aSPECT <strong>3D</strong> . It enables database enquiries of deliberate <strong>3D</strong> damage mappings and res<strong>to</strong>ration planning. Graphics: © Landmark<br />

Preservation Authority Berlin & res<strong>to</strong>ration office Helmich, based on a laser scan by the company Laserscan, Berlin. Software<br />

Screenshots: ArcTron <strong>3D</strong> GmbH, 2009.<br />

• Print versions<br />

Printed documentations on paper, CAD-maps and high-resolution, scaled and rectified<br />

screenshots of many mosaic <strong>image</strong>s (orthopho<strong>to</strong>s) can be derived from different<br />

perspectives.


<strong>From</strong> <strong>digital</strong> <strong>image</strong> <strong>series</strong> <strong>to</strong> <strong>3D</strong> <strong>models</strong><br />

• Presentations<br />

The <strong>3D</strong> <strong>models</strong> are able <strong>to</strong> rotate on the screen for presentational purposes. Interior rooms<br />

can be entered and walked through in full screen mode – similar <strong>to</strong> video games. <strong>3D</strong> stereo<br />

technologies are also supported if a respective <strong>3D</strong> moni<strong>to</strong>r is available. This allows viewing<br />

the objects in “real <strong>3D</strong>” using <strong>3D</strong> glasses.<br />

Future prospects<br />

The company ArcTron <strong>3D</strong> has been developing the presented software continually for the last<br />

10 years. By now, it provides <strong>to</strong>ols and applications for res<strong>to</strong>rers, heritage preservers and<br />

archaeologists. aSPECT <strong>3D</strong> continues on a developmental path. Currently, ArcTron <strong>3D</strong> is<br />

working on a process manager, which will lead the user systematically through the complete<br />

workflow process, from point cloud <strong>to</strong> textured <strong>3D</strong> model. This brand new development will<br />

be presented on the trade fair “denkmal” in November.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!