International Archives of the Photogrammetry, Remote Sensing
更新时间:2023-04-06 06:29:01 阅读量: 教育文库 文档下载
International Archives of the Photogrammetry, Remote Sensing & Spatial Information Sciences, 35(5):360-365, 2004.
ON AUTOMATIC ORTHOPROJECTION AND TEXTURE-MAPPING
OF 3D SURFACE MODELS
L. Grammatikopoulos a, I. Kalisperakis a, G. Karras a, T. Kokkinos a, E. Petsa b
a Laboratory of Photogrammetry, Department of Surveying,
National Technical University of Athens (NTUA), GR-15780 Athens, Greece
b Department of Surveying, The Technological Educational Institute of Athens (TEI-A),
Ag. Spyridonos Str., GR-12210 Athens, Greece
E-mail: lazaros@central.ntua.gr, ilias_k@central.ntua.gr, gkarras@central.ntua.gr, tk97010@survey.ntua.gr, petsa@teiath.gr KEY WORDS: Orthorectification, DEM/DTM, Laser scanning, Texture, Visualization, Automation, Heritage Conservation ABSTRACT
Photo-textured 3D surface models, and orthophotography in particular, are most important photogrammetric products, notably in he-ritage conservation. However, conventional software typically uses surface descriptions obtained via 2D triangulation; additionally, it cannot handle image visibility. Ignoring multiple elevations and image occlusions is clearly too restrictive for a complex surface shape. Geometric accuracy and visual quality are then possible only with tedious human interaction during surface modeling but also orthoprojection. Yet, laser scanning allows today fast collection of accurate, dense surface point clouds and creation of 3D meshes. Close-range photogrammetry is obviously expected to take full advantage of this.
The authors present their approach for an automated production of orthoimages from fully 3D surface representations derived from laser scanning. In a first step, the algorithm detects surface occlusions for the novel view. While common photogrammetric software needs operator-defined patches on inpidual original images as the source for image content, here all available images are combined for ‘viewer-independent’ texturing of the new image. To this end, bundle adjustment data allow all surface triangles to be back-pro-jected onto all initial images to establish visibilities. Texture blending is performed with suitable weighting, which controls the local radiometric contribution of each original image involved. Given more than two values, a statistical test allows to automatically ex-clude outlying colour data. The implemented algorithm was tested at the example of a Byzantine church in Athens to indicate that this coupling of laser scanning with photogrammetry is capable to automatically create novel views from several images, while com-bining geometric accuracy and visual quality with speed. Finally, future tasks and further elaborations are outlined.
1. INTRODUCTION
Among all photogrammetric products for the documentation of cultural heritage, digital orthomosaics – a combination of geo-metric accuracy with textured detail – are perhaps the most pro-minent. This, of course, is not intended to understate the signifi-cance of other related products, such as digital developments or cartographic projections, drapings or photorealistic visualisation and animation. In fact, orthophoto generation stands here as the paradigm for a core problem of photogrammetry, which incor-porates both surface modeling and photo-texturing. Compared to conventional aerial mapping, orthoimaging of cul-tural monuments often faces a number of significant problems. For instance, as discussed in Mavromati et al. (2002), these may include use of amateur cameras on unstable camera platforms; related problems concerning control over image configurations; resulting difficulties in bundle adjustment. However, a matter of primary importance is accurate surface modeling. It needs to be underlined here that a 3D model is not simply a prerequisite for orthoprojection or realistic rendering. Actually, in cases where only photo-realism or animated visualizations are required, then image-based rendering techniques may provide a direct solution (Beraldin et al., 2002). But photogrammetry typically relies on model-based texturing, as it is mostly asked to also produce ex-plicit 3D data and representations for the purposes of geometric or morphological documentation and analysis.
In many close-range applications object shapes may indeed be complex. As a rule, this implies significant occlusion problems. Thus, surface modeling is a key factor for producing orthophoto results, which will be geometrically reliable and visually correct (no ‘melting’ or ‘stretching’). Conventionally, all surface points are collected manually with stereoscopic viewing (the commer-cial matching algorithms usually require considerable editing in the case of an archaeological object). It has been demonstrated by Mavromati et al. (2003) that suitable collection strategies, as regards breaklines in particular, are capable of providing results of high quality. Notwithstanding its merits, however, this course is indeed tedious and time-consuming. Its limitations also in-clude registration problems among stereopair-based 3D models in the case of images all around the object. At the other far end of image-based modeling, powerful techniques are being deve-loped, notably in computer vision, for the automatic extraction of 3D surface models from an image sequence without any prior information about objects or camera. Although models of high visual quality can be thus produced, it appears that the obtained accuracies are not yet in position to meet the requirements for most mapping applications (Pollefeys et al., 2000). The metric potential of advanced techniques for an automatic dense recon-struction from small numbers of multiple wide-baseline images (Strecha et al., 2003) also remains to be further assessed. For certain objects classes, semi-automatic (hybrid) methods, based on a basic volumetric model of the scene which is subsequently exploited to constrain stereo matching, have also been presented (Debevec et al., 1996).
On the other hand, range-based modeling (notably through laser scanning) represents a powerful technology capable of sampling vast numbers of surface points at very fast rates. In this sense, it may well provide the required 3D support for orthorectification (Monti et al., 2002). In a wider sense, the same also holds true for creating photo-textured virtual models of real-world scenes, chiefly in computer graphics applications, where visual quality is a major concern (Bernardini et al., 2001; Corrêa et al., 2002). In fact, high-resolution recording of cultural sites and possibili-ties to promote them through virtual 3D visits, for instance, sti-mulates research, notably regarding fusion of laser scanning and colour imagery (Beraldin et al., 2002). Certain commercial 3D systems provide model-registered colour texture but, neverthe-
International Archives of the Photogrammetry, Remote Sensing & Spatial Information Sciences, 35(5):360-365, 2004.
less, the limited image quality usually makes the acquisition of
separate images with higher resolution necessary. Clearly, this
is equally true for the high requirements of orthoimaging. However, the laser scanning approach is faced with several pro-
blems. Leaving high cost aside, a major question is that of post-
processing vast volumes of data for mesh triangulation (includ-
ing noise removal and hole-filling), which is a very demanding
procedure indeed (B?hler et al., 2003). Yet, the next problem is
that commercial orthoprojection software handles only surfaces
described as a DTM with a unique elevation at each planimetric XY location. Thus, all scanned 3D points are typically processed by 2D triangulation into a 2.5D mesh – with predictable conse-
quences on the final orthoimage (Mavromati et al., 2003, give
such an example). Unless one finds suitable projections yielding
single-valued ‘depth’ functions for particular surface types (as
done by Knyaz & Zheltov, 2000), orthoimaging algorithms for
fully 3D models must necessarily be introduced.
The main task of such algorithms is to avoid the common flaws
of orthoimaging (displacement, blind areas, double-projection)
by handling the problem of visibility, which is twofold. On the
one hand, every inpidual surface unit (groundel) which is visi-
ble in the direction of the orthoimage should be established. In
this way, to each orthoimage pixel a unique elevation is assign-
ed. Next is to check whether these surface points are in fact vi-
sible from the perspective centre of the original image, too. In
case of occlusion, a colour value can be extracted from an adja-
cent image. Such approaches, based on dense regular DTM s de-
rived from laser scanning, have been implemented in aerial and
close-range projects (Kim et al., 2000; Boccardo et al., 2001).
Following the usual photogrammetric practice, in the above and other cases the texture for each visible surface unit is extracted from a corresponding single original image. Provided that over-lapping images exist, the source image can be selected accord-ing to different possible criteria, for instance: imaging distance, angle formed by the projective ray and the surface; size of the imaged surface triangle. Evidently, this ‘single-image’ texturing approach can lead to adjacent surface triangles receiving colour from different images with varying radiometric characteristics. The consequences on triangle borders can be radiometric distor-tion and discontinuity artifacts (El-Hakim et al., 2003). Alterna-tive elaborate responses to this, in the area of computer graphics and computer vision, rely on colour interpolation, or ‘blending’. For every point, appropriately weighted combinations of corres-ponding triangle textures from all available images – or from a suitably selected image subset – on which this point appears are used (Neugebauer & Klein, 1999; Bernardini et al., 2001; Bueh-ler et al., 2001; Wang et al., 2001; Rocchini et al., 2002). In this way of smoothing radiometric difference, seamless texture with no jumps in colour appearance can be obtained – at the possible cost of a certain blurring effect (El-Hakim et al., 2003).
The approaches referred to above have been developed in the field of computer graphics, where life-like animations, realism or illumination are evidently important. A weighting strategy is thus formulated mostly in the context of view-dependent texture mapping, where interpolation schemes favour images observing the object or scene closest in angle to the current viewing direc-tion. In this way, surface specularities and incorrect model geo-metry may be better captured (Debevec et al., 1996, 1998). However, it has been pointed out that using a single texture map in 3D models is usually sufficient (Wang et al., 2001). In this sense – though one obviously has much to benefit from research in this field – it appears that static rather than dynamic texturing is preferable for most instances of photogrammetric mapping. A view-independent algorithm weights the contribution of partici-pating original images according to their spatial relation to the model – e. g. distance, angle of view – and their characteristics – camera constant and resolution – in order to assign a unique colour value to each surface unit (see, for instance, Poulin et al., 1998; Grün et al., 2001).
Although colour blending may be regarded, to some extent, also as an ‘error filtering’ process, existing error sources may cause geometric and radiometric distortions. Obviously, the final pro-duct is significantly affected by the accuracy, with which image orientations – relative to each other as well as in object space – have been recovered. This holds also true for camera calibration parameters. Consequently, a self-calibrating bundle adjustment, including lens distortion, is indispensable. Further obvious error sources causing misalignments include accuracy of 3D record-ing, quality of surface description by 3D faces and model regi-stration. Finally, though such problems are rather uncommon in photogrammetric applications, significant differences in resolu-tion of the source images, which can blur texture, are also to be considered (Neugebauer & Klein, 1999; Buehler et al., 2001).
Here, an approach is presented for the automated generation of orthoimages from a 3D mesh, derived from laser scanning. The implemented algorithm identifies all surface triangles which are seen in the viewing direction and then establishes whether these appear or not on every available image. Each orthoimage pixel is coloured through weighted blending of texture from all view-ing images, whereby outlying colour data are automatically ex-cluded. Results of experimental applications are also given.
2. PROJECTION AND TEXTURING ALGORITHM
For this procedure, the following input data are required:
? a triangulated 3D mesh in the form of successive XYZ triplets describing the object surface;
?grayscale or colour images along with their interior and exte-rior orientation parameters;
?the equation in space of the projection plane;
?the endpoints in object space, if necessary, of the area to be projected;
?the pixel size of the new digital image.
It is seen that, besides orthogonal, oblique projections may also be accommodated.
2.1 Model visibility and occlusion
In the first step, the triangulated 3D mesh is projected orthogo-nally onto the specified plane of projection. In order to speed up the search process, the area of the orthoimage is tessellated into a rectangular grid, whose cell is larger than the one of the ortho-image, e.g. by 5 times (its size depends on factors such as the available computer memory, the model size and that of the new image). For each 2D triangle, the circumscribing orthogonal pa-rallelogram is formed. This occupies a number of adjacent grid cells, to which the identity number (ID) of the particular triangle is assigned.
This procedure is repeated for all triangles, resulting into a table containing all triangle ID s ascribed to each inpidual grid cell. In this way, all projected triangles actually containing a particu-lar pixel of the orthoimage may be established by checking only a limited number of triangles (namely, those ascribed to the cor-responding grid cell). Among these model triangles intersected in space by the projecting line of a particular orthoimage pixel, the one whose intersection yields the largest elevation value is selected; the elevation value, which provides the Z-value of the orthoimage pixel, and the triangle ID number are stored. In this mode, the model visibility/occlusion question has been handled.
International Archives of the Photogrammetry, Remote Sensing & Spatial Information Sciences, 35(5):360-365, 2004.
2.2 Image visibility and occlusion
Coming next to the image visibility aspect, all 3D triangles are now centrally projected, via the corresponding orientation data, onto all images involved. For every planimetric XY value of the orthoimage and its attributed Z-value, the corresponding image xy coordinates on all images are calculated. A similar technique as before is followed: among the model triangles intersected by a particular image ray, the one closer to the projective center is the triangle which has been actually recorded on the image. If the ID number of this triangle is not identical with that already assigned in the previous phase to the orthoimage pixel, it is in fact established that the model point corresponding to this parti-cular ortho pixel is occluded on the examined image. If, on the contrary, the two triangle ID s coincide, then the model point is visible on the particular image, and the RGB values are stored. Despite the computational burden, colour values are interpola-ted in the present implementation by bicubic convolution, since it provides an obviously smoother result. However, it is evident that adjacent pixels do not necessarily relate to adjacent model points. Although no discernible effects emerged in the applica-tions, checks may possibly be considered to omit such pixels. 2.3 Texture interpolation
Result of the preceding step for all orthoimage pixels is colour values from several images – unless, of course, the correspond-ing model point is occluded on all images. In this latter case, a specific colour value marks the particular orthoimage pixels as undefined. For such regions, ‘hole-filling’ processes can extract colour values from the surrounding model areas (Debevec et al., 1998; Poulin et al., 1998), which has not been examined here. If asked, however, the algorithm can create a map which displays all orthoimage areas visible on 0, 1, 2 and > 2 source images. In this way, additional images, if available, could be introduced in the process to fill the gaps. It is remarked that it is also useful to know which orthoimage areas are visible in more that 2 images, as this allows a test to detect and exclude outliers.
Indeed, in order to assign a final colour value to the orthoimage pixels, outlying values must first be excluded. Generally, these could originate not only from model faults, but also from view-dependent features – such as specular highlights, transparencies, mirrors, refractions, obstacles etc. (Poulin et al., 1998; Rocchini et al., 2001). However, more significant for photogrammetry is probably the case when one approaches model parts not seen by a camera, i. e. borders of occlusion (Neugebauer & Klein, 1999; Buehler et al., 2001). In these instances artifacts might appear, since even very small orientation or registration – or modeling – errors can lead to colour mistakenly derived from an occluding or, respectively, an occluded model point (Fig. 1 shows such an example; see also Fig. 5 but also Fig. 4). One may possibly eva-luate the ‘occlusion risk’ of pixels – for instance, by a compari-son of the imaging distance with those of adjacent pixels from their own visible surface point. This is a topic of future study. Here, a basic statistical test was adopted, provided that a suffi-cient number (> 2) of colour values are available for a particular orthoimage pixel. Mean (μ) and standard deviation (σ) of colour values are computed each time; inpidual colour values falling outside the range μ ± β×σ are excluded. It is estimated that the value of factor β could be around 1 (indeed, in the test presented in the next section using 7 images, it was set β = 1). After this procedure, the valid contributing colour values from all images are used to generate the final texture of each orhtoimage pixel.
A weighted mean of all contributing images is finally used for texturing each particular orthoimage pixel. In view-independent texture mapping, the main factors influencing colour quality are scale (i. e. imaging distance and camera constant) of the source image; its viewing angle (i. e. the angle formed by the intersec-tion of the image ray and the model triangle); and image resolu-tion. In fact, these factors are all combined to yield the size (in pixels) of the 2D triangle on each image, which is regarded as a good indication of the quality of the extracted colour. Hence, as suggested by Poulin et al. (1998), the contribution of all partici-pating colour values are weighted here as relative functions of the corresponding 2D triangle areas (this weighting scheme has also been used by Grün et al., 2001).
Figure 1. Due to small modeling, calibration and orientation errors, the texture of point B on image 1 may be assigned to A. The algorithm was initially developed in MatLab and was final-ly implemented in C. In order to verify its performance and also increase speed, tests were carried out with synthetic data, using images extracted from an existing photo-textured 3D model of a building (Kokkinos, 2004).
3. APPLICATION OF THE ALGORITHM
The object of the experimental application was the entrance of the 11th century church of ‘Kapnikarea’, an important Byzantine monument in the centre of Athens. Mainly due to its columns, the object is sufficiently complex for the task.
3.1 Scanning and modeling
For surface recording, the Mensi GS200 laser scanner was used. The device scans at a rate of 5000 points/sec, having a 60° ver-tical field of view. Three separate scans were carried out from a distance of about 5 m, for which a typical value ±1.4 mm of the standard deviation is given (the resolution is 4 mm at a distance of 10 m). For registration, 6 well distributed target spheres were employed, also measured geodetically. The RealWorks Survey 4.1.2 software was used for a target-based registration of scans. The precision of geo-referencing was about ±2.5 mm. In total, 7 million surface points were obtained. Subsequently, these were spatially down-sampled to provide a final 3D mesh, which con-sisted of about 3 million triangles. Special features of the soft-ware (smoothing and peak removal) were used to improve the 3D model. A grayscale intensity map was also obtained (in Fig.
2 the intensity map of the central scan is seen).
3.2 Bundle adjustment
The object has been recorded employing different analogue and digital cameras, which will be used in future tests. Here, results are given for images (2592×1944) from a Sony 5 MegaPixel ca-mera. A total of 7 overlapping images were selected, taken with fixed focusing, to keep interior orientation invariant. All images used are seen in Fig. 3.
International Archives of the Photogrammetry, Remote Sensing & Spatial Information Sciences, 35(5):360-365, 2004.
Figure 2. Intensity image of part of the central scan.
A set of 18 signalised points served as geodetic ground control, most of which appeared on all images. Using our own software, a bundle adjustment with self-calibration (including coefficients k 1 and k 2 of radial symmetric lens distortion) was carried out in-volving 8 tie points. It was possible to achieve a high accuracy, as seen in Table 1.
Table 1. Bundle Adjustment (7 images; 18 GCP s; 8 tie points)
σo = ±0.28 pixel
c (pixel) x o (pixel) y o (pixel)
k 1 (×108 )
k 2 (
×1015 )
2573.94 ± 1.36 10.27 ± 1.03
7.35 ± 0.87 ?2.98 ± 0.04
4.90 ± 0.19 rms (GCP ) = 0.3 mm σ (TIE
) = 1.3 mm
Using the data for camera calibration and exterior orientations, the images could be used to colour the projections of the model.
3.3 Model texturing
All images used here had been acquired successively, i .e. under the same lighting conditions. Thus, as no radiometric variations
among images were observed, texture averaging was considered as adequate (for pre-processing see e. g. Grün et al., 2001). The plane of the church fa?ade was selected as projection plane. The orthoimage pixel size was fixed to 2 mm. As already seen, the algorithm identifies image visibility of all model parts previous-ly established as being visible in the direction of projection. The results for an inpidual image are shown in Fig. 4.
A next question to be noted, also noticed in Fig. 4, concerns the
matter of outlying colour values, particularly close to an occlu-sion border (cf. section 2.3 and Fig. 1), a problem illustrated in Fig. 5. Above, an extract is seen of the orthoimage derived from all source images without blunder-filtering. The white artifact – clearly originating from the occlusion border of the image in the middle – disappears if colour values deviating more than ±σ from the mean are automatically ignored (below).
The final, automatically generated, orthoprojection is presented in Fig. 6, left. The result is essentially satisfactory. A main im-perfection are small holes due to lack of texture (the camera has not been elevated). Most other defects – for example, a certain aliasing at edges of some parts which, however, may not be per-
fectly perpendicular to the projection plane – are observed only with considerable zooming. On the right, a section is presented utilising all available model and texture information, including parts which are not visible on the left (e. g. a part of the arches).
Figure 3. The seven images used in the project.
Regarding speed, a PC (CPU Athlon XP 2.4 GH z, 512 MB RAM ) required 5 min to project 3 million triangles onto the images, in a suitable order to facilitate the last step; 1 min to establish the Z-values for each orthoimage pixel; and 13 minutes for the pro-duction of a final image 3720×2775 from 7 images 2592×1944.
5. CONCLUDING REMARKS
An algorithm has been implemented for the automatic synthesis of textured views, given a 3D triangulated mesh and precise ca-libration/orientation information for overlapping images. Model and image visibility are identified, to allow pixel colouring with the weighted average from all viewing images, whereby a basic blunder detection tool allows to avoid artifacts. The product for the paticular object studied here is indeed very satisfactory.
Yet, further enhancements of the technique are both necessary and feasible. Thus, some hole-filling tools should be introduced. Besides, more robust means for outlier avoidance need to be ex-perimented with. For instance, closeness of a source pixel to an occlusion border can be taken into account, probably combined
International Archives of the Photogrammetry, Remote Sensing & Spatial Information Sciences, 35(5):360-365, 2004.
with the viewing angle. Further, image matching has been used for model and registration refinement (Debevec et al., 1996; see also Bernardini et al., 2001); indeed, a multiple image coverage and precise starting values supplied by the 3D model may allow to refine the model and/or the textured output. The above points indicate possible future tasks of the work presented here.
Figure 4. Orthoprojection (below) of the image above. At the right silhouettes of the columns texturing problems are seen. Acknowledgements
The authors wish to sincerely thank Geotech Ltd., representative of Trimble in Greece, and particularly Drs. V. Balis and C. Liapakis, for providing the scans and the ground control.
Figure 5. Part of the orthoimage from 7 images without (above) and with (below) outlier filtering. In the middle is the source image which causes the artifact with its occlusion border.
REFERENCES
Beraldin J.-A., Picard. M., El-Hakim, S. F., Godin, G., Latou-che, C., Valzano, V., Bandiera, A., 2002. Exploring a Byzantine crypt through a high-resolution texture mapped 3D model: com-bining range data and photogrammetry. Proc. CIPA Int. Work-shop on Scanning for Cultutal Heritage Recoding, pp. 65-70. Bernardini, F., Martin, I. M., Rushmeier, H., 2001. High-quality texture reconstruction from multiple scans. IEEE Trans. Visuali-zation & Computer Graphics, 7(4):318-332.
Boccardo, P., Dequal, S., Lingua, A., Rinaudo, F., 2001. True digital orthophoto for architectural and archaeological applica-tions. Proc. Int. Workshop on Recreating the Past: Visualization & Animation of Cultural Heritage, Ayuttaya, Thailand (in CD). B?hler, W., Bordas Vicent, M., Hanke. K., Marbs, A., 2003. Documentation of German Emperor Maximilian I’s tomb. Proc. XIX CIPA Int. Symposium, Antalya, Turkey, pp. 474-479.
International Archives of the Photogrammetry, Remote Sensing & Spatial Information Sciences, 35(5):360-365, 2004.
Figure 6. Left: Extract of the orthoimage from 7 images. Right: orthoprojection of a model section defined by a maximum Z-value.
Buehler, C., Bosse, M., McMillan, L., Gortler, S., Cohen, M., 2001. Unstructured lumigraph rendering. Proc. ACM SIGGRAPH Annual Conference Series, pp. 425-432.
Corrêa, W., Oliveira, M., Silva, C., Wang, J., 2002. Modeling and rendering of real environments. RITA, IX(1), pp. 1-32. Debevec, P., Taylor, C.J., Malik, G., 1996. Modeling and ren-dering architecture from photographs: a hybrid geometry- and image-based approach. ACM SIGGRAPH, pp. 11–20.
Debevec P., Borshukov G., Yu Y., 1998. Efficient view-depen-dent image-based rendering with projective texture-mapping. Proc. 9th Eurographics Rendering Workshop, Rendering Tech-niques ’98, Springer, pp. 14-26.
El-Hakim, S., Gonzo, L., Picard, M., Girardi, S., Simoni, A., 2003. Visualization of frescoed surfaces: Buonconsiglio Castle – Aquila Tower, ‘Cycle of the Months’. Int. Workshop on Visu-alization & Animation of Reality-Based 3D Models (in CD).
Grün, A., Zhang, L., Visnovcova, J., 2001. Automatic recon-struction and visualization of a complex Buddha Tower of Ba-yon, Angkor, Cambodia. Proc. 21.Wissenschaftlich-Technische Jahrestagung der DGPF, pp. 289-301.
Kim, Y.-I., Kwon, O.-H., Kim, H.-T., 2000. Detecting and re-storing the occlusion area for generating a digital orthoimage. Proc. ASPRS Annual Conference (in CD).
Knyaz, V. A., Zheltov, S. Y, 2000. Approach to accurate photo-realistic model generation for complex 3D objects. Int. Arch. Phot. & Rem. Sens., 33(5), pp. 428-433.
Kokkinos, T., 2004. Automatic Photo-Texturing of Projections of 3D Models. Diploma Thesis, Dept. of Surveying, NTUA. Mavromati, D,. Petsa, E., Karras, G., 2002. Theoretical and pra-ctical aspects of archaeological orthoimaging. Int. Arch. Phot. & Rem. Sens., 34(B5), pp. 413-418.
Mavromati, D., Petsa, E., Karras, G., 2003. Experiences in pho-togrammetric archaeological recording. Proc. XIX CIPA Int. Symposium, Antalya, Turkey, pp. 666-669.
Monti C., Brumana R., Fregonese L., Savi C., Achille C., 2002. 3D models generation from laser canner data to support digital orthophoto. Int. Arch. Phot. & Rem. Sens., 34(5), pp. 423-428.
Neugebauer, P., Klein, K., 1999. Texturing 3D models of real world objects from multiple unregistered photographics views, Proc. Eurographics ’99, Computer Graphics Forum, 18(3). Pollefeys, M., Koch, R., Vergauwen, M., van Gool, L., 2000. Automated reconstruction of 3D scenes from sequences of ima-ges. ISPRS J. Phot. & Rem. Sens., 55:251–267
Poulin, P., Ouimet, M., Frasson, M.-C., 1998. Interactively mo-deling with photogrammetry. Proc. Eurographics Workshop on Rendering ’98, pp. 93-104.
Rocchini, C., Cignoni, P., Montani, C., Scopigno, R., 2002. Ac-quiring, stitching and blending diffuse appearance attributes on 3D models. The Visual Computer, 18:186–204.
Strecha, C., Tuytelaars, T., van Gool, L., 2003. Dense matching of multiple wide-baseline views. Proc. 9th Int. Conf. on Compu-ter Vision, ICCV ’03, Vol. 2, pp. 1994-1201.
Wang, L., Kang, S.B., Szeliski, R., Shum, H.-Y., 2001. Optimal texture map reconstruction from multiple views, Proc. Com-puter Vision & Pattern Recognition, CVPR '01, vol. 1.
正在阅读:
International Archives of the Photogrammetry, Remote Sensing04-06
《认识平均分一》教学反思08-06
现代教育原理形考第5次作业参考答案09-27
温人社发23号关于调整基本医疗保险服务项目有关政策的通知10-23
日出作文450字07-13
江苏省住宅设计标准 - 图文07-04
社区道德模范事迹材料03-08
扁鹊见蔡桓公02-03
- 1International Journal of Hospitality Management
- 2On Communication Skills of Language in International Busines
- 3International Flying Start Programme Application Form
- 4International Marketing(Chap001)
- 5压缩感知(Compressive Sensing)总结,毕设小节
- 6DMD Implementation of a Single Pixel Camera Based on Compressed Sensing
- 7英语论文-International Business Negotiation Strategies and S
- 8Test bank International Finance MCQ(word)Chap 13
- 9Test bank International Finance MCQ(word)Chap 8
- 10Test bank International Finance MCQ(word)Chap 4
- exercise2
- 铅锌矿详查地质设计 - 图文
- 厨余垃圾、餐厨垃圾堆肥系统设计方案
- 陈明珠开题报告
- 化工原理精选例题
- 政府形象宣传册营销案例
- 小学一至三年级语文阅读专项练习题
- 2014.民诉 期末考试 复习题
- 巅峰智业 - 做好顶层设计对建设城市的重要意义
- (三起)冀教版三年级英语上册Unit4 Lesson24练习题及答案
- 2017年实心轮胎现状及发展趋势分析(目录)
- 基于GIS的农用地定级技术研究定稿
- 2017-2022年中国医疗保健市场调查与市场前景预测报告(目录) - 图文
- 作业
- OFDM技术仿真(MATLAB代码) - 图文
- Android工程师笔试题及答案
- 生命密码联合密码
- 空间地上权若干法律问题探究
- 江苏学业水平测试《机械基础》模拟试题
- 选课走班实施方案
- Photogrammetry
- International
- Archives
- Sensing
- Remote
- 人教版七年级上册数学《角的相关知识》能力提升专题过关分类练习
- 2022年考研数学二真题与全面解析(Word版)
- 新课程高中数学测试题组(必修2)5含答案
- 自考《电子商务》电子商务英语2011年10月真题试题
- 2012年德州市武城县事业单位公开招聘工作人员简章
- 上海交通大学工程实践与科技创新【1】实验报告
- 《珍珠鸟》读书笔记
- 2015年版镇江市电子加工产业园区规划及招商策略咨询研究报告
- 机械制造基础(模拟试题及答案)73267
- 二级综合医院评审标准(2015年版)实施细则2
- how to be a good translator怎么样成为一名好的翻译
- 初中中考专题济南市 初中学业水平考试地理试题(有答案)
- 精选5篇广西王城导游词.doc
- 三甘醇脱水系统设计毕业论文
- 中国互联网络发展状况分析报告
- 北京体育大学运动训练学考研复习重点名词解释部分
- 行政职业能力测验(行测)题库:行测:数字推理题725道详解
- iOS移动平台架构设计说明
- 【最新文档】大学生道德素质调查报告(精选多篇)-精选word文档 (1
- 公司股权分配协议样本