Mid Sweden University

miun.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Modeling Depth Uncertainty of Desynchronized Multi-Camera Systems
Mid Sweden University, Faculty of Science, Technology and Media, Department of Information Systems and Technology. (Realistic 3D)ORCID iD: 0000-0002-4967-3033
Mid Sweden University, Faculty of Science, Technology and Media, Department of Information Systems and Technology. (Realistic 3D)ORCID iD: 0000-0003-3751-6089
Mid Sweden University, Faculty of Science, Technology and Media, Department of Information Systems and Technology. (Realistic 3D)
2017 (English)In: 2017 International Conference on 3D Immersion (IC3D), IEEE, 2017Conference paper, Published paper (Refereed)
Abstract [en]

Accurately recording motion from multiple perspectives is relevant for recording and processing immersive multi-media and virtual reality content. However, synchronization errors between multiple cameras limit the precision of scene depth reconstruction and rendering. In order to quantify this limit, a relation between camera de-synchronization, camera parameters, and scene element motion has to be identified. In this paper, a parametric ray model describing depth uncertainty is derived and adapted for the pinhole camera model. A two-camera scenario is simulated to investigate the model behavior and how camera synchronization delay, scene element speed, and camera positions affect the system's depth uncertainty. Results reveal a linear relation between synchronization error, element speed, and depth uncertainty. View convergence is shown to affect mean depth uncertainty up to a factor of 10. Results also show that depth uncertainty must be assessed on the full set of camera rays instead of a central subset.

Place, publisher, year, edition, pages
IEEE, 2017.
Keywords [en]
Camera synchronization, Synchronization error, Depth estimation error, Multi-camera system
National Category
Signal Processing Other Engineering and Technologies
Identifiers
URN: urn:nbn:se:miun:diva-31841DOI: 10.1109/IC3D.2017.8251891ISI: 000427148600001Scopus ID: 2-s2.0-85049401578ISBN: 978-1-5386-4655-7 (electronic)OAI: oai:DiVA.org:miun-31841DiVA, id: diva2:1149222
Conference
2017 International Conference on 3D Immersion (IC3D 2017), Brussels, Belgium, 11th-12th December 2017
Projects
LIFE project
Funder
Knowledge Foundation, 20140200Available from: 2017-10-13 Created: 2017-10-13 Last updated: 2025-02-18
In thesis
1. Multi-Camera Light Field Capture: Synchronization, Calibration, Depth Uncertainty, and System Design
Open this publication in new window or tab >>Multi-Camera Light Field Capture: Synchronization, Calibration, Depth Uncertainty, and System Design
2018 (English)Licentiate thesis, comprehensive summary (Other academic)
Abstract [en]

The digital camera is the technological counterpart to the human eye, enabling the observation and recording of events in the natural world. Since modern life increasingly depends on digital systems, cameras and especially multiple-camera systems are being widely used in applications that affect our society, ranging from multimedia production and surveillance to self-driving robot localization. The rising interest in multi-camera systems is mirrored by the rising activity in Light Field research, where multi-camera systems are used to capture Light Fields - the angular and spatial information about light rays within a 3D space. 

The purpose of this work is to gain a more comprehensive understanding of how cameras collaborate and produce consistent data as a multi-camera system, and to build a multi-camera Light Field evaluation system. This work addresses three problems related to the process of multi-camera capture: first, whether multi-camera calibration methods can reliably estimate the true camera parameters; second, what are the consequences of synchronization errors in a multi-camera system; and third, how to ensure data consistency in a multi-camera system that records data with synchronization errors. Furthermore, this work addresses the problem of designing a flexible multi-camera system that can serve as a Light Field capture testbed.

The first problem is solved by conducting a comparative assessment of widely available multi-camera calibration methods. A special dataset is recorded, giving known constraints on camera ground-truth parameters to use as reference for calibration estimates. The second problem is addressed by introducing a depth uncertainty model that links the pinhole camera model and synchronization error to the geometric error in the 3D projections of recorded data. The third problem is solved for the color-and-depth multi-camera scenario, by using a proposed estimation of the depth camera synchronization error and correction of the recorded depth maps via tensor-based interpolation. The problem of designing a Light Field capture testbed is addressed empirically, by constructing and presenting a multi-camera system based on off-the-shelf hardware and a modular software framework.

The calibration assessment reveals that target-based and certain target-less calibration methods are relatively similar at estimating the true camera parameters. The results imply that for general-purpose multi-camera systems, target-less calibration is an acceptable choice. For high-accuracy scenarios, even commonly used target-based calibration approaches are insufficiently accurate. The proposed depth uncertainty model is used to show that converged multi-camera arrays are less sensitive to synchronization errors. The mean depth uncertainty of a camera system correlates to the rendered result in depth-based reprojection, as long as the camera calibration matrices are accurate. The proposed depthmap synchronization method is used to produce a consistent, synchronized color-and-depth dataset for unsynchronized recordings without altering the depthmap properties. Therefore, the method serves as a compatibility layer between unsynchronized multi-camera systems and applications that require synchronized color-and-depth data. Finally, the presented multi-camera system demonstrates a flexible, de-centralized framework where data processing is possible in the camera, in the cloud, and on the data consumer's side. The multi-camera system is able to act as a Light Field capture testbed and as a component in Light Field communication systems, because of the general-purpose computing and network connectivity support for each sensor, small sensor size, flexible mounts, hardware and software synchronization, and a segmented software framework. 

Place, publisher, year, edition, pages
Sundsvall, Sweden: Mid Sweden University, 2018. p. 64
Series
Mid Sweden University licentiate thesis, ISSN 1652-8948 ; 139
Keywords
Light field, Camera systems, Multiview, Synchronization, Camera calibration
National Category
Computer and Information Sciences
Identifiers
urn:nbn:se:miun:diva-33622 (URN)978-91-88527-56-1 (ISBN)
Presentation
2018-06-15, L111, Holmgatan 10, Sundsvall, 13:00 (English)
Opponent
Supervisors
Funder
Knowledge Foundation, 20140200
Note

Vid tidpunkten för framläggning av avhandlingen var följande delarbete opublicerat: delarbete 3 manuskript.

At the time of the defence the following paper was unpublished: paper 3 manuscript.

Available from: 2018-05-16 Created: 2018-05-15 Last updated: 2018-05-16Bibliographically approved
2. Augmented Telepresence based on Multi-Camera Systems: Capture, Transmission, Rendering, and User Experience
Open this publication in new window or tab >>Augmented Telepresence based on Multi-Camera Systems: Capture, Transmission, Rendering, and User Experience
2021 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

 Observation and understanding of the world through digital sensors is an ever-increasing part of modern life. Systems of multiple sensors acting together have far-reaching applications in automation, entertainment, surveillance, remote machine control, and robotic self-navigation. Recent developments in digital camera, range sensor and immersive display technologies enable the combination of augmented reality and telepresence into Augmented Telepresence, which promises to enable more effective and immersive forms of interaction with remote environments.

The purpose of this work is to gain a more comprehensive understanding of how multi-sensor systems lead to Augmented Telepresence, and how Augmented Telepresence can be utilized for industry-related applications. On the one hand, the conducted research is focused on the technological aspects of multi-camera capture, rendering, and end-to-end systems that enable Augmented Telepresence. On the other hand, the research also considers the user experience aspects of Augmented Telepresence, to obtain a more comprehensive perspective on the application and design of Augmented Telepresence solutions.

This work addresses multi-sensor system design for Augmented Telepresence regarding four specific aspects ranging from sensor setup for effective capture to the rendering of outputs for Augmented Telepresence. More specifically, the following problems are investigated: 1) whether multi-camera calibration methods can reliably estimate the true camera parameters; 2) what the consequences are of synchronization errors in a multi-camera system; 3) how to design a scalable multi-camera system for low-latency, real-time applications; and 4) how to enable Augmented Telepresence from multi-sensor systems for mining, without prior data capture or conditioning. 

The first problem was solved by conducting a comparative assessment of widely available multi-camera calibration methods. A special dataset was recorded, enforcing known constraints on camera ground-truth parameters to use as a reference for calibration estimates. The second problem was addressed by introducing a depth uncertainty model that links the pinhole camera model and synchronization error to the geometric error in the 3D projections of recorded data. The third problem was addressed empirically - by constructing a multi-camera system based on off-the-shelf hardware and a modular software framework. The fourth problem was addressed by proposing a processing pipeline of an augmented remote operation system for augmented and novel view rendering.

The calibration assessment revealed that target-based and certain target-less calibration methods are relatively similar in their estimations of the true camera parameters, with one specific exception. For high-accuracy scenarios, even commonly used target-based calibration approaches are not sufficiently accurate with respect to the ground truth. The proposed depth uncertainty model was used to show that converged multi-camera arrays are less sensitive to synchronization errors. The mean depth uncertainty of a camera system correlates to the rendered result in depth-based reprojection as long as the camera calibration matrices are accurate. The presented multi-camera system demonstrates a flexible, de-centralized framework where data processing is possible in the camera, in the cloud, and on the data consumer's side. The multi-camera system is able to act as a capture testbed and as a component in end-to-end communication systems, because of the general-purpose computing and network connectivity support coupled with a segmented software framework. This system forms the foundation for the augmented remote operation system, which demonstrates the feasibility of real-time view generation by employing on-the-fly lidar de-noising and sparse depth upscaling for novel and augmented view synthesis.

In addition to the aforementioned technical investigations, this work also addresses the user experience impacts of Augmented Telepresence. The following two questions were investigated: 1) What is the impact of camera-based viewing position in Augmented Telepresence? 2) What is the impact of depth-aiding augmentations in Augmented Telepresence? Both are addressed through a quality of experience study with non-expert participants, using a custom Augmented Telepresence test system for a task-based experiment. The experiment design combines in-view augmentation, camera view selection, and stereoscopic augmented scene presentation via a head-mounted display to investigate both the independent factors and their joint interaction.

The results indicate that between the two factors, view position has a stronger influence on user experience. Task performance and quality of experience were significantly decreased by viewing positions that force users to rely on stereoscopic depth perception. However, position-assisting view augmentations can mitigate the negative effect of sub-optimal viewing positions; the extent of such mitigation is subject to the augmentation design and appearance.

In aggregate, the works presented in this dissertation cover a broad view of Augmented Telepresence. The individual solutions contribute general insights into Augmented Telepresence system design, complement gaps in the current discourse of specific areas, and provide tools for solving challenges found in enabling the capture, processing, and rendering in real-time-oriented end-to-end systems.

Place, publisher, year, edition, pages
Sundsvall: Mid Sweden University, 2021. p. 70
Series
Mid Sweden University doctoral thesis, ISSN 1652-893X ; 345
National Category
Computer and Information Sciences
Identifiers
urn:nbn:se:miun:diva-41860 (URN)978-91-89341-06-7 (ISBN)
Public defence
2021-05-17, C312, Mittuniversitetet Holmgatan 10, Sundsvall, 14:00 (English)
Opponent
Supervisors
Available from: 2021-04-15 Created: 2021-04-15 Last updated: 2021-11-22Bibliographically approved

Open Access in DiVA

fulltext(686 kB)1376 downloads
File information
File name FULLTEXT01.pdfFile size 686 kBChecksum SHA-512
f252afc8627d023471530cbc6794919c8fe7f52a5524ddeb48dea468a32cdd8c6516814f5bee35325aa73507aab9173ff5745261bd4d43a2a689755f7e698016
Type fulltextMimetype application/pdf

Other links

Publisher's full textScopus

Authority records

Dima, ElijsSjöström, MårtenOlsson, Roger

Search in DiVA

By author/editor
Dima, ElijsSjöström, MårtenOlsson, Roger
By organisation
Department of Information Systems and Technology
Signal ProcessingOther Engineering and Technologies

Search outside of DiVA

GoogleGoogle Scholar
Total: 1376 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

doi
isbn
urn-nbn

Altmetric score

doi
isbn
urn-nbn
Total: 901 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf