Mid Sweden University

miun.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
LIFE: A Flexible Testbed For Light Field Evaluation
Mid Sweden University, Faculty of Science, Technology and Media, Department of Information Systems and Technology. (Realistic 3D)ORCID iD: 0000-0002-4967-3033
Mid Sweden University, Faculty of Science, Technology and Media, Department of Information Systems and Technology. (Realistic 3D)
Mid Sweden University, Faculty of Science, Technology and Media, Department of Information Systems and Technology. (Realistic 3D)
Mid Sweden University, Faculty of Science, Technology and Media, Department of Information Systems and Technology. (Realistic 3D)
Show others and affiliations
2018 (English)Conference paper, Published paper (Refereed)
Abstract [en]

Recording and imaging the 3D world has led to the use of light fields. Capturing, distributing and presenting light field data is challenging, and requires an evaluation platform. We define a framework for real-time processing, and present the design and implementation of a light field evaluation system. In order to serve as a testbed, the system is designed to be flexible, scalable, and able to model various end-to-end light field systems. This flexibility is achieved by encapsulating processes and devices in discrete framework systems. The modular capture system supports multiple camera types, general-purpose data processing, and streaming to network interfaces. The cloud system allows for parallel transcoding and distribution of streams. The presentation system encapsulates rendering and display specifics. The real-time ability was tested in a latency measurement; the capture and presentation systems process and stream frames within a 40 ms limit.

Place, publisher, year, edition, pages
2018. article id 8478550
Keywords [en]
Multiview, 3DTV, Light field, Distributed surveillance, 360 video
National Category
Computer and Information Sciences
Identifiers
URN: urn:nbn:se:miun:diva-33620ISI: 000454903900016Scopus ID: 2-s2.0-85056147245ISBN: 978-1-5386-6125-3 (print)OAI: oai:DiVA.org:miun-33620DiVA, id: diva2:1205690
Conference
2018 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON), Stockholm – Helsinki – Stockholm, 3-5 June 2018
Projects
LIFE Project
Funder
Knowledge Foundation, 20140200Available from: 2018-05-15 Created: 2018-05-15 Last updated: 2021-04-15Bibliographically approved
In thesis
1. Multi-Camera Light Field Capture: Synchronization, Calibration, Depth Uncertainty, and System Design
Open this publication in new window or tab >>Multi-Camera Light Field Capture: Synchronization, Calibration, Depth Uncertainty, and System Design
2018 (English)Licentiate thesis, comprehensive summary (Other academic)
Abstract [en]

The digital camera is the technological counterpart to the human eye, enabling the observation and recording of events in the natural world. Since modern life increasingly depends on digital systems, cameras and especially multiple-camera systems are being widely used in applications that affect our society, ranging from multimedia production and surveillance to self-driving robot localization. The rising interest in multi-camera systems is mirrored by the rising activity in Light Field research, where multi-camera systems are used to capture Light Fields - the angular and spatial information about light rays within a 3D space. 

The purpose of this work is to gain a more comprehensive understanding of how cameras collaborate and produce consistent data as a multi-camera system, and to build a multi-camera Light Field evaluation system. This work addresses three problems related to the process of multi-camera capture: first, whether multi-camera calibration methods can reliably estimate the true camera parameters; second, what are the consequences of synchronization errors in a multi-camera system; and third, how to ensure data consistency in a multi-camera system that records data with synchronization errors. Furthermore, this work addresses the problem of designing a flexible multi-camera system that can serve as a Light Field capture testbed.

The first problem is solved by conducting a comparative assessment of widely available multi-camera calibration methods. A special dataset is recorded, giving known constraints on camera ground-truth parameters to use as reference for calibration estimates. The second problem is addressed by introducing a depth uncertainty model that links the pinhole camera model and synchronization error to the geometric error in the 3D projections of recorded data. The third problem is solved for the color-and-depth multi-camera scenario, by using a proposed estimation of the depth camera synchronization error and correction of the recorded depth maps via tensor-based interpolation. The problem of designing a Light Field capture testbed is addressed empirically, by constructing and presenting a multi-camera system based on off-the-shelf hardware and a modular software framework.

The calibration assessment reveals that target-based and certain target-less calibration methods are relatively similar at estimating the true camera parameters. The results imply that for general-purpose multi-camera systems, target-less calibration is an acceptable choice. For high-accuracy scenarios, even commonly used target-based calibration approaches are insufficiently accurate. The proposed depth uncertainty model is used to show that converged multi-camera arrays are less sensitive to synchronization errors. The mean depth uncertainty of a camera system correlates to the rendered result in depth-based reprojection, as long as the camera calibration matrices are accurate. The proposed depthmap synchronization method is used to produce a consistent, synchronized color-and-depth dataset for unsynchronized recordings without altering the depthmap properties. Therefore, the method serves as a compatibility layer between unsynchronized multi-camera systems and applications that require synchronized color-and-depth data. Finally, the presented multi-camera system demonstrates a flexible, de-centralized framework where data processing is possible in the camera, in the cloud, and on the data consumer's side. The multi-camera system is able to act as a Light Field capture testbed and as a component in Light Field communication systems, because of the general-purpose computing and network connectivity support for each sensor, small sensor size, flexible mounts, hardware and software synchronization, and a segmented software framework. 

Place, publisher, year, edition, pages
Sundsvall, Sweden: Mid Sweden University, 2018. p. 64
Series
Mid Sweden University licentiate thesis, ISSN 1652-8948 ; 139
Keywords
Light field, Camera systems, Multiview, Synchronization, Camera calibration
National Category
Computer and Information Sciences
Identifiers
urn:nbn:se:miun:diva-33622 (URN)978-91-88527-56-1 (ISBN)
Presentation
2018-06-15, L111, Holmgatan 10, Sundsvall, 13:00 (English)
Opponent
Supervisors
Funder
Knowledge Foundation, 20140200
Note

Vid tidpunkten för framläggning av avhandlingen var följande delarbete opublicerat: delarbete 3 manuskript.

At the time of the defence the following paper was unpublished: paper 3 manuscript.

Available from: 2018-05-16 Created: 2018-05-15 Last updated: 2018-05-16Bibliographically approved
2. Augmented Telepresence based on Multi-Camera Systems: Capture, Transmission, Rendering, and User Experience
Open this publication in new window or tab >>Augmented Telepresence based on Multi-Camera Systems: Capture, Transmission, Rendering, and User Experience
2021 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

 Observation and understanding of the world through digital sensors is an ever-increasing part of modern life. Systems of multiple sensors acting together have far-reaching applications in automation, entertainment, surveillance, remote machine control, and robotic self-navigation. Recent developments in digital camera, range sensor and immersive display technologies enable the combination of augmented reality and telepresence into Augmented Telepresence, which promises to enable more effective and immersive forms of interaction with remote environments.

The purpose of this work is to gain a more comprehensive understanding of how multi-sensor systems lead to Augmented Telepresence, and how Augmented Telepresence can be utilized for industry-related applications. On the one hand, the conducted research is focused on the technological aspects of multi-camera capture, rendering, and end-to-end systems that enable Augmented Telepresence. On the other hand, the research also considers the user experience aspects of Augmented Telepresence, to obtain a more comprehensive perspective on the application and design of Augmented Telepresence solutions.

This work addresses multi-sensor system design for Augmented Telepresence regarding four specific aspects ranging from sensor setup for effective capture to the rendering of outputs for Augmented Telepresence. More specifically, the following problems are investigated: 1) whether multi-camera calibration methods can reliably estimate the true camera parameters; 2) what the consequences are of synchronization errors in a multi-camera system; 3) how to design a scalable multi-camera system for low-latency, real-time applications; and 4) how to enable Augmented Telepresence from multi-sensor systems for mining, without prior data capture or conditioning. 

The first problem was solved by conducting a comparative assessment of widely available multi-camera calibration methods. A special dataset was recorded, enforcing known constraints on camera ground-truth parameters to use as a reference for calibration estimates. The second problem was addressed by introducing a depth uncertainty model that links the pinhole camera model and synchronization error to the geometric error in the 3D projections of recorded data. The third problem was addressed empirically - by constructing a multi-camera system based on off-the-shelf hardware and a modular software framework. The fourth problem was addressed by proposing a processing pipeline of an augmented remote operation system for augmented and novel view rendering.

The calibration assessment revealed that target-based and certain target-less calibration methods are relatively similar in their estimations of the true camera parameters, with one specific exception. For high-accuracy scenarios, even commonly used target-based calibration approaches are not sufficiently accurate with respect to the ground truth. The proposed depth uncertainty model was used to show that converged multi-camera arrays are less sensitive to synchronization errors. The mean depth uncertainty of a camera system correlates to the rendered result in depth-based reprojection as long as the camera calibration matrices are accurate. The presented multi-camera system demonstrates a flexible, de-centralized framework where data processing is possible in the camera, in the cloud, and on the data consumer's side. The multi-camera system is able to act as a capture testbed and as a component in end-to-end communication systems, because of the general-purpose computing and network connectivity support coupled with a segmented software framework. This system forms the foundation for the augmented remote operation system, which demonstrates the feasibility of real-time view generation by employing on-the-fly lidar de-noising and sparse depth upscaling for novel and augmented view synthesis.

In addition to the aforementioned technical investigations, this work also addresses the user experience impacts of Augmented Telepresence. The following two questions were investigated: 1) What is the impact of camera-based viewing position in Augmented Telepresence? 2) What is the impact of depth-aiding augmentations in Augmented Telepresence? Both are addressed through a quality of experience study with non-expert participants, using a custom Augmented Telepresence test system for a task-based experiment. The experiment design combines in-view augmentation, camera view selection, and stereoscopic augmented scene presentation via a head-mounted display to investigate both the independent factors and their joint interaction.

The results indicate that between the two factors, view position has a stronger influence on user experience. Task performance and quality of experience were significantly decreased by viewing positions that force users to rely on stereoscopic depth perception. However, position-assisting view augmentations can mitigate the negative effect of sub-optimal viewing positions; the extent of such mitigation is subject to the augmentation design and appearance.

In aggregate, the works presented in this dissertation cover a broad view of Augmented Telepresence. The individual solutions contribute general insights into Augmented Telepresence system design, complement gaps in the current discourse of specific areas, and provide tools for solving challenges found in enabling the capture, processing, and rendering in real-time-oriented end-to-end systems.

Place, publisher, year, edition, pages
Sundsvall: Mid Sweden University, 2021. p. 70
Series
Mid Sweden University doctoral thesis, ISSN 1652-893X ; 345
National Category
Computer and Information Sciences
Identifiers
urn:nbn:se:miun:diva-41860 (URN)978-91-89341-06-7 (ISBN)
Public defence
2021-05-17, C312, Mittuniversitetet Holmgatan 10, Sundsvall, 14:00 (English)
Opponent
Supervisors
Available from: 2021-04-15 Created: 2021-04-15 Last updated: 2021-11-22Bibliographically approved

Open Access in DiVA

fulltext(359 kB)1156 downloads
File information
File name FULLTEXT02.pdfFile size 359 kBChecksum SHA-512
fa9a4c86e9fc96acbc64aba5171a9ba07251cad8559680bb87ed93fe66f22505b86f66d288af2cc2c8589b01d85a7c62d6c41a8c7adf54776ca36619617ad1a8
Type fulltextMimetype application/pdf

Scopus

Authority records

Dima, ElijsSjöström, MårtenOlsson, RogerKjellqvist, Martin

Search in DiVA

By author/editor
Dima, ElijsSjöström, MårtenOlsson, RogerKjellqvist, Martin
By organisation
Department of Information Systems and Technology
Computer and Information Sciences

Search outside of DiVA

GoogleGoogle Scholar
Total: 1156 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

isbn
urn-nbn

Altmetric score

isbn
urn-nbn
Total: 923 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf