Mid Sweden University

miun.sePublications
Change search
Link to record
Permanent link

Direct link
Li, Yongwei
Publications (9 of 9) Show all publications
Xia, Z., Wang, C., Li, Y., Yu, B., Zhan, Y., Li, Q., . . . Ma, B. (2023). Geometrical attacks resilient statistical watermark decoder using polar harmonic Fourier moments. Journal of the Franklin Institute, 360(7), 4493-4518
Open this publication in new window or tab >>Geometrical attacks resilient statistical watermark decoder using polar harmonic Fourier moments
Show others...
2023 (English)In: Journal of the Franklin Institute, ISSN 0016-0032, E-ISSN 1879-2693, Vol. 360, no 7, p. 4493-4518Article in journal (Refereed) Published
Abstract [en]

This paper presents a new robust multiplicative watermark detector. Due to the strong robustness against various attacks, polar harmonic Fourier moment (PHFM) magnitudes are used as the employed watermark carrier. The distribution of PHFM magnitudes is highly non-Gaussian and can be properly modeled by a heavy-tailed probability density function (PDF). In this paper, we proved that Weibull distribution can suitably fit the distribution of PHFM magnitudes, and based on this, we presented a statistics-based watermark decoder by using the Weibull as a prior for the PHFM magnitudes. In watermark embedding, a multiplicative manner was used to embed watermark information in PHFM magnitudes of the highest entropy blocks to achieve better robustness and imperceptibility. In watermark detection, we developed a Weibull distribution-based statistical watermark decoder, which uses the maximum likelihood (ML) decision rule. Compared with Bessel K form (BKF), Cauchy, and generalized Gaussian (GG)-based decoders, the Weibull-based decoder demonstrates stronger robustness. In addition, the proposed watermark decoder is more robust against geometrical and common image processing attacks than existing statistical watermark decoders. 

National Category
Electrical Engineering, Electronic Engineering, Information Engineering
Identifiers
urn:nbn:se:miun:diva-48014 (URN)10.1016/j.jfranklin.2023.02.028 (DOI)000957725900001 ()2-s2.0-85150233835 (Scopus ID)
Available from: 2023-03-28 Created: 2023-03-28 Last updated: 2025-09-25Bibliographically approved
Li, Y., Pla, F., Sjöström, M. & Fernandez-Beltran, R. (2022). Simultaneous Color Restoration and Depth Estimation in Light Field Imaging. IEEE Access, 10, 49599-49610
Open this publication in new window or tab >>Simultaneous Color Restoration and Depth Estimation in Light Field Imaging
2022 (English)In: IEEE Access, E-ISSN 2169-3536, Vol. 10, p. 49599-49610Article in journal (Refereed) Published
Abstract [en]

Recent studies in the light field imaging have shown the potential and advantages of different light field information processes. In most of the existing techniques, the processing pipeline of light field has been treated in a step-by-step manner, and each step is considered to be independent from the others. For example, in light field color demosaicing, inferring the scene geometry is treated as an irrelevant and negligible task, and vice versa. Such processing techniques may fail due to the inherent connection among different steps, and result in both corrupted post-processing and defective pre-processing results. In this paper, we address the interaction between color interpolation and depth estimation in light field, and propose a probabilistic approach to handle these two processing steps jointly. This probabilistic framework is based on a Markov Random Fields —Collaborative Graph Model for simultaneous Demosaicing and Depth Estimation (CGMDD)—to explore the color-depth interdependence from general light field sampling. Experimental results show that both image interpolation quality and depth estimation can benefit from their interaction, mainly for processes such as image demosaicing which are shown to be sensitive to depth information, especially for light field sampling with large baselines.

Keywords
Image color analysis, Estimation, Pipelines, Imaging, Interpolation, Probabilistic logic, Markov random fields, Light field, demosaicing, depth estimation, graph model
National Category
Signal Processing Computer graphics and computer vision
Identifiers
urn:nbn:se:miun:diva-44953 (URN)10.1109/ACCESS.2022.3172343 (DOI)000795627800001 ()2-s2.0-85129643302 (Scopus ID)
Available from: 2022-05-05 Created: 2022-05-05 Last updated: 2025-09-25Bibliographically approved
Li, Y. (2020). Computational Light Field Photography: Depth Estimation, Demosaicing, and Super-Resolution. (Doctoral dissertation). Sundsvall: Mid Sweden University
Open this publication in new window or tab >>Computational Light Field Photography: Depth Estimation, Demosaicing, and Super-Resolution
2020 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

The transition of camera technology from film-based cameras to digital cameras has been witnessed in the past twenty years, along with impressive technological advances in processing massively digitized media content. Today, a new evolution emerged -- the migration from 2D content to immersive perception. This rising trend has a profound and long-term impact to our society, fostering technologies such as teleconferencing and remote surgery. The trend is also reflected in the scientific research community, and more intention has been drawn to the light field and its applications.

 

The purpose of this dissertation is to develop a better understanding of light field structure by analyzing its sampling behavior and to addresses three problems concerning the light field processing pipeline: 1) How to address the depth estimation problem when there is limited color and texture information. 2) How to improve the rendered image quality by using the inherent depth information. 3) How to solve the interdependence conflict of demosaicing and depth estimation.

 

The first problem is solved by a hybrid depth estimation approach that combines advantages of correspondence matching and depth-from-focus, where occlusion is handled by involving multiple depth maps in a voting scheme. The second problem is divided into two specific tasks -- demosaicing and super-resolution, where depth-assisted light field analysis is employed to surpass the competence of traditional image processing. The third problem is tackled with an inferential graph model that encodes the connections between demosaicing and depth estimation explicitly, and jointly performs a global optimization for both tasks.

 

The proposed depth estimation approach shows a noticeable improvement in point clouds and depth maps, compared with references methods. Furthermore, the objective metrics and visual quality are compared with classical sensor-based demosaicing and multi-image super-resolution to show the effectiveness of the proposed depth-assisted light field processing methods. Finally, a multi-task graph model is proposed to challenge the performance of the sequential light field image processing pipeline. The proposed method is validated with various kinds of light fields, and outperforms the state-of-the-art in both demosaicing and depth estimation tasks.

 

The works presented in this dissertation raise a novel view of the light field data structure in general, and provide tools to solve image processing problems in specific. The impact of the outcome can be manifold: To support scientific research with light field microscopes, to stabilize the performance of range cameras for industrial applications, as well as to provide individuals with a high-quality immersive experience.

Abstract [sv]

Under de senaste tjugo åren har det skett en övergång från filmbaserad till digital kamerateknik, parallellt med en imponerande teknisk utveckling inom bearbetning av omfattande digitaliserat medieinnehåll. På senare tid även en ny utvecklingslinje – övergången från 2D-innehåll till omslutande perception. Detta är en utveckling som har långtgående och långvarig påverkan på samhället och främjar arbetsmetoder såsom telekonferens och fjärrstyrd kirurgi. Den här utvecklingst trenden återspeglas också i det vetenskapliga forskningssamhället, och mer uppmärksamhet har lagts på light field och dess olika tillämpningsområden.

Syftet med avhandlingen är att nå en bättre förståelse av strukturen i light field genom att analysera hur light field samplas, och att lösa tre problem inom behandlingsprocessen av light field: 1) Hur problemet med djupestimering kan lösas med begränsad information om färg och textur. 2) Hur renderad bildkvalitet kan förbättras genom att utnyttja den inneboende djupinformationen. 3) Hur beroendekonflikten mellan demosaicing (färgfiltrering) och djupestimering kan lösas.

Det första problemet har lösts genom en hybridmetod för djupestimering, som kombinerar styrkorna med korrespondensmatchning och djup från fokus, där ocklusion hanteras genom att använda flera djupkartor i ett röstningssystem. Det andra problemet delas upp i två separata moment – demosaicing och superupplösning, där djupassisterad analys av light field används för att överträffa kapaciteten för traditionell bildbehandling. Det tredje problemet har angripits med en inferentiell grafmodell som explicit kopplar samman demosaicing och djupestimering, och samfällt utför en global optimering för båda dessa processteg.

Den metod för djupestimering som föreslås producerar visuellt tilltalande punktmoln och djupkartor, jämfört med andra referensmetoder. Objektiva mätvärden och visuell kvalitet jämförs vidare med klassisk sensorbaserad demosaicing och superupplösning från multipla bilder, för att visa effektiviteten hos de föreslagna metoderna för djupassisterad behandling av light field. En multitaskande grafmodell föreslås även för att matcha och överträffa prestandan hos sekventiell light field-baserad bildbehandling. Den metod som föreslås valideras med olika sorters light fields och överträffar de bästa existerande metoderna inom både demosaicing och djupestimering.

De arbeten som presenteras i avhandlingen utgör ett nytt sätt att betrakta den generella datastrukturen hos light field, och tillhandahåller verktyg för att lösa specifika bildbehandlingsproblem. Effekterna av dessa resultat kan vara många, till exempel som stöd för vetenskaplig forskning om light field-baserade mikroskop, för att förbättra prestandan hos avståndsmätande kameror i industriella tillämpningar, såväl som för att erbjuda högkvalitativa omslutande mediaupplevelser.

Place, publisher, year, edition, pages
Sundsvall: Mid Sweden University, 2020. p. 56
Series
Mid Sweden University doctoral thesis, ISSN 1652-893X ; 327
Keywords
Light field, computational photography, depth estimation, demosaicing, super-resolution
National Category
Computer and Information Sciences
Identifiers
urn:nbn:se:miun:diva-39005 (URN)978-91-88947-57-4 (ISBN)
Public defence
2020-06-10, C312, Holmgatan 10, Sundsvall, 09:00 (English)
Opponent
Supervisors
Projects
European Unions Horizon 2020 under the Marie Sklodowska-Curie grant agreement No 676401, European Training Network on Full Parallax Imaging
Note

Vid tidpunkten för disputationen var följande delarbeten opublicerade: delarbete 4 manuskript och delarbete 5 inskickat.

At the time of the doctoral defence the following papers were unpublished: paper 4 manuscript and paper 5 submitted.

Available from: 2020-05-18 Created: 2020-05-11 Last updated: 2025-09-25Bibliographically approved
Li, Y. & Sjöström, M. (2019). Depth-Assisted Demosaicing for Light Field Data in Layered Object Space. In: 2019 IEEE International Conference on Image Processing (ICIP): . Paper presented at 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22-25 September, 2019 (pp. 3746-3750). IEEE, Article ID 8803441.
Open this publication in new window or tab >>Depth-Assisted Demosaicing for Light Field Data in Layered Object Space
2019 (English)In: 2019 IEEE International Conference on Image Processing (ICIP), IEEE, 2019, p. 3746-3750, article id 8803441Conference paper, Published paper (Refereed)
Abstract [en]

Light field technology, which emerged as a solution to the increasing demands of visually immersive experience, has shown its extraordinary potential for scene content representation and reconstruction. Unlike conventional photography that maps the 3D scenery onto a 2D plane by a projective transformation, light field preserves both the spatial and angular information, enabling further processing steps such as computational refocusing and image-based rendering. However, there are still gaps that have been barely studied, such as the light field demosaicing process. In this paper, we propose a depth-assisted demosaicing method for light field data. First, we exploit the sampling geometry of the light field data with respect to the scene content using the ray-tracing technique and develop a sampling model of light field capture. Then we carry out the demosaicing process in a layered object space with object-space sampling adjacencies rather than pixel placement. Finally, we compare our results with state-of-art approaches and discuss about the potential research directions of the proposed sampling model to show the significance of our approach.

Place, publisher, year, edition, pages
IEEE, 2019
Keywords
Lenses, Cameras, Image color analysis, Three-dimensional displays, Microoptics, Interpolation, Two dimensional displays, Light field, demosaicing, object space, ray-tracing technique
National Category
Computer and Information Sciences
Identifiers
urn:nbn:se:miun:diva-37690 (URN)10.1109/ICIP.2019.8803441 (DOI)000521828603177 ()2-s2.0-85076819023 (Scopus ID)978-1-5386-6249-6 (ISBN)
Conference
2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22-25 September, 2019
Available from: 2019-11-15 Created: 2019-11-15 Last updated: 2025-09-25Bibliographically approved
Li, Y., Olsson, R. & Sjöström, M. (2018). An analysis of demosaicing for plenoptic capture based on ray optics. In: Proceedings of 3DTV Conference 2018: . Paper presented at 3D at any scale and any perspective, 3-5 June 2018, Stockholm – Helsinki – Stockholm. , Article ID 8478476.
Open this publication in new window or tab >>An analysis of demosaicing for plenoptic capture based on ray optics
2018 (English)In: Proceedings of 3DTV Conference 2018, 2018, article id 8478476Conference paper, Published paper (Refereed)
Abstract [en]

The plenoptic camera is gaining more and more attention as it capturesthe 4D light field of a scene with a single shot and enablesa wide range of post-processing applications. However, the preprocessing steps for captured raw data, such as demosaicing, have been overlooked. Most existing decoding pipelines for plenoptic cameras still apply demosaicing schemes which are developed for conventional cameras. In this paper, we analyze the sampling pattern of microlens-based plenoptic cameras by ray-tracing techniques and ray phase space analysis. The goal of this work is to demonstrate guidelines and principles for demosaicing the plenoptic captures by taking the unique microlens array design into account. We show that the sampling of the plenoptic camera behaves differently from that of a conventional camera and the desired demosaicing scheme is depth-dependent.

Keywords
Light field, plenoptic camera, depth, image demosaicing
National Category
Other Electrical Engineering, Electronic Engineering, Information Engineering
Identifiers
urn:nbn:se:miun:diva-33618 (URN)10.1109/3DTV.2018.8478476 (DOI)000454903900008 ()2-s2.0-85056161198 (Scopus ID)978-1-5386-6125-3 (ISBN)
Conference
3D at any scale and any perspective, 3-5 June 2018, Stockholm – Helsinki – Stockholm
Available from: 2018-05-15 Created: 2018-05-15 Last updated: 2025-09-25Bibliographically approved
Li, Y., Scrofani, G., Sjöström, M. & Martinez-Corraly, M. (2018). Area-Based Depth Estimation for Monochromatic Feature-Sparse Orthographic Capture. In: 2018 26th European Signal Processing Conference (EUSIPCO): . Paper presented at EUSIPCO 2018, 26th European Signal Processing Conference, Rome, Italy, September 3-7, 2018 (pp. 206-210). IEEE conference proceedings, Article ID 8553336.
Open this publication in new window or tab >>Area-Based Depth Estimation for Monochromatic Feature-Sparse Orthographic Capture
2018 (English)In: 2018 26th European Signal Processing Conference (EUSIPCO), IEEE conference proceedings, 2018, p. 206-210, article id 8553336Conference paper, Published paper (Refereed)
Abstract [en]

With the rapid development of light field technology, depth estimation has been highlighted as one of the critical problems in the field, and a number of approaches have been proposed to extract the depth of the scene. However, depthestimation by stereo matching becomes difficult and unreliable when the captured images lack both color and feature information. In this paper, we propose a scheme that extracts robust depth from monochromatic, feature-sparse scenes recorded in orthographic sub-aperture images. Unlike approaches which relyon the rich color and texture information across the sub-aperture views, our approach is based on depth from focus techniques. First, we superimpose shifted sub-aperture images on top of anarbitrarily chosen central image. To focus on different depths, the shift amount is varied based on the micro-lens array properties. Next, an area-based depth estimation approach is applied tofind the best match among the focal stack and generate the dense depth map. This process is repeated for each sub-aperture image. Finally, occlusions are handled by merging depth maps generated from different central images followed by a voting process. Results show that the proposed scheme is more suitable than conventional depth estimation approaches in the context of orthographic captures that have insufficient color and feature information, such as microscopic fluorescence imaging.

Place, publisher, year, edition, pages
IEEE conference proceedings, 2018
Keywords
Depth estimation, integral imaging, orthographic views, depth from focus
National Category
Computer Sciences
Identifiers
urn:nbn:se:miun:diva-34418 (URN)000455614900042 ()2-s2.0-85059811493 (Scopus ID)
Conference
EUSIPCO 2018, 26th European Signal Processing Conference, Rome, Italy, September 3-7, 2018
Available from: 2018-09-14 Created: 2018-09-14 Last updated: 2025-09-25Bibliographically approved
Wang, C., Wang, X., Li, Y., Xia, Z. & Zhang, C. (2018). Quaternion polar harmonic Fourier moments for color images. Information Sciences, 450, 141-156
Open this publication in new window or tab >>Quaternion polar harmonic Fourier moments for color images
Show others...
2018 (English)In: Information Sciences, ISSN 0020-0255, E-ISSN 1872-6291, Vol. 450, p. 141-156Article in journal (Refereed) Published
Abstract [en]

This paper proposes quaternion polar harmonic Fourier moments (QPHFM) for color image processing and analyzes the properties of QPHFM. After extending Chebyshev–Fourier moments (CHFM) to quaternion Chebyshev-Fourier moments (QCHFM), comparison experiments, including image reconstruction and color image object recognition, on the performance of QPHFM and quaternion Zernike moments (QZM), quaternion pseudo-Zernike moments (QPZM), quaternion orthogonal Fourier-Mellin moments (QOFMM), QCHFM, and quaternion radial harmonic Fourier moments (QRHFM) are carried out. Experimental results show QPHFM can achieve an ideal performance in image reconstruction and invariant object recognition in noise-free and noisy conditions. In addition, this paper discusses the importance of phase information of quaternion orthogonal moments in image reconstruction. 

Keywords
Image reconstruction, Moment invariant, Object recognition, Orthogonal moment, Phase, Quaternion polar harmonic Fourier moments
National Category
Information Systems, Social aspects
Identifiers
urn:nbn:se:miun:diva-33500 (URN)10.1016/j.ins.2018.03.040 (DOI)000432646100008 ()2-s2.0-85044451202 (Scopus ID)
Available from: 2018-04-16 Created: 2018-04-16 Last updated: 2025-09-25Bibliographically approved
Li, Y., Pla, F. & Sjöström, M.A Collaborative Graph Model for Light Field Demosaicing and Depth Estimation.
Open this publication in new window or tab >>A Collaborative Graph Model for Light Field Demosaicing and Depth Estimation
(English)Manuscript (preprint) (Other academic)
National Category
Computer and Information Sciences
Identifiers
urn:nbn:se:miun:diva-39013 (URN)
Available from: 2020-05-13 Created: 2020-05-13 Last updated: 2025-09-25Bibliographically approved
Li, Y. & Sjöström, M. Depth-Assisted Light Field Super-Resolution in Layered Object Space.
Open this publication in new window or tab >>Depth-Assisted Light Field Super-Resolution in Layered Object Space
(English)Manuscript (preprint) (Other academic)
Abstract [en]

The captured light field may fail to reconstruct fine details of the scene due to under-sampling problem of lightfield acquisition devices. Therefore,super-resolution is required to restore high-frequency information from the light field and to improve the quality of therendered views. Conventionalsuper-resolution algorithms are not ideal for light field data, as they do not utilize the full potential of light field 4D structure, while existing light fieldsuper-resolution algorithms rely heavily on the accuracy of the estimated depth and perform complex sub-pixeldisparity estimation. In this paper, we propose a new light field super-resolution algorithm which can address depthuncertainty with a layered object space. First, a pixel-wise depth estimation is performed from the resampled views.Then we divide the depth range into finite layers and back-project pixels onto these layers in order to address the sub-pixel depth error. Finally, two super-resolution schemes: in-depth warping and cross-depth learning, are introduced tosuper-resolve the views from light field data redundancy. The algorithms is tested with extensive datasets, and theresults show that our method attains favorable results in both visual assessment and objective metrics compared toother light field super-resolution methods.

Keywords
Light field, Image warping, Ray-tracing, Super-resolution
National Category
Computer and Information Sciences
Identifiers
urn:nbn:se:miun:diva-39012 (URN)
Available from: 2020-05-13 Created: 2020-05-13 Last updated: 2025-09-25Bibliographically approved
Organisations

Search in DiVA

Show all publications