17 research outputs found

    High Dynamic Range Imaging by Perceptual Logarithmic Exposure Merging

    Full text link
    In this paper we emphasize a similarity between the Logarithmic-Type Image Processing (LTIP) model and the Naka-Rushton model of the Human Visual System (HVS). LTIP is a derivation of the Logarithmic Image Processing (LIP), which further replaces the logarithmic function with a ratio of polynomial functions. Based on this similarity, we show that it is possible to present an unifying framework for the High Dynamic Range (HDR) imaging problem, namely that performing exposure merging under the LTIP model is equivalent to standard irradiance map fusion. The resulting HDR algorithm is shown to provide high quality in both subjective and objective evaluations.Comment: 14 pages 8 figures. Accepted at AMCS journa

    High Dynamic Range Spectral Imaging Pipeline For Multispectral Filter Array Cameras

    Get PDF
    Spectral filter arrays imaging exhibits a strong similarity with color filter arrays. This permits us to embed this technology in practical vision systems with little adaptation of the existing solutions. In this communication, we define an imaging pipeline that permits high dynamic range (HDR)-spectral imaging, which is extended from color filter arrays. We propose an implementation of this pipeline on a prototype sensor and evaluate the quality of our implementation results on real data with objective metrics and visual examples. We demonstrate that we reduce noise, and, in particular we solve the problem of noise generated by the lack o

    νŠΉμ§• ν˜Όν•© λ„€νŠΈμ›Œν¬λ₯Ό μ΄μš©ν•œ μ˜μƒ μ •ν•© 기법과 κ³  λͺ…μ•”λΉ„ μ˜μƒλ²• 및 λΉ„λ””μ˜€ κ³  ν•΄μƒν™”μ—μ„œμ˜ μ‘μš©

    Get PDF
    ν•™μœ„λ…Όλ¬Έ (박사) -- μ„œμšΈλŒ€ν•™κ΅ λŒ€ν•™μ› : κ³΅κ³ΌλŒ€ν•™ 전기·컴퓨터곡학뢀, 2020. 8. 쑰남읡.This dissertation presents a deep end-to-end network for high dynamic range (HDR) imaging of dynamic scenes with background and foreground motions. Generating an HDR image from a sequence of multi-exposure images is a challenging process when the images have misalignments by being taken in a dynamic situation. Hence, recent methods first align the multi-exposure images to the reference by using patch matching, optical flow, homography transformation, or attention module before the merging. In this dissertation, a deep network that synthesizes the aligned images as a result of blending the information from multi-exposure images is proposed, because explicitly aligning photos with different exposures is inherently a difficult problem. Specifically, the proposed network generates under/over-exposure images that are structurally aligned to the reference, by blending all the information from the dynamic multi-exposure images. The primary idea is that blending two images in the deep-feature-domain is effective for synthesizing multi-exposure images that are structurally aligned to the reference, resulting in better-aligned images than the pixel-domain blending or geometric transformation methods. Specifically, the proposed alignment network consists of a two-way encoder for extracting features from two images separately, several convolution layers for blending deep features, and a decoder for constructing the aligned images. The proposed network is shown to generate the aligned images with a wide range of exposure differences very well and thus can be effectively used for the HDR imaging of dynamic scenes. Moreover, by adding a simple merging network after the alignment network and training the overall system end-to-end, a performance gain compared to the recent state-of-the-art methods is obtained. This dissertation also presents a deep end-to-end network for video super-resolution (VSR) of frames with motions. To reconstruct an HR frame from a sequence of adjacent frames is a challenging process when the images have misalignments. Hence, recent methods first align the adjacent frames to the reference by using optical flow or adding spatial transformer network (STN). In this dissertation, a deep network that synthesizes the aligned frames as a result of blending the information from adjacent frames is proposed, because explicitly aligning frames is inherently a difficult problem. Specifically, the proposed network generates adjacent frames that are structurally aligned to the reference, by blending all the information from the neighbor frames. The primary idea is that blending two images in the deep-feature-domain is effective for synthesizing frames that are structurally aligned to the reference, resulting in better-aligned images than the pixel-domain blending or geometric transformation methods. Specifically, the proposed alignment network consists of a two-way encoder for extracting features from two images separately, several convolution layers for blending deep features, and a decoder for constructing the aligned images. The proposed network is shown to generate the aligned frames very well and thus can be effectively used for the VSR. Moreover, by adding a simple reconstruction network after the alignment network and training the overall system end-to-end, A performance gain compared to the recent state-of-the-art methods is obtained. In addition to each HDR imaging and VSR network, this dissertation presents a deep end-to-end network for joint HDR-SR of dynamic scenes with background and foreground motions. The proposed HDR imaging and VSR networks enhace the dynamic range and the resolution of images, respectively. However, they can be enhanced simultaneously by a single network. In this dissertation, the network which has same structure of the proposed VSR network is proposed. The network is shown to reconstruct the final results which have higher dynamic range and resolution. It is compared with several methods designed with existing HDR imaging and VSR networks, and shows both qualitatively and quantitatively better results.λ³Έ ν•™μœ„λ…Όλ¬Έμ€ λ°°κ²½ 및 μ „κ²½μ˜ μ›€μ§μž„μ΄ μžˆλŠ” μƒν™©μ—μ„œ κ³  λͺ…μ•”λΉ„ μ˜μƒλ²•μ„ μœ„ν•œ λ”₯ λŸ¬λ‹ λ„€νŠΈμ›Œν¬λ₯Ό μ œμ•ˆν•œλ‹€. μ›€μ§μž„μ΄ μžˆλŠ” μƒν™©μ—μ„œ 촬영된 λ…ΈμΆœμ΄ λ‹€λ₯Έ μ—¬λŸ¬ 영 상듀을 μ΄μš©ν•˜μ—¬ κ³  λͺ…μ•”λΉ„ μ˜μƒμ„ μƒμ„±ν•˜λŠ” 것은 맀우 μ–΄λ €μš΄ μž‘μ—…μ΄λ‹€. κ·Έλ ‡κΈ° λ•Œλ¬Έμ—, μ΅œκ·Όμ— μ œμ•ˆλœ 방법듀은 이미지듀을 ν•©μ„±ν•˜κΈ° 전에 패치 맀칭, μ˜΅ν‹°μ»¬ ν”Œλ‘œμš°, 호λͺ¨κ·Έλž˜ν”Ό λ³€ν™˜ 등을 μ΄μš©ν•˜μ—¬ κ·Έ 이미지듀을 λ¨Όμ € μ •λ ¬ν•œλ‹€. μ‹€μ œλ‘œ λ…ΈμΆœ 정도가 λ‹€λ₯Έ μ—¬λŸ¬ 이미지듀을 μ •λ ¬ν•˜λŠ” 것은 μ•„μ£Ό μ–΄λ €μš΄ μž‘μ—…μ΄κΈ° λ•Œλ¬Έμ—, 이 λ…Όλ¬Έμ—μ„œλŠ” μ—¬λŸ¬ μ΄λ―Έμ§€λ“€λ‘œλΆ€ν„° 얻은 정보λ₯Ό μ„žμ–΄μ„œ μ •λ ¬λœ 이미지λ₯Ό ν•©μ„±ν•˜λŠ” λ„€νŠΈμ›Œν¬λ₯Ό μ œμ•ˆν•œλ‹€. 특히, μ œμ•ˆν•˜λŠ” λ„€νŠΈμ›Œν¬λŠ” 더 밝게 ν˜Ήμ€ μ–΄λ‘‘κ²Œ 촬영된 이미지듀을 쀑간 밝기둜 촬영된 이미지λ₯Ό κΈ°μ€€μœΌλ‘œ μ •λ ¬ν•œλ‹€. μ£Όμš”ν•œ μ•„μ΄λ””μ–΄λŠ” μ •λ ¬λœ 이미지λ₯Ό ν•©μ„±ν•  λ•Œ νŠΉμ§• λ„λ©”μΈμ—μ„œ ν•©μ„±ν•˜λŠ” 것이며, μ΄λŠ” ν”½μ…€ λ„λ©”μΈμ—μ„œ ν•©μ„±ν•˜κ±°λ‚˜ κΈ°ν•˜ν•™μ  λ³€ν™˜μ„ μ΄μš©ν•  λ•Œ 보닀 더 쒋은 μ •λ ¬ κ²°κ³Όλ₯Ό κ°–λŠ”λ‹€. 특히, μ œμ•ˆν•˜λŠ” μ •λ ¬ λ„€νŠΈμ›Œν¬λŠ” 두 갈래의 인코더와 μ»¨λ³Όλ£¨μ…˜ λ ˆμ΄μ–΄λ“€ 그리고 λ””μ½”λ”λ‘œ 이루어져 μžˆλ‹€. 인코더듀은 두 μž…λ ₯ μ΄λ―Έμ§€λ‘œλΆ€ν„° νŠΉμ§•μ„ μΆ”μΆœν•˜κ³ , μ»¨λ³Όλ£¨μ…˜ λ ˆμ΄μ–΄λ“€μ΄ 이 νŠΉμ§•λ“€μ„ μ„žλŠ”λ‹€. λ§ˆμ§€λ§‰μœΌλ‘œ λ””μ½”λ”μ—μ„œ μ •λ ¬λœ 이미지λ₯Ό μƒμ„±ν•œλ‹€. μ œμ•ˆν•˜λŠ” λ„€νŠΈμ›Œν¬λŠ” κ³  λͺ…μ•”λΉ„ μ˜μƒλ²•μ—μ„œ μ‚¬μš©λ  수 μžˆλ„λ‘ λ…ΈμΆœ 정도가 크게 μ°¨μ΄λ‚˜λŠ” μ˜μƒμ—μ„œλ„ 잘 μž‘λ™ν•œλ‹€. κ²Œλ‹€κ°€, κ°„λ‹¨ν•œ 병합 λ„€νŠΈμ›Œν¬λ₯Ό μΆ”κ°€ν•˜κ³  전체 λ„€νŠΈμ›Œν¬λ“€μ„ ν•œ λ²ˆμ— ν•™μŠ΅ν•¨μœΌλ‘œμ„œ, μ΅œκ·Όμ— μ œμ•ˆλœ 방법듀 보닀 더 쒋은 μ„±λŠ₯을 κ°–λŠ”λ‹€. λ˜ν•œ, λ³Έ ν•™μœ„λ…Όλ¬Έμ€ λ™μ˜μƒ λ‚΄ ν”„λ ˆμž„λ“€μ„ μ΄μš©ν•˜λŠ” λΉ„λ””μ˜€ κ³  해상화 방법을 μœ„ν•œ λ”₯ λŸ¬λ‹ λ„€νŠΈμ›Œν¬λ₯Ό μ œμ•ˆν•œλ‹€. λ™μ˜μƒ λ‚΄ μΈμ ‘ν•œ ν”„λ ˆμž„λ“€ μ‚¬μ΄μ—λŠ” μ›€μ§μž„μ΄ μ‘΄μž¬ν•˜κΈ° λ•Œλ¬Έμ—, 이듀을 μ΄μš©ν•˜μ—¬ κ³  ν•΄μƒλ„μ˜ ν”„λ ˆμž„μ„ ν•©μ„±ν•˜λŠ” 것은 μ•„μ£Ό μ–΄λ €μš΄ μž‘μ—…μ΄λ‹€. λ”°λΌμ„œ, μ΅œκ·Όμ— μ œμ•ˆλœ 방법듀은 이 μΈμ ‘ν•œ ν”„λ ˆμž„λ“€μ„ μ •λ ¬ν•˜κΈ° μœ„ν•΄ μ˜΅ν‹°μ»¬ ν”Œλ‘œμš°λ₯Ό κ³„μ‚°ν•˜κ±°λ‚˜ STN을 μΆ”κ°€ν•œλ‹€. μ›€μ§μž„μ΄ μ‘΄μž¬ν•˜λŠ” ν”„λ ˆμž„λ“€μ„ μ •λ ¬ν•˜λŠ” 것은 μ–΄λ €μš΄ 과정이기 λ•Œλ¬Έμ—, 이 λ…Όλ¬Έμ—μ„œλŠ” μΈμ ‘ν•œ ν”„λ ˆμž„λ“€λ‘œλΆ€ν„° 얻은 정보λ₯Ό μ„žμ–΄μ„œ μ •λ ¬λœ ν”„λ ˆμž„μ„ ν•©μ„±ν•˜λŠ” λ„€νŠΈμ›Œν¬λ₯Ό μ œμ•ˆν•œλ‹€. 특히, μ œμ•ˆν•˜λŠ” λ„€νŠΈμ›Œν¬λŠ” μ΄μ›ƒν•œ ν”„λ ˆμž„λ“€μ„ λͺ©ν‘œ ν”„λ ˆμž„μ„ κΈ°μ€€μœΌλ‘œ μ •λ ¬ν•œλ‹€. λ§ˆμ°¬κ°€μ§€λ‘œ μ£Όμš” μ•„μ΄λ””μ–΄λŠ” μ •λ ¬λœ ν”„λ ˆμž„μ„ ν•©μ„±ν•  λ•Œ νŠΉμ§• λ„λ©”μΈμ—μ„œ ν•©μ„±ν•˜λŠ” 것이닀. μ΄λŠ” ν”½μ…€ λ„λ©”μΈμ—μ„œ ν•©μ„±ν•˜κ±°λ‚˜ κΈ°ν•˜ν•™μ  λ³€ν™˜μ„ μ΄μš©ν•  λ•Œ 보닀 더 쒋은 μ •λ ¬ κ²°κ³Όλ₯Ό κ°–λŠ”λ‹€. 특히, μ œμ•ˆν•˜λŠ” μ •λ ¬ λ„€νŠΈμ›Œν¬λŠ” 두 갈래의 인코더와 μ»¨λ³Όλ£¨μ…˜ λ ˆμ΄μ–΄λ“€ 그리고 λ””μ½”λ”λ‘œ 이루어져 μžˆλ‹€. 인코더듀은 두 μž…λ ₯ ν”„λ ˆμž„μœΌλ‘œλΆ€ν„° νŠΉμ§•μ„ μΆ”μΆœν•˜κ³ , μ»¨λ³Όλ£¨μ…˜ λ ˆμ΄μ–΄λ“€μ΄ 이 νŠΉμ§•λ“€μ„ μ„žλŠ”λ‹€. λ§ˆμ§€λ§‰μœΌλ‘œ λ””μ½”λ”μ—μ„œ μ •λ ¬λœ ν”„λ ˆμž„μ„ μƒμ„±ν•œλ‹€. μ œμ•ˆν•˜λŠ” λ„€νŠΈμ›Œν¬λŠ” μΈμ ‘ν•œ ν”„λ ˆμž„λ“€μ„ 잘 μ •λ ¬ν•˜λ©°, λΉ„λ””μ˜€ κ³  해상화에 효과적으둜 μ‚¬μš©λ  수 μžˆλ‹€. κ²Œλ‹€κ°€ 병합 λ„€νŠΈμ›Œν¬λ₯Ό μΆ”κ°€ν•˜κ³  전체 λ„€νŠΈμ›Œν¬λ“€μ„ ν•œ λ²ˆμ— ν•™μŠ΅ν•¨μœΌλ‘œμ„œ, μ΅œκ·Όμ— μ œμ•ˆλœ μ—¬λŸ¬ 방법듀 보닀 더 쒋은 μ„±λŠ₯을 κ°–λŠ”λ‹€. κ³  λͺ…μ•”λΉ„ μ˜μƒλ²•κ³Ό λΉ„λ””μ˜€ κ³  해상화에 λ”ν•˜μ—¬, λ³Έ ν•™μœ„λ…Όλ¬Έμ€ λͺ…암비와 해상도λ₯Ό ν•œ λ²ˆμ— ν–₯μƒμ‹œν‚€λŠ” λ”₯ λ„€νŠΈμ›Œν¬λ₯Ό μ œμ•ˆν•œλ‹€. μ•žμ—μ„œ μ œμ•ˆλœ 두 λ„€νŠΈμ›Œν¬λ“€μ€ 각각 λͺ…암비와 해상도λ₯Ό ν–₯μƒμ‹œν‚¨λ‹€. ν•˜μ§€λ§Œ, 그듀은 ν•˜λ‚˜μ˜ λ„€νŠΈμ›Œν¬λ₯Ό 톡해 ν•œ λ²ˆμ— ν–₯상될 수 μžˆλ‹€. 이 λ…Όλ¬Έμ—μ„œλŠ” λΉ„λ””μ˜€ 고해상화λ₯Ό μœ„ν•΄ μ œμ•ˆν•œ λ„€νŠΈμ›Œν¬μ™€ 같은 ꡬ쑰의 λ„€νŠΈμ›Œν¬λ₯Ό μ΄μš©ν•˜λ©°, 더 높은 λͺ…암비와 해상도λ₯Ό κ°–λŠ” μ΅œμ’… κ²°κ³Όλ₯Ό 생성해낼 수 μžˆλ‹€. 이 방법은 기쑴의 κ³  λͺ…μ•”λΉ„ μ˜μƒλ²•κ³Ό λΉ„λ””μ˜€ 고해상화λ₯Ό μœ„ν•œ λ„€νŠΈμ›Œν¬λ“€μ„ μ‘°ν•©ν•˜λŠ” 것 보닀 μ •μ„±μ μœΌλ‘œ 그리고 μ •λŸ‰μ μœΌλ‘œ 더 쒋은 κ²°κ³Όλ₯Ό λ§Œλ“€μ–΄ λ‚Έλ‹€.1 Introduction 1 2 Related Work 7 2.1 High Dynamic Range Imaging 7 2.1.1 Rejecting Regions with Motions 7 2.1.2 Alignment Before Merging 8 2.1.3 Patch-based Reconstruction 9 2.1.4 Deep-learning-based Methods 9 2.1.5 Single-Image HDRI 10 2.2 Video Super-resolution 11 2.2.1 Deep Single Image Super-resolution 11 2.2.2 Deep Video Super-resolution 12 3 High Dynamic Range Imaging 13 3.1 Motivation 13 3.2 Proposed Method 14 3.2.1 Overall Pipeline 14 3.2.2 Alignment Network 15 3.2.3 Merging Network 19 3.2.4 Integrated HDR imaging network 20 3.3 Datasets 21 3.3.1 Kalantari Dataset and Ground Truth Aligned Images 21 3.3.2 Preprocessing 21 3.3.3 Patch Generation 22 3.4 Experimental Results 23 3.4.1 Evaluation Metrics 23 3.4.2 Ablation Studies 23 3.4.3 Comparisons with State-of-the-Art Methods 25 3.4.4 Application to the Case of More Numbers of Exposures 29 3.4.5 Pre-processing for other HDR imaging methods 32 4 Video Super-resolution 36 4.1 Motivation 36 4.2 Proposed Method 37 4.2.1 Overall Pipeline 37 4.2.2 Alignment Network 38 4.2.3 Reconstruction Network 40 4.2.4 Integrated VSR network 42 4.3 Experimental Results 42 4.3.1 Dataset 42 4.3.2 Ablation Study 42 4.3.3 Capability of DSBN for alignment 44 4.3.4 Comparisons with State-of-the-Art Methods 45 5 Joint HDR and SR 51 5.1 Proposed Method 51 5.1.1 Feature Blending Network 51 5.1.2 Joint HDR-SR Network 51 5.1.3 Existing VSR Network 52 5.1.4 Existing HDR Network 53 5.2 Experimental Results 53 6 Conclusion 58 Abstract (In Korean) 71Docto

    YDA gârüntü gâlgeleme gidermede gelişmişlik seviyesi ve YDA gârüntüler için nesnel bir gâlgeleme giderme kalite metriği.

    Get PDF
    Despite the emergence of new HDR acquisition methods, the multiple exposure technique (MET) is still the most popular one. The application of MET on dynamic scenes is a challenging task due to the diversity of motion patterns and uncontrollable factors such as sensor noise, scene occlusion and performance concerns on some platforms with limited computational capability. Currently, there are already more than 50 deghosting algorithms proposed for artifact-free HDR imaging of dynamic scenes and it is expected that this number will grow in the future. Due to the large number of algorithms, it is a difficult and time-consuming task to conduct subjective experiments for benchmarking recently proposed algorithms. In this thesis, first, a taxonomy of HDR deghosting methods and the key characteristics of each group of algorithms are introduced. Next, the potential artifacts which are observed frequently in the outputs of HDR deghosting algorithms are defined and an objective HDR image deghosting quality metric is presented. It is found that the proposed metric is well correlated with the human preferences and it may be used as a reference for benchmarking current and future HDR image deghosting algorithmsPh.D. - Doctoral Progra

    Inverse tone mapping

    Get PDF
    The introduction of High Dynamic Range Imaging in computer graphics has produced a novelty in Imaging that can be compared to the introduction of colour photography or even more. Light can now be captured, stored, processed, and finally visualised without losing information. Moreover, new applications that can exploit physical values of the light have been introduced such as re-lighting of synthetic/real objects, or enhanced visualisation of scenes. However, these new processing and visualisation techniques cannot be applied to movies and pictures that have been produced by photography and cinematography in more than one hundred years. This thesis introduces a general framework for expanding legacy content into High Dynamic Range content. The expansion is achieved avoiding artefacts, producing images suitable for visualisation and re-lighting of synthetic/real objects. Moreover, it is presented a methodology based on psychophysical experiments and computational metrics to measure performances of expansion algorithms. Finally, a compression scheme, inspired by the framework, for High Dynamic Range Textures, is proposed and evaluated

    High quality high dynamic range imaging

    Get PDF
    corecore