4,635 research outputs found
Detecting and removing visual distractors for video aesthetic enhancement
Personal videos often contain visual distractors, which are objects that are accidentally captured that can distract viewers from focusing on the main subjects. We propose a method to automatically detect and localize these distractors through learning from a manually labeled dataset. To achieve spatially and temporally coherent detection, we propose extracting features at the Temporal-Superpixel (TSP) level using a traditional SVM-based learning framework. We also experiment with end-to-end learning using Convolutional Neural Networks (CNNs), which achieves slightly higher performance than other methods. The classification result is further refined in a post-processing step based on graph-cut optimization. Experimental results show that our method achieves an accuracy of 81% and a recall of 86%. We demonstrate several ways of removing the detected distractors to improve the video quality, including video hole filling; video frame replacement; and camera path re-planning. The user study results show that our method can significantly improve the aesthetic quality of videos
Video Upright Adjustment and Stabilization
Upright adjustment, Video stabilization, Camera pathWe propose a novel video upright adjustment method that can reliably correct slanted video contents that are often found in casual videos. Our approach combines deep learning and Bayesian inference to estimate accurate rotation angles from video frames. We train a convolutional neural network to obtain initial estimates of the rotation angles of input video frames. The initial estimates from the network are temporally inconsistent and inaccurate. To resolve this, we use Bayesian inference. We analyze estimation errors of the network, and derive an error model. We then use the error model to formulate video upright adjustment as a maximum a posteriori problem where we estimate consistent rotation angles from the initial estimates, while respecting relative rotations between consecutive frames. Finally, we propose a joint approach to video stabilization and upright adjustment, which minimizes information loss caused by separately handling stabilization and upright adjustment. Experimental results show that our video upright adjustment method can effectively correct slanted video contents, and its combination with video stabilization can achieve visually pleasing results from shaky and slanted videos.openI. INTRODUCTION
1.1. Related work
II. ROTATION ESTIMATION NETWORK
III. ERROR ANALYSIS
IV. VIDEO UPRIGHT ADJUSTMENT
4.1. Initial angle estimation
4.2. Robust angle estimation
4.3. Optimization
4.4. Warping
V. JOINT UPRIGHT ADJUSTMENT AND STABILIZATION
5.1. Bundled camera paths for video stabilization
5.2. Joint approach
VI. EXPERIMENTS
VII. CONCLUSION
ReferencesCNN)μ νλ ¨μν¨λ€. μ κ²½λ§μ μ΄κΈ° μΆμ μΉλ μμ ν μ ννμ§ μμΌλ©° μκ°μ μΌλ‘λ μΌκ΄λμ§ μλλ€. μ΄λ₯Ό ν΄κ²°νκΈ° μν΄ λ² μ΄μ§μ μΈνΌλ°μ€λ₯Ό μ¬μ©νλ€. λ³Έ λ
Όλ¬Έμ μ κ²½λ§μ μΆμ μ€λ₯λ₯Ό λΆμνκ³ μ€λ₯ λͺ¨λΈμ λμΆνλ€. κ·Έλ° λ€μ μ€λ₯ λͺ¨λΈμ μ¬μ©νμ¬ μ°μ νλ μ κ°μ μλ νμ κ°λ(Relative rotation angle)λ₯Ό λ°μνλ©΄μ μ΄κΈ° μΆμ μΉλ‘λΆν° μκ°μ μΌλ‘ μΌκ΄λ νμ κ°λλ₯Ό μΆμ νλ μ΅λ μ¬ν λ¬Έμ (Maximum a posteriori problem)λ‘ λμμ μν 보μ μ 곡μννλ€. λ§μ§λ§μΌλ‘, λμμ μν 보μ λ° λμμ μμ ν(Video stabilization)μ λν λμ μ κ·Ό λ°©λ²μ μ μνμ¬ μν 보μ κ³Ό μμ νλ₯Ό λ³λλ‘ μνν λ λ°μνλ κ³΅κ° μ 보 μμ€κ³Ό μ°μ°λμ μ΅μννλ©° μμ νμ μ±λ₯μ μ΅λννλ€. μ€ν κ²°κ³Όμ λ°λ₯΄λ©΄ λμμ μν 보μ μΌλ‘ κΈ°μΈμ΄μ§ λμμμ ν¨κ³Όμ μΌλ‘ 보μ ν μ μμΌλ©° λμμ μμ ν λ°©λ²κ³Ό κ²°ν©νμ¬ νλ€λ¦¬κ³ κΈ°μΈμ΄μ§ λμμμΌλ‘λΆν° μκ°μ μΌλ‘ λ§μ‘±μ€λ¬μ΄ μλ‘μ΄ λμμμ νλν μ μλ€.λ³Έ λ
Όλ¬Έμ μΌλ°μΈλ€μ΄ 촬μν λμμμμ νν λ°μνλ λ¬Έμ μΈ κΈ°μΈμ΄μ§μ μ κ±°νμ¬ μνμ΄ μ¬λ°λ₯Έ λμμμ νλν μ μκ² νλ λμμ μν 보μ (Video upright adjustment) λ°©λ²μ μ μνλ€. λ³Έ λ
Όλ¬Έμ μ κ·Ό λ°©μμ λ₯ λ¬λ(Deep learning)κ³Ό λ² μ΄μ§μ μΈνΌλ°μ€(Bayesian inference)λ₯Ό κ²°ν©νμ¬ λμμ νλ μ(Frame)μμ μ νν κ°λλ₯Ό μΆμ νλ€. λ¨Όμ μ
λ ₯ λμμ νλ μμ νμ κ°λμ μ΄κΈ° μΆμ μΉλ₯Ό μ»κΈ° μν΄ νμ μ κ²½λ§(Convolutional neural networkMasterdCollectio
Lessons in looking : the digital audiovisual essay
This thesis examines the contemporary practice of the digital audiovisual essay,
which is defined as a material form of thinking at the crossroads of academic
textual analysis, personal cinephilia, and popular online fandom practices, to
suggest that it allows rich epistemological discoveries not only about individual
films and viewing experiences, but also about how cinema is perceived in the
context of digitally mediated audiovisual culture.
Chapter one advances five key defining tensions of the digital audiovisual
essay: its object is the investigation of specific films and cinephiliac experiences;
it uses a performative research methodology based on the affordances of digital
viewing and editing technologies; it exists primarily in Web 2.0 and takes
advantage of its collaborative and dialogical modes of production; it is a βrich
text objectβ that continuously tests the different contributions of both verbal and
audiovisual forms of communication to the production of knowledge about the
cinema; and finally, the digital audiovisual essay has an important pedagogical
potential, not only for those who watch it, but especially for those who practice
it.
Chapter two presents the theoretical framework of the dissertation,
challenges the βnewnessβ of the digital audiovisual essay, and suggests that any
investigation of this cultural practice must address its ideological implications
and its role in the context of contemporary audiovisual culture. Accordingly, it
relates the editing and compositional techniques of the digital audiovisual essay
with modernist montage and suggests that the audiovisual essay has not only
inherited, but has also updated and enhanced the dialectical interdependency
between critical and consumerism drives that shaped modernismβs ambiguous
relation to mass culture.
The final chapter examines four case studies (David Bordwell, Catherine
Grant, ::kogonada, and Kevin B. Lee) that showcase the contradictory tensions of
this cultural practice and broad our understanding of the politics of the digital
audiovisual essay
Content Seeking Students: Site-and-Sound Bites as Participants in Ubiquitous Social Computing
Discussion of digital, collaborative environments for architectural work often focuses on the structure of discourse, rather than upon its substance. An implied assumption is that the various means of electronic-based communication are suitable for any kind of subject matter, whether visual, sound-based, or text. Our project team has chosen to challenge this assumption by example: We have created new media artifacts for collaborative architectural education. Our project is an attempt to leverage on-going research concerning the efficacy of "ubiquitous social computingβ (USC) for design-studio teaching. With a pilot project already put in place by one of our team's leaders, we have supplemented graphic and verbal communication among participants with purpose-crafted video for their use and exchange. Smart-screens, placed strategically within students' "social enclaves,β provide access to curated content. Our approach challenges traditional educational emphasis upon explicit types of architectural knowledge. The construction of tacit knowledge, usually derived from first-hand architectural experiences, is here effected by mediated, digital-based experiences. Nevertheless, the social dimension of the USC framework may be significant in negotiating the interface between immediate and mediated experiences
Bootlegger : Turning Fans into Film Crew
Abstract Bootlegger is a system for creating multi-camera films of live music events using mobile devices. Using readily available technology and a synthesis of film-making conventions, the system coordinates music fans at live shows into an improvised film crew, suggesting shots, collating footage and generating rich metadata in real time. Bootlegger is part of a research project exploring adapting professional media workflows to amateur contexts in order to lower the bar to entry for media production. By enabling concert-goers ..
- β¦