4.8 Article

One-Shot Imitation Drone Filming of Human Motion Videos

出版社

IEEE COMPUTER SOC
DOI: 10.1109/TPAMI.2021.3067359

关键词

Videos; Cameras; Drones; Training; Task analysis; Feature extraction; Robot vision systems; Autonomous drone cinematography; imitation learning; one-shot imitation filming; style feature; motion prediction

资金

  1. Hong Kong General Research Fund (GRF) [16203319]
  2. Natural Science Foundation of China [U1909203]

向作者/读者索取更多资源

Imitation learning is applied to autonomous camera systems, but current methods require a large number of training videos with similar styles and struggle to generalize to different styles. To address this, a framework called one-shot imitation filming is proposed, which can imitate a filming style without style-specific model training using two key techniques: filming style feature extraction and camera motion prediction.
Imitation learning has recently been applied to mimic the operation of a cameraman in existing autonomous camera systems. To imitate a certain demonstration video, existing methods require users to collect a significant number of training videos with a similar filming style. Because the trained model is style-specific, it is challenging to generalize the model to imitate other videos with a different filming style. To address this problem, we propose a framework that we term one-shot imitation filming, which can imitate a filming style by seeing only a single demonstration video of the target style without style-specific model training. This is achieved by two key enabling techniques: 1) filming style feature extraction, which encodes sequential cinematic characteristics of a variable-length video clip into a fixed-length feature vector; and 2) camera motion prediction, which dynamically plans the camera trajectory to reproduce the filming style of the demo video. We implemented the approach with a deep neural network and deployed it on a 6 degrees of freedom (DOF) drone system by first predicting the future camera motions, and then converting them into the drone's control commands via an odometer. Our experimental results on comprehensive datasets and showcases exhibit that the proposed approach achieves significant improvements over conventional baselines, and our approach can mimic the footage of an unseen style with high fidelity.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据