英语
前馈
计算机科学
录像
人工智能
代表(政治)
计算机视觉
神经科学
生物
工程类
政治学
政治
广告
控制工程
业务
法学
作者
Lu Fang,Mengqi Ji,Xiaoyun Yuan,Jing He,Jianing Zhang,Yinheng Zhu,Tian Zheng,Leyao Liu,Bin Wang,Qionghai Dai
出处
期刊:Engineering
[Elsevier]
日期:2022-02-17
卷期号:25: 101-109
标识
DOI:10.1016/j.eng.2021.12.012
摘要
Sensing and understanding large-scale dynamic scenes require a high-performance imaging system. Conventional imaging systems pursue higher capability by simply increasing the pixel resolution via stitching cameras at the expense of a bulky system. Moreover, they strictly follow the feedforward pathway: that is, their pixel-level sensing is independent of semantic understanding. Differently, a human visual system owns superiority with both feedforward and feedback pathways: The feedforward pathway extracts object representation (referred to as memory engram) from visual inputs, while, in the feedback pathway, the associated engram is reactivated to generate hypotheses about an object. Inspired by this, we propose a dual-pathway imaging mechanism, called engram-driven videography. We start by abstracting the holistic representation of the scene, which is associated bidirectionally with local details, driven by an instance-level engram. Technically, the entire system works by alternating between the excitation–inhibition and association states. In the former state, pixel-level details become dynamically consolidated or inhibited to strengthen the instance-level engram. In the association state, the spatially and temporally consistent content becomes synthesized driven by its engram for outstanding videography quality of future scenes. The association state serves as the imaging of future scenes by synthesizing spatially and temporally consistent content driven by its engram. Results of extensive simulations and experiments demonstrate that the proposed system revolutionizes the conventional videography paradigm and shows great potential for videography of large-scale scenes with multi-objects.
科研通智能强力驱动
Strongly Powered by AbleSci AI