论文标题
手语识别的视觉方法:基于方式的评论
Visual Methods for Sign Language Recognition: A Modality-Based Review
论文作者
论文摘要
连续多模式流的手语视觉识别仍然是最具挑战性的领域之一。 人类行动识别的最新进展是利用大量数据的基于GPU的学习的提升,并且正在接近类似人类的表现。 然后,他们容易为聋哑和听力障碍社区创建互动服务。 预计将在未来几年中增长的人口。 本文旨在以标志性的视觉理解作为范围来审查人类行为识别文献。 分析的方法将主要根据所利用的单峰输入,其相对多模式组合和管道步骤进行组织。 在每个部分中,我们将详细介绍并比较相关数据集,方法,然后区分适合创建手语相关服务的仍在开放的贡献路径。 将特别注意处理面部表情和连续签名的方法和商业解决方案。
Sign language visual recognition from continuous multi-modal streams is still one of the most challenging fields. Recent advances in human actions recognition are exploiting the ascension of GPU-based learning from massive data, and are getting closer to human-like performances. They are then prone to creating interactive services for the deaf and hearing-impaired communities. A population that is expected to grow considerably in the years to come. This paper aims at reviewing the human actions recognition literature with the sign-language visual understanding as a scope. The methods analyzed will be mainly organized according to the different types of unimodal inputs exploited, their relative multi-modal combinations and pipeline steps. In each section, we will detail and compare the related datasets, approaches then distinguish the still open contribution paths suitable for the creation of sign language related services. Special attention will be paid to the approaches and commercial solutions handling facial expressions and continuous signing.