Signall和Kintrans等许多公司都开发了手动跟踪软件,这些软件试图使数百万依靠手语的人轻松与任何人交流。但是,他们几乎没有成功。在这里,Google AI实验室的最新手动跟踪算法带来了它的承诺,可以使该软件成为原本本来可以的。
Google AI Labs的手工跟踪算法使用您的智能手机及其相机创建一个复杂的详细片段,详细介绍了一个人的手,然后可以为沟通而进行跟踪。Google researchers Valentin Bazarevsky and Fan Zhang said in a blog post, ‘Whereas current state-of-the-art approaches rely primarily on powerful desktop environments for inference, our method achieves real-time performance on a mobile phone, and even scales to multiple hands.’
They further added, ‘Robust real-time hand perception is a decidedly challenging computer vision task, as hands often occlude themselves or each other (e.g., finger/palm occlusions and handshakes) and lack high contrast patterns.’ The researchers have managed to make the new hand-tracking algorithm capable of calculating the hand signals faster by simplifying the process as much as they possibly could, thus leading to lesser data and eventually lesser processing time.
新的手跟踪算法首先是在人的手掌上训练,而不是占据完整的尺寸。然后,单独的算法开始看手指和手掌,然后在指关节,指尖等上分配了21个坐标。摆姿势和照明条件使AI能够学习这些坐标。
开发人员团队已经开发了其代码,同时希望其他人能够提出创新的使用方式并改善它。新系统还利用了Google现有的MediaPipe增强现实框架。研究人员在其博客文章中说:“我们希望向更广泛的研发社区提供这种手感的功能,将导致创造性用例的出现,刺激新的应用程序和新的研究途径。”