CN101739122A - Gesture Recognition and Tracking Method - Google Patents

Gesture Recognition and Tracking Method Download PDF

Info

Publication number
CN101739122A
CN101739122A CN200810177689A CN200810177689A CN101739122A CN 101739122 A CN101739122 A CN 101739122A CN 200810177689 A CN200810177689 A CN 200810177689A CN 200810177689 A CN200810177689 A CN 200810177689A CN 101739122 A CN101739122 A CN 101739122A
Authority
CN
China
Prior art keywords
gesture
image
tracking
block
identification according
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN200810177689A
Other languages
Chinese (zh)
Inventor
陈水来
许哲豪
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
TOP-SEED TECHNOLOGY CORP
Original Assignee
TOP-SEED TECHNOLOGY CORP
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by TOP-SEED TECHNOLOGY CORP filed Critical TOP-SEED TECHNOLOGY CORP
Priority to CN200810177689A priority Critical patent/CN101739122A/en
Publication of CN101739122A publication Critical patent/CN101739122A/en
Pending legal-status Critical Current

Links

Images

Landscapes

  • Image Analysis (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

一种手势辨识及追踪的方法,利用一影像传感器撷取一手势影像,再由对该影像进行处理,用以辨识并追踪该影像,而执行该手势影像所对应的动作。首先,该方法先对该影像进行前置处理,然后进行影像运动检测,接着分析该影像的特征,进而判断该影像的手势状态。若该手势影像为一移动手势,则检测并追踪该移动手势的一中心坐标,并输出该移动手势的该中心坐标;若该手势影像为一命令手势,则输出该命令手势所对应的动作指令。为此,达到以该数字信号处理器为硬件平台,进行自然手势辨识及追踪的目的。

Figure 200810177689

A method for gesture recognition and tracking uses an image sensor to capture a gesture image, and then processes the image to recognize and track the image, and execute the action corresponding to the gesture image. First, the method pre-processes the image, then performs image motion detection, and then analyzes the characteristics of the image to determine the gesture state of the image. If the gesture image is a moving gesture, a center coordinate of the moving gesture is detected and tracked, and the center coordinate of the moving gesture is output; if the gesture image is a command gesture, the action instruction corresponding to the command gesture is output. To this end, the purpose of natural gesture recognition and tracking is achieved using the digital signal processor as a hardware platform.

Figure 200810177689

Description

手势辨识及追踪的方法 Gesture Recognition and Tracking Method

技术领域technical field

本发明涉及一种手势辨识及追踪的方法,尤其涉及一种利用一数字信号处理器对该手势进行辨识及追踪的方法。The invention relates to a gesture recognition and tracking method, in particular to a method for using a digital signal processor to recognize and track the gesture.

背景技术Background technique

随着计算机科技的进步,人机互动的改进一直是许多研究所专注的课题,从早期的键盘、鼠标及游戏杆,都是为了能让使用者能够更方便地操作计算机。在许多虚拟实境(virtual reality)和多媒体系统(multimedia system)的应用中,对于3D对象操作、3D虚拟产品展示系统、计算机绘图系统及动作类或运动类的电玩游戏…等范围的应用,必须具有3D和高自由度的输入装置。然而,前述的键盘、鼠标及游戏杆等输入装置,并无法方便且合适地提供使用者与系统间自然且直接的互动效果。With the advancement of computer technology, the improvement of human-computer interaction has always been the focus of many research institutes. From the early days of keyboards, mice and joysticks, they are all designed to allow users to operate computers more conveniently. In many virtual reality (virtual reality) and multimedia system (multimedia system) applications, for 3D object manipulation, 3D virtual product display systems, computer graphics systems, and action or sports video games... and other applications, must Input device with 3D and high degree of freedom. However, the aforementioned input devices such as keyboards, mice, and joysticks cannot conveniently and properly provide natural and direct interaction effects between the user and the system.

由于人机界面(human computer interface)的应用日益普及,包括手势辨识、语音辨识、或是肢体语言辨识等等,都已经被广泛研究并且应用在日常生活之中,其中以利用手势作为输入界面是最自然且最直接的。所以,将手势辨识应用在机器视觉及虚拟实境等方面,已成为新的发展趋势。Since the application of human computer interface (human computer interface) is becoming more and more popular, including gesture recognition, voice recognition, or body language recognition, etc., it has been widely studied and applied in daily life. The most natural and direct. Therefore, applying gesture recognition to machine vision and virtual reality has become a new development trend.

在手势辨识与手势追踪作为计算机输入界面的实际应用上,以手套为基础的方法(glove-based method)的使用,可提供精确并迅速的感应及辨识效果。所谓以手套为基础的方法,是指使用者需要戴上数据手套,而数据手套上装设有接触式感应器,可精确地撷取到使用者手指的弯曲度或手部动作,并可将手的活动以电子信号方式传送到计算机。由分析这些电子信号,系统能迅速地辨识出手势的动作状态。惟,此类装设有接触式感应器的数据手套的产品单价相当昂贵,并且数据手套的尺寸规格种类也不多样,造成使用者在穿戴合适度上需有更多取舍,另外穿戴厚重的数据手套也容易使手感到疲劳而限制了使用者的活动,导致造成使用上的不方便。In the practical application of gesture recognition and gesture tracking as a computer input interface, the use of a glove-based method can provide accurate and rapid sensing and recognition effects. The so-called glove-based method means that the user needs to wear a data glove, and the data glove is equipped with a contact sensor, which can accurately capture the curvature of the user's finger or hand movement, and can move the hand activity is transmitted to the computer as an electronic signal. By analyzing these electronic signals, the system can quickly identify the action state of the gesture. However, the unit price of such data gloves equipped with contact sensors is quite expensive, and the size specifications of data gloves are not diverse, causing users to have more choices in terms of wearing fit, and wearing heavy data gloves. Gloves also tend to make hands feel tired and restrict the user's activities, resulting in inconvenience in use.

因此,如何设计出一种能降低开发成本及简化操作程序的手势辨识及追踪的方法,使得缩短使用者与机器间的距离,并且使人机界面朝向更有效率、更合乎人性化及更多样化的方向迈进,乃为本案所欲行克服并加以解决的一大课题。Therefore, how to design a gesture recognition and tracking method that can reduce development costs and simplify operating procedures, shorten the distance between the user and the machine, and make the man-machine interface more efficient, more humane and more Moving in the direction of diversification is a major issue to be overcome and resolved in this case.

发明内容Contents of the invention

有鉴于此,本发明提供一种手势辨识及追踪的方法,利用一影像传感器撷取一手势影像,再由一数字信号处理器对该影像进行处理,用以辨识并追踪该手势影像,而执行该手势影像所对应的动作指令。为此,达到以该数字信号处理器为硬件平台,进行自然手势辨识及追踪的目的。In view of this, the present invention provides a gesture recognition and tracking method, using an image sensor to capture a gesture image, and then processing the image by a digital signal processor to identify and track the gesture image, and execute The action command corresponding to the gesture image. Therefore, the purpose of using the digital signal processor as a hardware platform to recognize and track natural gestures is achieved.

为了解决上述问题,本发明提供一种手势辨识及追踪的方法。该方法的步骤为:首先,对该手势影像进行前置处理。然后,检测该手势影像的一最大运动区块而定义为一手势区块。接着,分析该手势区块的特征,进而判断该手势区块为一移动确认手势、一命令手势或其它未定义手势。最后,若该手势区块为该移动确认手势,接着变换为一移动手势,且该移动手势在移动过程中未停止超过一动作时间,则检测并追踪该移动手势的一中心坐标,并输出该移动手势的该中心坐标。In order to solve the above problems, the present invention provides a gesture recognition and tracking method. The steps of the method are as follows: firstly, performing pre-processing on the gesture image. Then, a maximum motion block of the gesture image is detected and defined as a gesture block. Then, the feature of the gesture block is analyzed, and then it is determined that the gesture block is a movement confirmation gesture, a command gesture or other undefined gestures. Finally, if the gesture block is the movement confirmation gesture, then transformed into a movement gesture, and the movement gesture does not stop for more than an action time during the movement, then detect and track a center coordinate of the movement gesture, and output the The center coordinates of the move gesture.

本发明的有益功效在于,可以达到以数字信号处理器为硬件平台,进行自然手势辨识及追踪的目的。The beneficial effect of the present invention is that the purpose of recognizing and tracking natural gestures can be achieved by using the digital signal processor as the hardware platform.

以下结合附图和具体实施例对本发明进行详细描述,但不作为对本发明的限定。The present invention will be described in detail below in conjunction with the accompanying drawings and specific embodiments, but not as a limitation of the present invention.

附图说明Description of drawings

图1为本发明一手势辨识及追踪的方法流程图;FIG. 1 is a flowchart of a gesture recognition and tracking method of the present invention;

图2为本发明动态影像差分计算的示意图;Fig. 2 is a schematic diagram of dynamic image differential calculation of the present invention;

图3为本发明水平及垂直投影量计算的示意图;Fig. 3 is the schematic diagram of horizontal and vertical projection amount calculation of the present invention;

图4A至图4C为本发明区块标签化过程的示意图;4A to 4C are schematic diagrams of the block labeling process of the present invention;

图5A至图5B为本发明持续追踪移动手势中心坐标过程的示意图;及5A to 5B are schematic diagrams of the process of continuously tracking the coordinates of the center of the moving gesture in the present invention; and

图6为本发明一手势辨识及追踪的装置方块图。FIG. 6 is a block diagram of a gesture recognition and tracking device of the present invention.

其中,附图标记Among them, reference signs

10    影像传感器10 image sensor

20    数字信号处理器20 digital signal processor

30    第一内存30 first memory

40    第二内存40 second memory

50    视频输出模块50 video output modules

60    数据输入/输出模块60 data input/output modules

Ps    样板Ps template

S102-S502    步骤S102-S502 steps

具体实施方式Detailed ways

有关本发明的技术内容及详细说明,配合图式说明如下:Relevant technical content and detailed description of the present invention, cooperate drawing description as follows:

请参见图1为本发明一手势辨识及追踪的方法流程图。该方法是利用一影像传感器撷取一手势影像,再由一数字信号处理器对该手势影像进行处理。该方法的步骤如下详述,并且以手势动作取代鼠标动作的应用为例。Please refer to FIG. 1 , which is a flowchart of a gesture recognition and tracking method of the present invention. In the method, an image sensor is used to capture a gesture image, and then a digital signal processor is used to process the gesture image. The steps of the method are described in detail as follows, and an application in which a gesture action replaces a mouse action is taken as an example.

首先,该数字信号处理器对该手势影像进行前置处理(S102)。由于未经处理的影像通常会含有一些噪声,使得辨识错误的机率增加,并且,过多而无用的信息也会降低整体执行的效率,故此,撷取得到的影像在分析之前都会经过前置处理。该影像前置处理步骤(S102)依序为:先调该手势影像大小为适合演算范围;然后对该手势影像进行色彩转换,将影像由全彩影像(24bit RGB)降为灰阶影像(8bit gray level);最后再通过影像低通滤波器(image low pass filter)滤除该手势影像的点状噪声,以利后续实际演算。如此,通过影像前置处理不仅增加辨识的准确度,而且可以节省储存数据的空间以及提升传输速度。First, the digital signal processor performs pre-processing on the gesture image (S102). Because unprocessed images usually contain some noise, which increases the probability of identification errors, and too much useless information will also reduce the efficiency of the overall execution. Therefore, the captured images will be pre-processed before analysis . The image pre-processing step (S102) is as follows: first adjust the size of the gesture image to fit the calculation range; then perform color conversion on the gesture image, and reduce the image from a full-color image (24bit RGB) to a grayscale image (8bit gray level); finally, the point-like noise of the gesture image is filtered out by an image low pass filter to facilitate subsequent actual calculations. In this way, image pre-processing not only increases the accuracy of recognition, but also saves data storage space and improves transmission speed.

然后检测该手势影像的一最大运动区块而定义为一手势区块(S104)。该影像运动检测步骤(S104)依序为:先利用动态影像差分,再产生二进制影像,并采用逻辑运算该些二进制影像,计算出该手势影像中所有移动部分;然后利用统计该手势影像的垂直及水平亮点数量,并选取垂直轴及水平轴的最大投影量区域的逻辑,而找出该手势影像的最大移动区域;接着利用膨胀(dilation)技术,对该垂直轴及水平轴最大投影量的逻辑区域的细部破碎影像进行填补;最后使用标签编号,以排除非手势运动区域,并且保留且计算出最大连通区域以检测出该最大运动区块。Then detect a maximum motion block of the gesture image and define it as a gesture block ( S104 ). The image motion detection step (S104) is as follows in sequence: First, use dynamic image difference to generate binary images, and use logic operations on these binary images to calculate all moving parts in the gesture image; and the number of horizontal bright spots, and select the logic of the maximum projection area of the vertical axis and the horizontal axis to find the maximum moving area of the gesture image; The detailed fragmented images of the logical area are filled; finally, the label number is used to exclude the non-gesture motion area, and the maximum connected area is reserved and calculated to detect the maximum motion block.

请参见图2本发明动态影像差分计算的示意图。如图所示,是利用连续三帧手势影像而计算出真正移动物体。其中,目前帧灰阶影像标示为M2,前一帧灰阶影像标示为M1,以及前两帧灰阶影像标示为M0。并且,设定一门阈值(threshold value),用以做为转换该些灰阶影像成为二进制影像(binary image)的依据。先将目前帧灰阶影像M2减去前一帧灰阶影像M1,得到一新的灰阶影像,然后,再将该新的灰阶影像的像素与该门阈值比较:若该灰阶影像的像素大于或等于该门阈值,则设为亮点;若该灰阶影像的像素小于该门阈值,则设为暗点,而得到一新的二进制影像M3。同样地,将前一帧灰阶影像M1减去前二帧灰阶影像M0,而得到另一新的灰阶影像,并且再于该门阈值比较,而得到另一新的二进制影像M4。最后,再将两二进制影像M3、M4进行逻辑(AND)运算,即可得到最后移动部分的二进制影像M5。Please refer to FIG. 2 for a schematic diagram of dynamic image difference calculation in the present invention. As shown in the figure, the real moving object is calculated by using three consecutive frames of gesture images. Wherein, the current grayscale image frame is marked as M2, the previous grayscale image frame is marked as M1, and the previous two grayscale image frames are marked as M0. In addition, a threshold value is set to serve as a basis for converting the grayscale images into binary images. First subtract the previous grayscale image M1 from the current frame grayscale image M2 to obtain a new grayscale image, and then compare the pixels of the new grayscale image with the gate threshold: if the grayscale image’s If the pixel is greater than or equal to the gate threshold, it is set as a bright spot; if the pixel of the grayscale image is smaller than the gate threshold, it is set as a dark spot, and a new binary image M3 is obtained. Similarly, the previous two frames of gray-scale image M0 are subtracted from the previous frame of gray-scale image M1 to obtain another new gray-scale image, and then compared with the gate threshold to obtain another new binary image M4. Finally, logical (AND) operation is performed on the two binary images M3 and M4 to obtain the binary image M5 of the last moving part.

请参见图3为本发明水平及垂直投影量计算的示意图。将所得到的最后移动部分的二进制影像M5进行统计垂直亮点数量及水平亮点数量,以找出最大移动区域。如图所示,有两块较大移动量区块,分别标示为X与Y,经计算后,水平投影量得到A、B两个较大区域;垂直投影量得到C、D两个较大区域。然后,取水平轴最大投影量区域B及垂直轴最大投影量区域C的逻辑区域,即可得到最大移动量区块X。Please refer to FIG. 3 which is a schematic diagram of calculation of horizontal and vertical projection amounts in the present invention. Count the number of vertical bright spots and the number of horizontal bright spots on the obtained binary image M5 of the last moving part to find out the maximum moving area. As shown in the figure, there are two large movement blocks, marked as X and Y respectively. After calculation, two large areas A and B are obtained by the horizontal projection amount; two large areas C and D are obtained by the vertical projection amount. area. Then, the logical area of the maximum projection amount area B on the horizontal axis and the maximum projection amount area C on the vertical axis can be obtained to obtain the maximum movement amount block X.

请参见图4A至图4C为本发明区块标签化过程的示意图。�����垂直轴及水平轴最大投影量的逻辑区域的细部破碎影像填补后,以二进制数值表示二进制影像的亮点与暗点(如图4A标示为0和1)。然后,再针对连通区域予以重新编号并计算面积,最后只保留最大部分(如图4B及图4C标示为2的区域),而排除非手势运动的区域。Please refer to FIG. 4A to FIG. 4C which are schematic diagrams of the block labeling process of the present invention. After the detailed fragmented images in the logical area of the maximum projection amount on the vertical and horizontal axes are filled, the bright and dark points of the binary image are represented by binary values (marked as 0 and 1 in FIG. 4A ). Then, renumber the connected regions and calculate the area, and finally only keep the largest part (the region marked as 2 in FIG. 4B and FIG. 4C ), and exclude the non-gesture movement region.

然后分析该手势区块的特征,进而判断该手势区块为一移动确认手势、一命令手势或其它未定义手势(S106)。该手势区块的特征比对是利用各手势所产生的相对极值点发生位置及各极值差值,与数据库内的手势影像数据逐一进行比对,并且将比对结果储存于内存缓冲区中。例如,当操作者张开五根手指时,因为定义相对极大值出现在指尖部分,相对极小值出现在两手指蹼连接处及手掌左右两边。所以,该手势区块的特征则具有五个极大值及六个极小值。故此,当该手势区块与数据库内的手势影像数据比对后,则辨识出该操作者的手势为伸出五根手指的状态。Then analyze the feature of the gesture block, and then determine that the gesture block is a movement confirmation gesture, a command gesture or other undefined gestures (S106). The feature comparison of the gesture block is to use the relative extreme point occurrence position and the extreme value difference generated by each gesture to compare with the gesture image data in the database one by one, and store the comparison result in the memory buffer middle. For example, when the operator spreads five fingers, because it is defined that the relative maximum value appears at the fingertips, and the relative minimum value appears at the junction of the webs of the two fingers and the left and right sides of the palm. Therefore, the feature of the gesture block has five maximum values and six minimum values. Therefore, when the gesture block is compared with the gesture image data in the database, it is recognized that the operator's gesture is a state of extending five fingers.

再请参见图1。然后若该手势区块为该移动确认手势(S108),则判断是否接续变换为一移动手势(S200)。若非接续变换为该移动手势,则重新执行步骤(S102)。若判断出操作者手势由该移动确认手势变换为该移动手势,则可控制一光标为移动动作。其中,该移动确认手势可定义为由食指及中指形成的一V字形状,当操作者伸出食指及中指形成该V字形状时,该手势区块与数据库内的手势影像数据比对后,则辨识出该操作者的手势是为伸出食指及中指的状态,而为该移动确认手势。当该移动手势为移动状态时,则产生移动该光标的动作。接着判断该移动手势是否持续停止移动超过一动作时间(S300)。若操作者停止手势移动超过该动作时间,则重新执行步骤(S102)。其中,该动作时间可依操作者的使用方式或使用需要予以设定不同的时间长度,例如,该动作时间可设定为1秒钟。若操作者停止手势移动未超过该动作时间,则判断是否检测到该移动手势的一中心坐标(S400)。若未检测到该移动手势的该中心坐标,则重新检测该移动手势的该中心坐标(S404),然后再执行步骤(S400),重新判断是否检测到该移动手势的该中心坐标。由于该移动手势定义为五指握拳形成的一拳头形状,因此,当该中心坐标的启始追踪时,利用圆形霍夫转换(circularHough transfer),以统计方式找出最多圆心相同的点,即可判断为该中心坐标所在。See Figure 1 again. Then if the gesture block is the movement confirmation gesture (S108), it is determined whether to continue to transform into a movement gesture (S200). If the movement gesture is not continued, the step (S102) is re-executed. If it is determined that the operator's gesture is changed from the movement confirmation gesture to the movement gesture, a cursor can be controlled to move. Wherein, the movement confirmation gesture can be defined as a V shape formed by the index finger and the middle finger. When the operator stretches out the index finger and the middle finger to form the V shape, after comparing the gesture block with the gesture image data in the database, Then it is recognized that the operator's gesture is a state of extending the index finger and the middle finger, and the gesture is confirmed for the movement. When the moving gesture is in a moving state, an action of moving the cursor is generated. Then it is judged whether the moving gesture continues to stop moving for more than an action time ( S300 ). If the operator stops the gesture movement for more than the action time, step (S102) is re-executed. Wherein, the action time can be set to different lengths of time according to the operator's use mode or needs, for example, the action time can be set to 1 second. If the operator stops the movement of the gesture within the action time, it is determined whether a center coordinate of the movement gesture is detected (S400). If the center coordinate of the movement gesture is not detected, re-detect the center coordinate of the movement gesture (S404), and then perform step (S400) to re-determine whether the center coordinate of the movement gesture is detected. Since the movement gesture is defined as a fist shape formed by clenching a fist with five fingers, when the center coordinates are started to be tracked, the circular Hough transfer is used to statistically find the most points with the same center of circle, that is, Determine where the center coordinates are located.

若检测到该移动手势的该中心坐标,则利用快速积分表(sum ofaccumulator table,SAT)方式(S402),进行判断是否追踪到该移动手势的该中心坐标(S500)。若未追踪到该移动手势的该中心坐标,则再执行步骤(S404),重新检测该移动手势的该中心坐标。若追踪到该移动手势的该中心坐标,则输出该移动手势的该中心坐标(S502),然后重新执行步骤(S102)。If the center coordinate of the moving gesture is detected, a fast integration table (sum ofaccumulator table, SAT) method (S402) is used to determine whether the center coordinate of the moving gesture is tracked (S500). If the center coordinate of the moving gesture is not tracked, step (S404) is executed again to re-detect the center coordinate of the moving gesture. If the center coordinate of the movement gesture is tracked, output the center coordinate of the movement gesture (S502), and then re-execute the step (S102).

请参见图5A至图5B本发明持续追踪移动手势中心坐标过程的示意图。当检测出该移动手势的该中心坐标,以中心坐标上、下、左、右各取20像素,产生一40像素*40像素大小的样板Ps并计算所有像素灰阶值的加总值,之后再利用快速积分表方式,于中心坐标正负60像素的区域进行逐一搜寻比对加总值差值最小或相等的位置,即为该移动手势新的中心位置,此时产生新坐标储存至该内存缓冲区中,并将搜寻区域移到新的中心坐标。如此通过该方形样板Ps在该搜寻区域由左上至右下逐一比对,以保持持续追踪该中心坐标。至终,若该移动手势停止不再移动并持续超过该动作时间,则结束追踪,并且重新执行步骤(S102)。Please refer to FIG. 5A to FIG. 5B for schematic diagrams of the process of continuously tracking the center coordinates of the moving gesture in the present invention. When the center coordinates of the moving gesture are detected, take 20 pixels from the center coordinates up, down, left, and right to generate a template Ps with a size of 40 pixels*40 pixels and calculate the sum of the grayscale values of all pixels, and then Then use the fast integral table method to search one by one in the area of plus or minus 60 pixels of the center coordinates for the position where the difference of the total value of the comparison is the smallest or equal, which is the new center position of the movement gesture. At this time, the new coordinates are generated and stored in the memory buffer and move the search area to the new center coordinates. In this way, the square template Ps is compared one by one in the search area from the upper left to the lower right, so as to keep tracking the center coordinates. Finally, if the moving gesture stops and lasts longer than the action time, the tracking is ended and the step ( S102 ) is re-executed.

此外,在步骤S106中,若该手势区块被检测为该命令手势(S110),则输出该命令手势所对应的动作指令(S112),然后,再重新执行步骤(S102)。例如,该命令手势可被定义为食指形成的一1字形状,且该命令手势所对应的动作指令为一单击动作。当操作者伸出食指形成1字形状时,该手势区块与数据库内的手势影像数据比对后,则辨识出该操作者的手势是为伸出食指的状态,因此,在光标所在坐标的处执行单击动作。该命令手势可依操作者的操作习惯,自行定义为其它手势形状,或也可以定义其它有效命令手势,以执行各别对应的动作指令。In addition, in step S106, if the gesture block is detected as the command gesture (S110), an action instruction corresponding to the command gesture is output (S112), and then step (S102) is re-executed. For example, the command gesture can be defined as a 1 shape formed by the index finger, and the action instruction corresponding to the command gesture is a click action. When the operator stretches out the index finger to form a 1 shape, after comparing the gesture block with the gesture image data in the database, it is recognized that the operator’s gesture is in the state of extending the index finger. Therefore, at the coordinates where the cursor is located Perform a click action. The command gesture can be defined as other gesture shapes according to the operator's operating habits, or other valid command gestures can be defined to execute respective corresponding action commands.

此外,在步骤S106中,若该手势区块被检测为该未定义手势(S114),即该命令手势非为该移动确认手势(食指及中指形成的一V字形状)、该移动手势(五指握拳形成的一拳头形状)或该命令手势(食指形成的一1字形状)的任一手势,而为一无效的未定义手势,则重新执行步骤(S102)。In addition, in step S106, if the gesture block is detected as the undefined gesture (S114), that is, the command gesture is not the movement confirmation gesture (a V shape formed by the index finger and the middle finger), the movement gesture (five-finger A fist shape formed by clenching a fist) or any gesture of the command gesture (a 1-shaped shape formed by the index finger), if it is an invalid undefined gesture, then re-execute the step (S102).

请参见图6为本发明一手势辨识及追踪的装置方块图。该装置包含一影像传感器10、一数字信号处理器20、一第一内存30、一第二内存40及一视频输出模块50。Please refer to FIG. 6 which is a block diagram of a gesture recognition and tracking device of the present invention. The device includes an image sensor 10 , a digital signal processor 20 , a first memory 30 , a second memory 40 and a video output module 50 .

该影像传感器10用以撷取一手势影像。该数字信号处理器20电性连接该影像传感器10,用以提供一算法对该手势影像进行处理。该第一内存30电性连接该数字信号处理器20,用以储存该数字信号处理器20的该算法,并提供大量演算数据的储存。其中,该第一内存30可为一闪存(flash memory)。该第二内存40电性连接该数字信号处理器20,用以提供该数字信号处理器20运算时所需的记忆缓冲区。其中,该第二内存40可为一随机存取内存(randomaccess memory)。该视频输出模块50电性连接该数字信号处理器20,用以输出该数字信号处理器20运算后的一影像演算结果。其中,该此影像演算结果可输出至模拟显示装置(图未示),如电视或监视器;或数字显示装置(图未示),如液晶显示器。该数字信号处理器20更可电性连接一数据输入/输出模块60,用以不仅通过不同输出界面传送该影像演算结果至其它装置(图未示),如计算机或电玩主机等独立运作装置,同时也接受外界控制命令,用以调整该数字信号处理器20的运算。The image sensor 10 is used to capture a gesture image. The digital signal processor 20 is electrically connected to the image sensor 10 for providing an algorithm to process the gesture image. The first memory 30 is electrically connected to the digital signal processor 20 for storing the algorithm of the digital signal processor 20 and providing storage of a large amount of calculation data. Wherein, the first memory 30 can be a flash memory (flash memory). The second memory 40 is electrically connected to the digital signal processor 20 to provide a memory buffer required by the digital signal processor 20 for operation. Wherein, the second memory 40 can be a random access memory (random access memory). The video output module 50 is electrically connected to the digital signal processor 20 for outputting an image calculation result obtained by the digital signal processor 20 . Wherein, the image calculation result can be output to an analog display device (not shown), such as a TV or a monitor; or a digital display device (not shown), such as a liquid crystal display. The digital signal processor 20 can be further electrically connected to a data input/output module 60, so as to not only transmit the image calculation result to other devices (not shown), such as independent operating devices such as computers or video game hosts, through different output interfaces, At the same time, it also accepts external control commands to adjust the operation of the digital signal processor 20 .

综上所述,本发明具有以下的优点:In summary, the present invention has the following advantages:

1、利用该数字信号处理器为硬件平台,进行自然手势辨识及追踪,不须额外穿戴手套或特殊图标、色彩或发光装置,可大大降低开发成本及简化操作程序。1. Use the digital signal processor as a hardware platform to recognize and track natural gestures without wearing additional gloves or special icons, colors or light-emitting devices, which can greatly reduce development costs and simplify operating procedures.

2、该数字信号处理器的硬件平台可以连接其它外部独立装置,提高可携式的便利及弹性的扩充应用。2. The hardware platform of the digital signal processor can be connected to other external independent devices, which improves the convenience of portability and flexible expansion of applications.

当然,本发明还可有其它多种实施例,在不背离本发明精神及其实质的情况下,熟悉本领域的技术人员当可根据本发明作出各种相应的改变和变形,但这些相应的改变和变形都应属于本发明所附的权利要求的保护范围。Certainly, the present invention also can have other multiple embodiments, without departing from the spirit and essence of the present invention, those skilled in the art can make various corresponding changes and deformations according to the present invention, but these corresponding Changes and deformations should belong to the scope of protection of the appended claims of the present invention.

Claims (13)

1. the method for gesture identification and tracking utilizes an image sensor to capture a gesture image, by this gesture image is handled, it is characterized in that again the step of this method comprises:
(a) this gesture image is carried out pre-process;
(b) detect a largest motion block of this gesture image and be defined as a gesture block;
(c) analyzing the feature of this gesture block, is to move to confirm a gesture or an order gesture to judge this gesture block;
(d) if this gesture block, judge then whether this gesture block continues for should move confirming gesture and be transformed to one and move gesture;
(e) if this gesture block continues and is transformed to this and moves gesture, and this mobile gesture does not stop then to detect and follow the trail of the centre coordinate that this moves gesture above an actuation time in moving process; And
(f) export this centre coordinate that this moves gesture, and re-execute step (a).
2. the method for gesture identification according to claim 1 and tracking is characterized in that, this step (a) comprises:
(a1) adjust this gesture image size for being fit to the calculation scope;
(a2) this gesture image is carried out color conversion; And
(a3) the point-like noise of this gesture image of filtering.
3. the method for gesture identification according to claim 1 and tracking is characterized in that, this step (b) comprises:
(b1) utilize the dynamic image difference, calculate all movable parts in this gesture image;
(b2) utilize the vertical and horizontal bright spot quantity of adding up this gesture image, find out the maximum moving area of this gesture image;
(b3) utilize expansion technique, calculate the broken image of thin portion and fill up; And
(b4) use tag number, calculate largest connected zone to detect this largest motion block.
4. the method for gesture identification according to claim 1 and tracking is characterized in that, this step (e) comprises:
(e1) utilize circular Hough conversion, detect this centre coordinate that this moves gesture; And
(e2) utilize the quick point table, follow the trail of this centre coordinate that this moves gesture.
5. the method for gesture identification according to claim 1 and tracking is characterized in that, in step (d), if this gesture block was confirmed gesture for moving, but the conversion that continues is non-for this mobile gesture, then re-executes step (a).
6. the method for gesture identification according to claim 1 and tracking is characterized in that, in step (d), if this gesture block is then exported the pairing action command of this order gesture, and re-executed step (a) for this order gesture.
7. the method for gesture identification according to claim 1 and tracking is characterized in that, in step (d), if this gesture block is a undefined gesture, then re-executes step (a).
8. the method for gesture identification according to claim 1 and tracking is characterized in that, in step (e), if this mobile gesture stops to surpass this actuation time, then re-executes step (a) in moving process.
9. the method for gesture identification according to claim 1 and tracking is characterized in that, in step (e), if can't detect side or track this centre coordinate that this moves gesture, this moves this centre coordinate of gesture then to detect tracking again.
10. the method for gesture identification according to claim 1 and tracking is characterized in that be set at 1 second this actuation time.
11. the method for gesture identification according to claim 1 and tracking is characterized in that, this moves confirms that gesture is defined as a V-shape of forefinger and middle finger formation.
12. the method for gesture identification according to claim 1 and tracking is characterized in that, this moves gesture and is defined as the fist shape that the five fingers are clenched fist and formed.
13. the method for gesture identification according to claim 1 and tracking is characterized in that, this order gesture is defined as one 1 word shapes that forefinger forms.
CN200810177689A 2008-11-24 2008-11-24 Gesture Recognition and Tracking Method Pending CN101739122A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN200810177689A CN101739122A (en) 2008-11-24 2008-11-24 Gesture Recognition and Tracking Method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN200810177689A CN101739122A (en) 2008-11-24 2008-11-24 Gesture Recognition and Tracking Method

Publications (1)

Publication Number Publication Date
CN101739122A true CN101739122A (en) 2010-06-16

Family

ID=42462678

Family Applications (1)

Application Number Title Priority Date Filing Date
CN200810177689A Pending CN101739122A (en) 2008-11-24 2008-11-24 Gesture Recognition and Tracking Method

Country Status (1)

Country Link
CN (1) CN101739122A (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102452591A (en) * 2010-10-19 2012-05-16 由田新技股份有限公司 Elevator control system
CN103019389A (en) * 2013-01-12 2013-04-03 福建华映显示科技有限公司 Gesture recognition system and gesture recognition method
CN103699212A (en) * 2012-09-27 2014-04-02 纬创资通股份有限公司 Interactive system and movement detection method
CN103778405A (en) * 2012-10-17 2014-05-07 原相科技股份有限公司 Gesture recognition method by natural image
CN104461276A (en) * 2013-09-25 2015-03-25 联想(北京)有限公司 Switching method and information processing equipment
CN104508680A (en) * 2012-08-03 2015-04-08 科智库公司 Improved Video Tracking
CN107678551A (en) * 2017-10-19 2018-02-09 京东方科技集团股份有限公司 Gesture identification method and device, electronic equipment
CN112850406A (en) * 2015-04-03 2021-05-28 奥的斯电梯公司 Traffic list generation for passenger transport
CN115167675A (en) * 2011-09-19 2022-10-11 视力移动技术有限公司 Augmented reality device

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102452591A (en) * 2010-10-19 2012-05-16 由田新技股份有限公司 Elevator control system
CN115167675A (en) * 2011-09-19 2022-10-11 视力移动技术有限公司 Augmented reality device
CN104508680B (en) * 2012-08-03 2018-08-17 科智库公司 Improved video signal is tracked
CN104508680A (en) * 2012-08-03 2015-04-08 科智库公司 Improved Video Tracking
CN103699212B (en) * 2012-09-27 2016-08-31 纬创资通股份有限公司 interactive system and movement detection method
CN103699212A (en) * 2012-09-27 2014-04-02 纬创资通股份有限公司 Interactive system and movement detection method
CN107256089B (en) * 2012-10-17 2020-07-03 原相科技股份有限公司 Gesture recognition method by natural image
CN103778405B (en) * 2012-10-17 2017-07-04 原相科技股份有限公司 Gesture recognition method based on natural images
CN107256089A (en) * 2012-10-17 2017-10-17 原相科技股份有限公司 Gesture recognition method by natural image
CN103778405A (en) * 2012-10-17 2014-05-07 原相科技股份有限公司 Gesture recognition method by natural image
CN103019389A (en) * 2013-01-12 2013-04-03 福建华映显示科技有限公司 Gesture recognition system and gesture recognition method
CN104461276A (en) * 2013-09-25 2015-03-25 联想(北京)有限公司 Switching method and information processing equipment
CN104461276B (en) * 2013-09-25 2019-02-05 联想(北京)有限公司 A kind of switching method and information processing equipment
CN112850406A (en) * 2015-04-03 2021-05-28 奥的斯电梯公司 Traffic list generation for passenger transport
CN107678551A (en) * 2017-10-19 2018-02-09 京东方科技集团股份有限公司 Gesture identification method and device, electronic equipment
CN107678551B (en) * 2017-10-19 2021-12-28 京东方科技集团股份有限公司 Gesture recognition method and device and electronic equipment
US11402918B2 (en) 2017-10-19 2022-08-02 Boe Technology Group Co., Ltd. Method for controlling terminal apparatus, apparatus for controlling terminal apparatus, and computer-program product

Similar Documents

Publication Publication Date Title
US8270670B2 (en) Method for recognizing and tracing gesture
CN101739122A (en) Gesture Recognition and Tracking Method
US8290210B2 (en) Method and system for gesture recognition
TW201019241A (en) Method for identifying and tracing gesture
Le et al. Investigating the feasibility of finger identification on capacitive touchscreens using deep learning
US20200097091A1 (en) Method and Apparatus of Interactive Display Based on Gesture Recognition
CN103809733B (en) Human-computer interaction system and method
US20130120250A1 (en) Gesture recognition system and method
TWI471815B (en) Gesture recognition device and method
Ahuja et al. TouchPose: hand pose prediction, depth estimation, and touch classification from capacitive images
US20140327611A1 (en) Information processing apparatus and method, and program
US20150153832A1 (en) Visual feedback by identifying anatomical features of a hand
JP6651388B2 (en) Gesture modeling device, gesture modeling method, program for gesture modeling system, and gesture modeling system
CN104793738A (en) Non-contact type computer operating method based on Leap Motion
CN106598422B (en) hybrid control method, control system and electronic equipment
EP2204760A1 (en) Method for recognizing and tracing gesture
Mishra et al. Virtual mouse input control using hand gestures
JP4965590B2 (en) How to recognize and track gestures
US20220050528A1 (en) Electronic device for simulating a mouse
US20170139545A1 (en) Information processing apparatus, information processing method, and program
KR101171239B1 (en) Non-touch data input and operating method using image processing
KR102861897B1 (en) Apparatus and Method for Recognizing Motion
Dudile et al. AI Driven-Virtual Mouse Using Hand Gestures
Babu et al. Touchless User Interface for Sketching Using Hand Gesture Recognition
TW201537389A (en) Gesture recognition method for motion sensing detector

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C02 Deemed withdrawal of patent application after publication (patent law 2001)
WD01 Invention patent application deemed withdrawn after publication

Open date: 20100616