With the fingertip detection in place, it's now a lot easier to recognize the "grab" gesture. The old version monitored the cluster size and if it changed quickly would then guess that the hand was closed. The new version uses the number of fingers that are detected in each frame to decide whether the hand is open or closed.
Plus the hand's shape is now transfered to the image view:
I'm thinking of implementing a simple event based API for this.