Ex Parte HARVILL et al - Page 3



          Appeal No. 2003-1121                                                        
          Application 09/189,179                                                      

                                       OPINION                                        
               We reverse the aforementioned rejections.  We need to                  
          address only the broadest independent claim, i.e., claim 17.  The           
          examiner does not rely upon Leysieffer and King for any                     
          disclosure that remedies the deficiency in Fisher and Fallacaro             
          as to the broadest claim.                                                   
               Fisher discloses a system for visceral interaction of a user           
          with a surrounding virtual environment through gesture technology           
          (page 4).  Fisher discloses (pages 4-6):                                    
               For tactile interaction with the displayed three                       
               dimensional environment, the user wears lightweight                    
               glove-like devices that transmit data-records of arm,                  
               hand and finger shape and position to a host computer.                 
               The gloves are instrumented with flex-sensing devices                  
               at each finger joint, between fingers and across the                   
               palm of the hand (see Fig. 7).  Motion tracking sensors                
               like that described for tracking head motion are                       
               mounted on each glove to transmit position and                         
               orientation of the hands and arms to the host system                   
               (see Fig. 8).  One application of this technology is to                
               provide a three-dimensional cursor in the displayed                    
               environment.  And, in coordination with connected                      
               speech recognition technology, the hand and arm gesture                
               information is used to effect indicated gestures in the                
               synthesized or remote environment (e.g. control of                     
               robotic arms and end-effectors, and associated control                 
               of auxiliary camera positions).  Current                               
               implementations of this research include a three-                      
               dimensional graphic database of an articulated hand                    
               that, in the display environment, is spatially                         
               correspondent with the viewer’s real hand and is                       
               directly controlled by the instrumented glove device                   
               (see Fig. 9).  With this capability, the operator can                  
               pick-up and manipulate virtual objects that appear in                  
                                          3                                           




Page:  Previous  1  2  3  4  5  6  7  Next 

Last modified: November 3, 2007