Appeal No. 2000-0189 Page 4 Application No. 08/864,460 Holmes teaches generating a continuous distribution probability density HMM from a quantized vector series for training and recognition: “A more widely used method for coping with the fact that particular sets of finely quantized feature values will occur only very rarely is to represent the distribution of feature vectors by some simple parametric model, and to use the calculated probabilities from this model to supply the probability distributions in the training and recognition processes. The Baum-Welch re-estimation must then be used to optimize the parameters of the feature distribution model, rather than the probabilities of the particular feature vectors" (p. 143). Said computation of optimum parameters of the feature distribution model (for each state, tacitly understood) is just the recited calculation of the incidence of the labels in each state, from the HMM state likelihood functions described by said parameters (claim 3), determined from the training vectors. Holmes also teaches clustering and using nearest-neighbor templates representing the average properties in each cluster (p. 125), and vector quantizing training (and test) patterns into a label series of clusters to which they belong ("It is possible to make a useful approximation to the feature vectors that actually occur by choosing only a small subset (perhaps about 100) of feature vectors, and replacing each measured vector by the one in the subset that is `nearest' to it according to a suitable distance metric. This process is known as vector quantization", p. 142, emphasis in original). As discussed above, since the Specification does not teach a two-step quantization, the examiner has interpreted the recited "vectors so quantized" as a reference to the inherent quantization involved in the measurement of continuous data.Page: Previous 1 2 3 4 5 6 7 8 9 NextLast modified: November 3, 2007