String Kernels, Fisher Kernels and Finite State Automata

Abstract

In this paper we show how the generation of documents can be thought of as a k-stage Markov process, which leads to a Fisher ker(cid:173) nel from which the n-gram and string kernels can be re-constructed. The Fisher kernel view gives a more flexible insight into the string kernel and suggests how it can be parametrised in a way that re(cid:173) flects the statistics of the training corpus. Furthermore, the prob(cid:173) abilistic modelling approach suggests extending the Markov pro(cid:173) cess to consider sub-sequences of varying length, rather than the standard fixed-length approach used in the string kernel. We give a procedure for determining which sub-sequences are informative features and hence generate a Finite State Machine model, which can again be used to obtain a Fisher kernel. By adjusting the parametrisation we can also influence the weighting received by the features . In this way we are able to obtain a logarithmic weighting in a Fisher kernel. Finally, experiments are reported comparing the different kernels using the standard Bag of Words kernel as a baseline.

Cite

Text

Saunders et al. "String Kernels, Fisher Kernels and Finite State Automata." Neural Information Processing Systems, 2002.

Markdown

[Saunders et al. "String Kernels, Fisher Kernels and Finite State Automata." Neural Information Processing Systems, 2002.](https://mlanthology.org/neurips/2002/saunders2002neurips-string/)

BibTeX

@inproceedings{saunders2002neurips-string,
  title     = {{String Kernels, Fisher Kernels and Finite State Automata}},
  author    = {Saunders, Craig and Vinokourov, Alexei and Shawe-taylor, John S.},
  booktitle = {Neural Information Processing Systems},
  year      = {2002},
  pages     = {649-656},
  url       = {https://mlanthology.org/neurips/2002/saunders2002neurips-string/}
}