Learning via automaton minimization and matrix factorization

<p>This thesis focuses on weighted-automaton learning and minimization, matrix factorization, and the relationships between these areas.</p> <p>The first part of the thesis studies minimization and learning of weighted automata with weights in a field, focusing on tree automata as...

Full description

Bibliographic Details
Main Author: Marusic, I
Other Authors: Worrell, J
Format: Thesis
Published: 2016
Subjects:
_version_ 1797083296901890048
author Marusic, I
author2 Worrell, J
author_facet Worrell, J
Marusic, I
author_sort Marusic, I
collection OXFORD
description <p>This thesis focuses on weighted-automaton learning and minimization, matrix factorization, and the relationships between these areas.</p> <p>The first part of the thesis studies minimization and learning of weighted automata with weights in a field, focusing on tree automata as representations of distributions over trees. We give a minimization algorithm that runs in polynomial time assuming unit-cost arithmetic, and show that a polynomial bound in the Turing model would require a breakthrough in the complexity of polynomial identity testing. We also improve the complexity of minimizing weighted word automata. Secondly, we look at learning minimal weighted automata in both active and passive learning frameworks, analysing both the computational and query complexity. For active learning, we give a new algorithm for learning weighted tree automata in Angluin's exact learning model that efficiently processes DAG counterexamples, and show a complexity lower bound in terms of polynomial identity testing. For passive learning, we characterise the complexity of finding a minimal weighted automaton that is consistent with a set of observations. </p> <p>The second part of the thesis studies nonnegative matrix factorization (NMF), a powerful dimension-reduction and feature-extraction technique with numerous applications in machine learning and other scientific disciplines. Despite being an important technique in practice, there are a number of open questions about the theory of NMF---especially about its complexity and the existence of good heuristics. We answer a major open problem posed in 1993 by Cohen and Rothblum: the nonnegative ranks over the rationals and over the reals differ. We first tackle a variant of the problem, restricted NMF, which has applications in topic modelling and a geometric interpretation as the nested polytope problem. Lastly we show that NMF is polynomial-time reducible to automaton minimization, and then use our results on NMF to answer old open problems on minimizing labelled Markov chains and probabilistic automata.</p>
first_indexed 2024-03-07T01:39:45Z
format Thesis
id oxford-uuid:9672c380-6b26-48f7-8766-b60adaeec453
institution University of Oxford
last_indexed 2024-03-07T01:39:45Z
publishDate 2016
record_format dspace
spelling oxford-uuid:9672c380-6b26-48f7-8766-b60adaeec4532022-03-26T23:52:58ZLearning via automaton minimization and matrix factorizationThesishttp://purl.org/coar/resource_type/c_db06uuid:9672c380-6b26-48f7-8766-b60adaeec453Computer scienceORA Deposit2016Marusic, IWorrell, JKiefer, SBenedikt, M<p>This thesis focuses on weighted-automaton learning and minimization, matrix factorization, and the relationships between these areas.</p> <p>The first part of the thesis studies minimization and learning of weighted automata with weights in a field, focusing on tree automata as representations of distributions over trees. We give a minimization algorithm that runs in polynomial time assuming unit-cost arithmetic, and show that a polynomial bound in the Turing model would require a breakthrough in the complexity of polynomial identity testing. We also improve the complexity of minimizing weighted word automata. Secondly, we look at learning minimal weighted automata in both active and passive learning frameworks, analysing both the computational and query complexity. For active learning, we give a new algorithm for learning weighted tree automata in Angluin's exact learning model that efficiently processes DAG counterexamples, and show a complexity lower bound in terms of polynomial identity testing. For passive learning, we characterise the complexity of finding a minimal weighted automaton that is consistent with a set of observations. </p> <p>The second part of the thesis studies nonnegative matrix factorization (NMF), a powerful dimension-reduction and feature-extraction technique with numerous applications in machine learning and other scientific disciplines. Despite being an important technique in practice, there are a number of open questions about the theory of NMF---especially about its complexity and the existence of good heuristics. We answer a major open problem posed in 1993 by Cohen and Rothblum: the nonnegative ranks over the rationals and over the reals differ. We first tackle a variant of the problem, restricted NMF, which has applications in topic modelling and a geometric interpretation as the nested polytope problem. Lastly we show that NMF is polynomial-time reducible to automaton minimization, and then use our results on NMF to answer old open problems on minimizing labelled Markov chains and probabilistic automata.</p>
spellingShingle Computer science
Marusic, I
Learning via automaton minimization and matrix factorization
title Learning via automaton minimization and matrix factorization
title_full Learning via automaton minimization and matrix factorization
title_fullStr Learning via automaton minimization and matrix factorization
title_full_unstemmed Learning via automaton minimization and matrix factorization
title_short Learning via automaton minimization and matrix factorization
title_sort learning via automaton minimization and matrix factorization
topic Computer science
work_keys_str_mv AT marusici learningviaautomatonminimizationandmatrixfactorization