Learning via automaton minimization and matrix factorization
<p>This thesis focuses on weighted-automaton learning and minimization, matrix factorization, and the relationships between these areas.</p> <p>The first part of the thesis studies minimization and learning of weighted automata with weights in a field, focusing on tree automata as...
Main Author: | |
---|---|
Other Authors: | |
Format: | Thesis |
Published: |
2016
|
Subjects: |
_version_ | 1797083296901890048 |
---|---|
author | Marusic, I |
author2 | Worrell, J |
author_facet | Worrell, J Marusic, I |
author_sort | Marusic, I |
collection | OXFORD |
description | <p>This thesis focuses on weighted-automaton learning and minimization, matrix factorization, and the relationships between these areas.</p> <p>The first part of the thesis studies minimization and learning of weighted automata with weights in a field, focusing on tree automata as representations of distributions over trees. We give a minimization algorithm that runs in polynomial time assuming unit-cost arithmetic, and show that a polynomial bound in the Turing model would require a breakthrough in the complexity of polynomial identity testing. We also improve the complexity of minimizing weighted word automata. Secondly, we look at learning minimal weighted automata in both active and passive learning frameworks, analysing both the computational and query complexity. For active learning, we give a new algorithm for learning weighted tree automata in Angluin's exact learning model that efficiently processes DAG counterexamples, and show a complexity lower bound in terms of polynomial identity testing. For passive learning, we characterise the complexity of finding a minimal weighted automaton that is consistent with a set of observations. </p> <p>The second part of the thesis studies nonnegative matrix factorization (NMF), a powerful dimension-reduction and feature-extraction technique with numerous applications in machine learning and other scientific disciplines. Despite being an important technique in practice, there are a number of open questions about the theory of NMF---especially about its complexity and the existence of good heuristics. We answer a major open problem posed in 1993 by Cohen and Rothblum: the nonnegative ranks over the rationals and over the reals differ. We first tackle a variant of the problem, restricted NMF, which has applications in topic modelling and a geometric interpretation as the nested polytope problem. Lastly we show that NMF is polynomial-time reducible to automaton minimization, and then use our results on NMF to answer old open problems on minimizing labelled Markov chains and probabilistic automata.</p> |
first_indexed | 2024-03-07T01:39:45Z |
format | Thesis |
id | oxford-uuid:9672c380-6b26-48f7-8766-b60adaeec453 |
institution | University of Oxford |
last_indexed | 2024-03-07T01:39:45Z |
publishDate | 2016 |
record_format | dspace |
spelling | oxford-uuid:9672c380-6b26-48f7-8766-b60adaeec4532022-03-26T23:52:58ZLearning via automaton minimization and matrix factorizationThesishttp://purl.org/coar/resource_type/c_db06uuid:9672c380-6b26-48f7-8766-b60adaeec453Computer scienceORA Deposit2016Marusic, IWorrell, JKiefer, SBenedikt, M<p>This thesis focuses on weighted-automaton learning and minimization, matrix factorization, and the relationships between these areas.</p> <p>The first part of the thesis studies minimization and learning of weighted automata with weights in a field, focusing on tree automata as representations of distributions over trees. We give a minimization algorithm that runs in polynomial time assuming unit-cost arithmetic, and show that a polynomial bound in the Turing model would require a breakthrough in the complexity of polynomial identity testing. We also improve the complexity of minimizing weighted word automata. Secondly, we look at learning minimal weighted automata in both active and passive learning frameworks, analysing both the computational and query complexity. For active learning, we give a new algorithm for learning weighted tree automata in Angluin's exact learning model that efficiently processes DAG counterexamples, and show a complexity lower bound in terms of polynomial identity testing. For passive learning, we characterise the complexity of finding a minimal weighted automaton that is consistent with a set of observations. </p> <p>The second part of the thesis studies nonnegative matrix factorization (NMF), a powerful dimension-reduction and feature-extraction technique with numerous applications in machine learning and other scientific disciplines. Despite being an important technique in practice, there are a number of open questions about the theory of NMF---especially about its complexity and the existence of good heuristics. We answer a major open problem posed in 1993 by Cohen and Rothblum: the nonnegative ranks over the rationals and over the reals differ. We first tackle a variant of the problem, restricted NMF, which has applications in topic modelling and a geometric interpretation as the nested polytope problem. Lastly we show that NMF is polynomial-time reducible to automaton minimization, and then use our results on NMF to answer old open problems on minimizing labelled Markov chains and probabilistic automata.</p> |
spellingShingle | Computer science Marusic, I Learning via automaton minimization and matrix factorization |
title | Learning via automaton minimization and matrix factorization |
title_full | Learning via automaton minimization and matrix factorization |
title_fullStr | Learning via automaton minimization and matrix factorization |
title_full_unstemmed | Learning via automaton minimization and matrix factorization |
title_short | Learning via automaton minimization and matrix factorization |
title_sort | learning via automaton minimization and matrix factorization |
topic | Computer science |
work_keys_str_mv | AT marusici learningviaautomatonminimizationandmatrixfactorization |