posted on 2023-12-11, 05:13authored byMaxwell Clarke
<p><strong>Transformer models have been quickly taking over many tasks in the field of deep learning, due to their versatility while maintaining high performance. This thesis provides an introduction to transformer models, presents experiments with new ways of sampling data with them, and then applies them to the domain of hand motion modeling.</strong></p><p>Firstly, a comprehensive introduction to transformer models is given, including the attention operation, masking, architecture variants, and different pre-training tasks.</p><p>Secondly, an experiment is presented using a probabilistic transformer model on the MNIST dataset, which was trained so that it is capable of arbitrary-order sampling. The experiment compares different sampling orders, including some dynamic sampling order heuristics based on the entropy. The experiments find that such sampling orders introduce a statistical bias into the samples.</p><p>Lastly, the problem domain of hand motion modeling is introduced, and transformer models are trained to generate hand-motion sequences, via self-supervised learning on a motion-capture dataset. Both deterministic and probabilistic models are trained. The deterministic models can generate realistic-looking hand motions, but cannot be directed to generate specific motions. The probabilistic model performs poorly.</p>
History
Copyright Date
2023-12-11
Date of Award
2023-12-11
Publisher
Te Herenga Waka—Victoria University of Wellington
Rights License
CC BY-SA 4.0
Degree Discipline
Computer Science
Degree Grantor
Te Herenga Waka—Victoria University of Wellington
Degree Level
Masters
Degree Name
Master of Science
ANZSRC Socio-Economic Outcome code
220403 Artificial intelligence;
130205 Visual communication;
130603 Recreation and leisure activities (excl. sport and exercise)