Skip to content

The goal of this project is to create multi-modal implementation of Transformer architecture in Swift.

Notifications You must be signed in to change notification settings

wojtekcz/language2motion

Repository files navigation

Language2motion

The goal of this project was to create multi-modal implementation of Transformer architecture in Swift for Tensorflow.

Also it was an attempt to answer the question if Swift for Tensorflow is ready for non-trivial work.

The use-case is based on a paper "Learning a bidirectional mapping between human whole-body motion and natural language using deep recurrent neural networks" by Matthias Plappert. He created a nice dataset of few thousand motions "The KIT Motion-Language Dataset (paper)", website.

The Motion2Language and Lang2motion Transformer-based models were implemented. Also some more sophisticated motion generation strategies were tried.

Modified Swift Transformer implementation by Andre Carrera was used.

Models

  • motion 2 language
    • Transformer from motion to annotation
  • language 2 motion
    • Transformer from annotation to motion

Dataset files

Motion player

Runtime env

About

The goal of this project is to create multi-modal implementation of Transformer architecture in Swift.

Topics

Resources

Stars

Watchers

Forks

Packages

No packages published