Skip to content
Switch branches/tags

Latest commit


Git stats


Failed to load latest commit information.
Latest commit message
Commit time

High-Performance Tucker Factorization on Heterogeneous Platforms (GTA)


High-Performance Tucker Factorization on Heterogeneous Platforms
Sejoon Oh, Namyong Park, Jungi Jang, Lee Sael, and U Kang
IEEE Transactions on Parallel and Distributed Systems (TPDS), 2019
[Paper] [Supplementary Material] [Offical Page]


GTA requires OpenMP (2.0+ version) and OpenCL (1.2+ version) libraries! (if you use gcc/g++ compiler, OpenMP is installed by default)

"make" command will create a single executable file, which is "GTA".

Please review the below requirements before you run the code.

1. Input tensor must follow tab- or space-separated format (see the demo tensor).
2. It is highly recommended to modify the pre-defined size of tmp and tmp2 in kernel files (src/ and src/
3. For tmp size (must be larger than the tensor rank), 10~25 is recommended.
4. For tmp2 size (must be larger than the tensor order), 3~10 is recommended.

The executable file takes five arguments, which are the path of input tensor file, path of directory for storing results, tensor rank (10~25 are recommended for Tucker), local size (64, 256, 512, and 1024 are recommended), number of GPUs (depends on your machine), binary number indicating POTF (1) or FOTF (0). The arguments MUST BE valid and in the above order.

	ex) ./GTA input.txt result/ 10 256 1 1

If you put the command properly, GTA will write all values of factor matrices and a core tensor in the result directory set by an argument. (PLEASE MAKE SURE THAT YOU HAVE A WRITE PERMISSION TO THE RESULT DIRECTORY!).

	ex) result/FACTOR1, result/CORETENSOR

We note that input tensors must follow base-1 indexing and outputs are based on base-0 indexing.


To run the demo, please follow the following procedure. Sample tensor is created as 100x100x100 size with 1,000 observable entries.

1. Type "make demo"
2. Check "result" directory for the demo factorization results


Name Structure Size Number of Nonzeros Download
Netflix (User, Movie, Year-month; Rating) 480K × 18K × 74 100M DOWN
MovieLens (User, Movie, Year, Hour; Rating) 138K × 27K × 21 × 24 20M DOWN
DBLP (Author, Conference, Year; Count) 418K × 4K × 50 1.3M DOWN
Facebook (User 1, User 2, Date; Friendship) 64K × 64K × 870 1.5M DOWN
Synthetic Synthetic random tensors 10K × 10K × 10K × 10K ~10M

Orthogonalization of Factor Matrices

You can apply QR decompositions to output factor matrices according to the main paper using MATLAB or other languages. Notice that current version of GTA does not orthogonalize factor matrices by default.


No releases published


No packages published