Skip to content
master
Switch branches/tags
Code

Latest commit

 

Git stats

Files

Permalink
Failed to load latest commit information.
Type
Name
Latest commit message
Commit time
 
 
 
 
 
 
src
 
 
 
 
 
 
 
 
 
 

High-Performance Tucker Factorization on Heterogeneous Platforms (GTA)

Overview

High-Performance Tucker Factorization on Heterogeneous Platforms
Sejoon Oh, Namyong Park, Jungi Jang, Lee Sael, and U Kang
IEEE Transactions on Parallel and Distributed Systems (TPDS), 2019
[Paper] [Supplementary Material] [Offical Page]

Usage

GTA requires OpenMP (2.0+ version) and OpenCL (1.2+ version) libraries! (if you use gcc/g++ compiler, OpenMP is installed by default)

"make" command will create a single executable file, which is "GTA".

Please review the below requirements before you run the code.

1. Input tensor must follow tab- or space-separated format (see the demo tensor).
2. It is highly recommended to modify the pre-defined size of tmp and tmp2 in kernel files (src/GTA_GPU_Delta.cl and src/GTA_reconstruction.cl).
3. For tmp size (must be larger than the tensor rank), 10~25 is recommended.
4. For tmp2 size (must be larger than the tensor order), 3~10 is recommended.

The executable file takes five arguments, which are the path of input tensor file, path of directory for storing results, tensor rank (10~25 are recommended for Tucker), local size (64, 256, 512, and 1024 are recommended), number of GPUs (depends on your machine), binary number indicating POTF (1) or FOTF (0). The arguments MUST BE valid and in the above order.

	ex) ./GTA input.txt result/ 10 256 1 1

If you put the command properly, GTA will write all values of factor matrices and a core tensor in the result directory set by an argument. (PLEASE MAKE SURE THAT YOU HAVE A WRITE PERMISSION TO THE RESULT DIRECTORY!).

	ex) result/FACTOR1, result/CORETENSOR

We note that input tensors must follow base-1 indexing and outputs are based on base-0 indexing.

Demo

To run the demo, please follow the following procedure. Sample tensor is created as 100x100x100 size with 1,000 observable entries.

1. Type "make demo"
2. Check "result" directory for the demo factorization results

Dataset

Name Structure Size Number of Nonzeros Download
Netflix (User, Movie, Year-month; Rating) 480K × 18K × 74 100M DOWN
MovieLens (User, Movie, Year, Hour; Rating) 138K × 27K × 21 × 24 20M DOWN
DBLP (Author, Conference, Year; Count) 418K × 4K × 50 1.3M DOWN
Facebook (User 1, User 2, Date; Friendship) 64K × 64K × 870 1.5M DOWN
Synthetic Synthetic random tensors 10K × 10K × 10K × 10K ~10M

Orthogonalization of Factor Matrices

You can apply QR decompositions to output factor matrices according to the main paper using MATLAB or other languages. Notice that current version of GTA does not orthogonalize factor matrices by default.

Releases

No releases published

Packages

No packages published

Languages