Skip to content

Transformer Text AutoEncoder: An autoencoder is a type of artificial neural network used to learn efficient encodings of unlabeled data, the same is employed for textual data employing pre-trained models from the hugging-face library.

License

Notifications You must be signed in to change notification settings

AmanPriyanshu/Transformer-Text-AutoEncoder

Repository files navigation

Transformer-Text-AutoEncoder

Transformer Text AutoEncoder: An autoencoder is a type of artificial neural network used to learn efficient encodings of unlabeled data, the same is employed for textual data employing pre-trained models from the hugging-face library.

Installation:

pip install Transformer-Text-AutoEncoder

Execution:

Training

from Transformer_Text_AutoEncoder.AutoEncoder import TTAE

def read(path='./Transformer-Text-AutoEncoder/data/FinancialNews.txt'):
  with open(path, "r", encoding='utf-8', errors='ignore') as f:
    data = [i.strip() for i in f.readlines()]
  return data

sentences = read()
print(sentences[:3])
ttae = TTAE(sentences)
ttae.train(100, batch_size=8)

Predictions

predicted_sentence, encoded_vec = ttae.predict("According to Gran , the company has no plans to move all production to Russia , although that is where the company is growing.")

returns the predicted sentence as well as the encoded_vec.

Cite Work:

@inproceedings{ttae,
  title = {Transformer-Text-AutoEncoder},
  author = {Aman Priyanshu},
  year = {2022},
  publisher = {{GitHub}},
  url = {https://github.com/AmanPriyanshu/Transformer-Text-AutoEncoder/}
}

About

Transformer Text AutoEncoder: An autoencoder is a type of artificial neural network used to learn efficient encodings of unlabeled data, the same is employed for textual data employing pre-trained models from the hugging-face library.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages