We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
I am trying to train on google colab using the following code
from random import choice import numpy as np import pandas as pd import tensorflow as tf from tensorflow import keras from hyperlib.manifold.lorentz import Lorentz from hyperlib.manifold.poincare import Poincare from hyperlib.models.pehr import HierarchicalEmbeddings def load_wordnet_data(file, negatives=20): noun_closure = pd.read_csv(file) noun_closure_np = noun_closure[["id1","id2"]].values edges = set() for i, j in noun_closure_np: edges.add((i,j)) unique_nouns = list(set( noun_closure["id1"].tolist()+noun_closure["id2"].tolist() )) noun_closure["neg_pairs"] = noun_closure["id1"].apply(get_neg_pairs, args=(edges, unique_nouns, 20,)) return noun_closure, unique_nouns def get_neg_pairs(noun, edges, unique_nouns, negatives=20): neg_list = [] while len(neg_list) < negatives: neg_noun = choice(unique_nouns) if neg_noun != noun \ and not neg_noun in neg_list \ and not ((noun, neg_noun) in edges or (neg_noun, noun) in edges): neg_list.append(neg_noun) return neg_list # Make training dataset noun_closure, unique_nouns = load_wordnet_data("mammal_closure.csv", negatives=15) noun_closure_dataset = noun_closure[["id1","id2"]].values batch_size = 16 train_dataset = tf.data.Dataset.from_tensor_slices( (noun_closure_dataset, noun_closure["neg_pairs"].tolist())) train_dataset = train_dataset.shuffle(buffer_size=1024).batch(batch_size) # Create model model = HierarchicalEmbeddings(vocab=unique_nouns, embedding_dim=10) sgd = keras.optimizers.SGD(learning_rate=1e-2, momentum=0.9) # Run custom training loop model.fit(train_dataset, sgd, epochs=20) embs = model.get_embeddings() M = Poincare() mammal = M.expmap0(model(tf.constant('dog.n.01')), c=1) dists = M.dist(mammal, embs, c=1.0) top = tf.math.top_k(-dists[:,0], k=20) for i in top.indices: print(unique_nouns[i],': ',-dists[i,0].numpy())
I see that the GPU is not being used when I inspect the GPU usage. Kindly help
The text was updated successfully, but these errors were encountered:
No branches or pull requests
I am trying to train on google colab using the following code
I see that the GPU is not being used when I inspect the GPU usage. Kindly help
The text was updated successfully, but these errors were encountered: