Skip to content

Latest commit

 

History

History
34 lines (20 loc) · 550 Bytes

README.md

File metadata and controls

34 lines (20 loc) · 550 Bytes

Multi-Modality

GeidiPrime

This is an extremely experimental Transformer architecture with Macaron like FFNs with local attention. Perhap's we can add the visual expert from Zeta and make it multi-modal!

Install

Usage

import torch
from geidi_prime.model import GeidiPrimeTransformer

model = GeidiPrimeTransformer(
    dim=4096,
    depth=6,
    heads=8,
    num_tokens=20000,
)

x = torch.randint(0, 20000, (1, 4096))

out = model(x)
print(out.shape)

License

MIT