Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add coca trained (#307) #308

Merged
merged 34 commits into from
Jan 29, 2023
Merged

Add coca trained (#307) #308

merged 34 commits into from
Jan 29, 2023

Commits on Dec 20, 2022

  1. Add coca trained (#307)

    * initial setup
    
    * add coca loss
    
    * remove loss from the model
    
    * fix loss
    
    * add underscores
    
    * name changes
    
    * add cross attention to Residual and CustomResidual
    
    * fix if
    
    * ädd transformer 'decoder'
    
    * minor fix
    
    * looks better
    
    * initlize coca model structure
    
    * clean
    
    * typo and format
    
    * checkpoint signature
    
    * adjust multimodal decoder and add CoCaTransformer
    
    * keep older logic
    
    * remove chunk
    
    * typo
    
    * fix
    
    * make chunk dim explicit
    
    * adjust cfg names
    
    * add attentionalpooling
    
    * add attentional pooling to coca
    
    * small change
    
    * add cocatransformer variants and AttentionPooling
    
    * remoive older attention pooler
    
    * adapt embed text to coca text transformer
    
    * rm coca layers
    
    * rename and remove useless CoCa models
    
    * make attentionpooler pooler only
    
    * refactor for one transformer only
    
    * coca forward works
    
    * separatae context and n_queries
    
    * add inital coca_base config
    
    * remove config
    
    * small loss change
    
    * init training file
    
    * make variable order right
    
    * remove print
    
    * uniform names
    
    * renaming
    
    * add coca funcs to init
    
    * add coca config and exclude from testing
    
    * add and comment simple test (no trained model)
    
    * add L2 norm
    
    * make L2 same as in clip
    
    * remove unused temperature
    
    * type
    
    * clean
    
    * fix config
    
    * make rename and move cfg
    
    * rename
    
    * temptative add coca to factory
    
    * fix config
    
    * update config
    
    * embed contrastive cls token in model
    
    * remove unused arg
    
    * import create_loss
    
    * make factory accept coca
    
    * make caption loss distributed
    
    * make loss customizable
    
    * pass loss trhough training_epoch
    
    * add coca specific params to params
    
    * removed decoder unused parameters
    
    * remove unused attributes
    
    * adjust coca_config
    
    * fix config and remove unused parameters
    
    * remove comment
    
    * remove more comments
    
    * rename attention pooler
    
    * rename TransformerDecoder
    
    * make AttentionalPooler clearer
    
    * add local loss logic to cocaloss
    
    * only create loss if train in data
    
    * remove wrong file
    
    * fix attentional pooler call
    
    * not ready for testing
    
    * really not ready for testing
    
    * eof lien
    
    * uniform names
    
    * add possible generative loss to evaluate
    
    * change _build function names
    
    * remove wrong import
    
    * remove local_loss from captioning loss
    
    * indexing error
    
    * finish renaming
    
    * adjust configs
    
    * add training test for coca
    
    * simplify captioning loss
    
    * remove hf
    
    * fix evaluate and loss
    
    * remove print
    
    * move projection
    
    * add coca vit 32 config
    
    * test on new config
    
    * adjust coca_base config
    
    * remove coca from test_inference
    
    * maybe fix regression test
    
    * make logits and labels contiguous
    
    * simpler logic
    
    * make contiguous after transpose
    
    * last test
    
    * try fix loss
    
    * CoCa PR: loss fix + rename file
    
    * wait for feedback on this
    
    * cleanup
    
    * CoCa PR: add set_grad_checkpointing + fix checkpoint API
    
    * CoCa PR: fix eval (which uses encode_x instead of forward)
    
    * move making space for CLS token into encode_text
    
    * rever zs changes + fix
    
    Co-authored-by: gpucce <g.puccetti92@gmail.com>
    Co-authored-by: gpucce <g.puccetti@gmail.com>
    Co-authored-by: iejmac <iejmac@ip-172-31-44-155.ec2.internal>
    4 people committed Dec 20, 2022
    Configuration menu
    Copy the full SHA
    1b86601 View commit details
    Browse the repository at this point in the history

Commits on Dec 21, 2022

  1. Add coca to CI

    rom1504 committed Dec 21, 2022
    Configuration menu
    Copy the full SHA
    29fa332 View commit details
    Browse the repository at this point in the history
  2. Add coca to CI pr

    rom1504 committed Dec 21, 2022
    Configuration menu
    Copy the full SHA
    911c737 View commit details
    Browse the repository at this point in the history
  3. simplify encode_iamge (#313)

    Co-authored-by: Romain Beaumont <romain.rom1@gmail.com>
    gpucce and rom1504 committed Dec 21, 2022
    Configuration menu
    Copy the full SHA
    b4881bc View commit details
    Browse the repository at this point in the history
  4. Add cls mask (#312)

    * buil_cls_mask
    
    * add cls_mask to encode_text
    
    * add model properties
    
    Co-authored-by: Romain Beaumont <romain.rom1@gmail.com>
    Co-authored-by: gpucce <g.puccetti@gmail.com>
    3 people committed Dec 21, 2022
    Configuration menu
    Copy the full SHA
    50bc599 View commit details
    Browse the repository at this point in the history

Commits on Dec 22, 2022

  1. Ignore pad tokens in captioning loss (#316)

    * add ignore_index
    
    * just need to pick right index
    
    Co-authored-by: gpucce <g.puccetti@gmail.com>
    gpucce and gpucce committed Dec 22, 2022
    Configuration menu
    Copy the full SHA
    279e088 View commit details
    Browse the repository at this point in the history
  2. add generate to coca model (#314)

    * add initial generative support
    
    * make generation context_length independend
    
    * remove kwargs
    
    * last positional embeddings for CLS
    
    * typo
    
    * fix mask len
    
    * add comment
    
    * remove unused args
    
    * simpler logic for input shorter than context length
    
    Co-authored-by: gpucce <g.puccetti@gmail.com>
    gpucce and gpucce committed Dec 22, 2022
    Configuration menu
    Copy the full SHA
    dee1ea5 View commit details
    Browse the repository at this point in the history

Commits on Jan 6, 2023

  1. use TextEncoder in coca encode_image (#321)

    * use self.text in encode image
    
    * unused var
    
    * rever aAtention and CustoResidualAttentionBlock
    
    * remove whiteline
    
    * add dict output
    
    * bintegrate self.text attributes
    
    * HF compatibility
    
    * better config and minor fixes
    
    * clean
    
    * remove eembed_cls option from HF
    
    * use cls_token_position
    
    * fix cls masking
    
    * resize labels
    
    * text -> self.text
    
    * split loss logging
    
    * add total loss
    
    * minor logs formatting
    
    * fix generate
    
    * simpler logic
    
    * disentangle proj for HF too
    
    * adjust config
    
    * only norm cls
    
    * move attn_pool to VisionTransformer
    
    * adjust coca_base config
    
    * fix grad checkpointing in MultimodalTransformer
    
    Co-authored-by: gpucce <g.puccetti@gmail.com>
    Co-authored-by: iejMac <kilianmaciej6@gmail.com>
    3 people committed Jan 6, 2023
    Configuration menu
    Copy the full SHA
    30a73d4 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    f616050 View commit details
    Browse the repository at this point in the history

Commits on Jan 9, 2023

  1. Configuration menu
    Copy the full SHA
    061482b View commit details
    Browse the repository at this point in the history

Commits on Jan 21, 2023

  1. Add tests bis (#355)

    * make jit compilable
    
    * redundant annotation
    
    * less tests
    
    * less annotations
    
    * even less annotations
    
    * fix name check in ci
    
    * some annotations back
    
    * make it simpler
    
    * make hf simpler too
    
    * better jit support with tests
    
    * remove extra line
    
    * add customtextclip
    
    * more jit tests
    
    * missing assert
    
    * add eval
    
    * typo
    
    * rever forward changes
    
    * clean coca model
    
    * more cleaning
    
    * last cleaning
    gpucce committed Jan 21, 2023
    Configuration menu
    Copy the full SHA
    d0bd09e View commit details
    Browse the repository at this point in the history
  2. Merge branch 'main' into coca

    rom1504 committed Jan 21, 2023
    Configuration menu
    Copy the full SHA
    ef80b7b View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    2ab47b7 View commit details
    Browse the repository at this point in the history

Commits on Jan 22, 2023

  1. add README (#365)

    * add README
    
    * multimodal_cfg info
    
    * multimodal
    iejMac committed Jan 22, 2023
    Configuration menu
    Copy the full SHA
    c0e5950 View commit details
    Browse the repository at this point in the history
  2. Merge branch 'main' into coca

    rom1504 committed Jan 22, 2023
    Configuration menu
    Copy the full SHA
    9ab881e View commit details
    Browse the repository at this point in the history
  3. remove output_dict argument (#368)

    * remove output_dict argument
    
    * cleaner
    gpucce committed Jan 22, 2023
    Configuration menu
    Copy the full SHA
    3f5b0fb View commit details
    Browse the repository at this point in the history
  4. do same thing for _encode_image (#366)

    * do same thing for _encode_image
    
    * encoder
    
    * try this
    
    * adjust inference tests
    
    * fix syntax
    
    * True not None
    
    * dumb
    iejMac committed Jan 22, 2023
    Configuration menu
    Copy the full SHA
    de343fb View commit details
    Browse the repository at this point in the history

Commits on Jan 23, 2023

  1. Configuration menu
    Copy the full SHA
    88aa6ce View commit details
    Browse the repository at this point in the history

Commits on Jan 24, 2023

  1. Revert "do same thing for _encode_image (#366)"

    This reverts commit de343fb.
    gpucce authored and rwightman committed Jan 24, 2023
    Configuration menu
    Copy the full SHA
    3b66f37 View commit details
    Browse the repository at this point in the history
  2. refactor

    gpucce authored and rwightman committed Jan 24, 2023
    Configuration menu
    Copy the full SHA
    cdb91dd View commit details
    Browse the repository at this point in the history
  3. white space

    gpucce authored and rwightman committed Jan 24, 2023
    Configuration menu
    Copy the full SHA
    58eb5bd View commit details
    Browse the repository at this point in the history
  4. remove extra layer norm

    gpucce authored and rwightman committed Jan 24, 2023
    Configuration menu
    Copy the full SHA
    cbd66ed View commit details
    Browse the repository at this point in the history
  5. move to_logits into decoder

    gpucce authored and rwightman committed Jan 24, 2023
    Configuration menu
    Copy the full SHA
    bf6ef3e View commit details
    Browse the repository at this point in the history
  6. leave for later

    gpucce authored and rwightman committed Jan 24, 2023
    Configuration menu
    Copy the full SHA
    03dfeab View commit details
    Browse the repository at this point in the history

Commits on Jan 25, 2023

  1. better torchscript

    gpucce authored and rwightman committed Jan 25, 2023
    Configuration menu
    Copy the full SHA
    15d6223 View commit details
    Browse the repository at this point in the history
  2. annotate hf too

    gpucce authored and rwightman committed Jan 25, 2023
    Configuration menu
    Copy the full SHA
    9beb0d4 View commit details
    Browse the repository at this point in the history

Commits on Jan 27, 2023

  1. Configuration menu
    Copy the full SHA
    fde2aee View commit details
    Browse the repository at this point in the history
  2. Merge branch 'main' into coca

    rom1504 committed Jan 27, 2023
    Configuration menu
    Copy the full SHA
    24e454d View commit details
    Browse the repository at this point in the history

Commits on Jan 28, 2023

  1. Remove dead LN code, refactor attn_pool conditional for more clarity,…

    … minor formatting tweaks
    rwightman committed Jan 28, 2023
    Configuration menu
    Copy the full SHA
    f7c566b View commit details
    Browse the repository at this point in the history
  2. latent_dim to embed_dim

    gpucce authored and rwightman committed Jan 28, 2023
    Configuration menu
    Copy the full SHA
    9533575 View commit details
    Browse the repository at this point in the history
  3. remove extra cfg

    gpucce authored and rwightman committed Jan 28, 2023
    Configuration menu
    Copy the full SHA
    f5e0c5a View commit details
    Browse the repository at this point in the history
  4. A bit more cleanup, keep context_length as context len, 'num_pos' to …

    …incl extra tokens. None type check for embed_cls instead of getattr
    rwightman committed Jan 28, 2023
    Configuration menu
    Copy the full SHA
    1ba2ab6 View commit details
    Browse the repository at this point in the history

Commits on Jan 29, 2023

  1. CoCa: add B/32 pretrained (#389)

    * add B/32 pretrained
    
    * fix
    
    * no capital
    
    * slash
    iejMac committed Jan 29, 2023
    Configuration menu
    Copy the full SHA
    f0847fa View commit details
    Browse the repository at this point in the history
  2. remove coca from ci.yml

    rom1504 committed Jan 29, 2023
    Configuration menu
    Copy the full SHA
    ba081d3 View commit details
    Browse the repository at this point in the history