/
bert-base-multilingual-cased-cdo-wiki_pfeiffer.yaml
83 lines (67 loc) 路 3 KB
/
bert-base-multilingual-cased-cdo-wiki_pfeiffer.yaml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
# Adapter-Hub adapter entry
# Defines a single adapter entry in Adapter-Hub
# --------------------
# The name of the author(s) of this adapter.
author: "Jonas Pfeiffer"
# A bibtex citation of the work related to this adapter.
citation: |
@article{pfeiffer20madx,
title={{MAD-X}: An {A}dapter-based {F}ramework for {M}ulti-task {C}ross-lingual {T}ransfer},
author={Pfeiffer, Jonas and Vuli\'{c}, Ivan and Gurevych, Iryna and Ruder, Sebastian},
journal={arXiv preprint},
year={2020},
url={https://arxiv.org/pdf/2005.00052.pdf},
}
# The string identifier of the adapter architecture (must be available in architecture).
# Describes the adapter architecture used by this adapter
config: # TODO: REQUIRED
# The name of the adapter config used by this adapter (a short name available in the `architectures` folder).
# Example: pfeiffer
using: pfeiffer
# Overrides the default activation function of the specified adapter architecture.
# Example: tanh
non_linearity: gelu
# Overrides the default reduction factor of the specified adapter architecture
# Example: 64
reduction_factor: 2
# The version to be downloaded if no version is explicitly stated.
default_version: "nd"
# A short description of this adapter.
description: |
Pfeiffer Adapter trained with Masked Language Modelling on Min Dong Wikipedia Articles for 50k steps and a batch size of 64.
# A contact email of the author(s).
email: "pfeiffer@ukp.informatik.tu-darmstadt"
# A list of different versions of this adapter available for download.
files:
- sha1: "c4deba6321c5c7e721abb83e83754bac2373e51c"
sha256: "df43c8ebd986847e3e945b27405b492c7464c49d8403772b1500a8fc84c286b9"
# Download URL pointing to a zip folder containing the adapter module.
url: "https://public.ukp.informatik.tu-darmstadt.de/AdapterHub/text_lang/cdo/bert-base-multilingual-cased/pfeiffer/cdo_pfeiffer_gelu_nd.zip"
version: "nd"
- sha1: "a5a7a3166275f4b9d4ac7fccc18f85e8f87e73ba"
sha256: "35cbfef64fe13867dfcb85703b2d6afda2d13cc23ee003fc3ffd280315c6b5e9"
# Download URL pointing to a zip folder containing the adapter module.
url: "https://public.ukp.informatik.tu-darmstadt.de/AdapterHub/text_lang/cdo/bert-base-multilingual-cased/pfeiffer/cdo_pfeiffer_gelu.zip"
version: "wd"
# - ...
# A GitHub handle associated with the author(s).
github: "jopfeiff"
# The hidden size of the model
hidden_size: 768
# The string identifier of the pre-trained model (by which it is identified at Huggingface).
# Example: bert-base-uncased
model_name: "bert-base-multilingual-cased"
# The model type.
# Example: bert
model_type: bert
# The string identifier of the subtask this adapter belongs to.
subtask: wiki
# The string identifier of the task this adapter belongs to.
task: cdo
# A Twitter handle associated with the author(s).
twitter: "@PfeiffJo"
# The type of adapter (one of the options available in `adapter_type`.
type: "text_lang"
# A URL providing more information on this adapter/ the authors/ the organization.
url: "https://pfeiffer.ai"
prediction_head: false