Upload README.md with huggingface_hub
Browse files
README.md
CHANGED
|
@@ -21,10 +21,10 @@ Models are named as follows:
|
|
| 21 |
|
| 22 |
- `CodeBERTa` → `CodeBERTa-ft-coco-[1,2,5]e-05lr`
|
| 23 |
- e.g. `CodeBERTa-ft-coco-2e-05lr`, which was trained with learning rate of `2e-05`.
|
| 24 |
-
- `JonBERTa-head` → `JonBERTa-head-ft-
|
| 25 |
-
- e.g. `JonBERTa-head-ft-
|
| 26 |
-
- `JonBERTa-attn` → `JonBERTa-attn-ft-
|
| 27 |
-
- e.g. `JonBERTa-attn-ft-
|
| 28 |
|
| 29 |
Other hyperparameters may be found in the paper or the replication package (see below).
|
| 30 |
|
|
|
|
| 21 |
|
| 22 |
- `CodeBERTa` → `CodeBERTa-ft-coco-[1,2,5]e-05lr`
|
| 23 |
- e.g. `CodeBERTa-ft-coco-2e-05lr`, which was trained with learning rate of `2e-05`.
|
| 24 |
+
- `JonBERTa-head` → `JonBERTa-head-ft-[dense,proj,reinit]`
|
| 25 |
+
- e.g. `JonBERTa-head-ft-dense-proj`, where all have `2e-05` learning rate, but may differ in the head layer in which the telemetry features are introduced (either `head` or `proj`, with optional `reinit`ialisation of all its weights).
|
| 26 |
+
- `JonBERTa-attn` → `JonBERTa-attn-ft-[0,1,2,3,4,5]L`
|
| 27 |
+
- e.g. `JonBERTa-attn-ft-012L` , where all have `2e-05` learning rate, but may differ in the attention layer(s) in which the telemetry features are introduced (either `0`, `1`, `2`, `3`, `4`, or `5L`).
|
| 28 |
|
| 29 |
Other hyperparameters may be found in the paper or the replication package (see below).
|
| 30 |
|