Update README.md
Browse files
README.md
CHANGED
|
@@ -27,27 +27,20 @@ More information needed
|
|
| 27 |
|
| 28 |
## Intended uses & limitations
|
| 29 |
|
| 30 |
-
|
| 31 |
-
|
| 32 |
-
## Training and evaluation data
|
| 33 |
-
|
| 34 |
-
More information needed
|
| 35 |
|
| 36 |
## Training procedure
|
| 37 |
|
| 38 |
-
|
| 39 |
feature annotation). To make sure that, given the same report, the model would produce a specific HVDC feature, we simply append at
|
| 40 |
-
the beginning of each report a string of the form ``HVDC-Feature
|
| 41 |
|
| 42 |
-
After this procedure to the original dataset (\~1.8K) we obtain approximately 6.6K items.
|
| 43 |
-
Activities, Emotion, Friendliness, Misfortune, and Good Fortune.
|
| 44 |
-
this would have excluded Good Fortune (see Figure \ref{fig:train_set_prefix_dist}). We include either way the feature to control for
|
| 45 |
-
memorisation and counterbalance the Misfortune feature.
|
| 46 |
-
In the present study, we focused on a subset of six HVDC features: Characters, Activities, Emotion, Friendliness, Misfortune, and Good Fortune.
|
| 47 |
This selection was made to exclude features that represented less than 10\% of the total instances. Notably, Good Fortune would have been excluded under this
|
| 48 |
-
criterion
|
| 49 |
memorisation effects and to provide a counterbalance to the Misfortune feature. After filtering out instances whose annotation
|
| 50 |
-
feature is not one of the six selected features, we are left with \~5.3K
|
| 51 |
dream reports. We then generate a random split of 80\%-20\% for the training (i.e., 4,311 reports) and testing (i.e. 1,078 reports) sets.
|
| 52 |
|
| 53 |
### Training
|
|
|
|
| 27 |
|
| 28 |
## Intended uses & limitations
|
| 29 |
|
| 30 |
+
This model is designed for research purposes. See the disclaimer for more details.
|
|
|
|
|
|
|
|
|
|
|
|
|
| 31 |
|
| 32 |
## Training procedure
|
| 33 |
|
| 34 |
+
The overall idea of our approach is to disentangle each dream report from its annotation as a whole and to create an augmented set of (dream report; single
|
| 35 |
feature annotation). To make sure that, given the same report, the model would produce a specific HVDC feature, we simply append at
|
| 36 |
+
the beginning of each report a string of the form ``HVDC-Feature:'', in a manner that closely mimics T5 task-specific prefix fine-tuning.
|
| 37 |
|
| 38 |
+
After this procedure to the original dataset (\~1.8K) we obtain approximately 6.6K items. In the present study, we focused on a subset of six HVDC features:
|
| 39 |
+
Characters, Activities, Emotion, Friendliness, Misfortune, and Good Fortune.
|
|
|
|
|
|
|
|
|
|
| 40 |
This selection was made to exclude features that represented less than 10\% of the total instances. Notably, Good Fortune would have been excluded under this
|
| 41 |
+
criterion, but we intentionally retained this feature to control against potential
|
| 42 |
memorisation effects and to provide a counterbalance to the Misfortune feature. After filtering out instances whose annotation
|
| 43 |
+
feature is not one of the six selected features, we are left with \~5.3K
|
| 44 |
dream reports. We then generate a random split of 80\%-20\% for the training (i.e., 4,311 reports) and testing (i.e. 1,078 reports) sets.
|
| 45 |
|
| 46 |
### Training
|