lorenzoscottb commited on
Commit
58cd61e
·
verified ·
1 Parent(s): 2ab6b54

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -14
README.md CHANGED
@@ -27,27 +27,20 @@ More information needed
27
 
28
  ## Intended uses & limitations
29
 
30
- More information needed
31
-
32
- ## Training and evaluation data
33
-
34
- More information needed
35
 
36
  ## Training procedure
37
 
38
- the overall idea of our approach is to disentangle each dream report from its annotation as a whole and to create an augmented set of (dream report; single
39
  feature annotation). To make sure that, given the same report, the model would produce a specific HVDC feature, we simply append at
40
- the beginning of each report a string of the form ``HVDC-Feature :'', in a manner that closely mimics T5 task-specific prefix fine-tuning.
41
 
42
- After this procedure to the original dataset (\~1.8K) we obtain approximately 6.6K items. %For this work, we focused on six HVDC features, namely Characters,
43
- Activities, Emotion, Friendliness, Misfortune, and Good Fortune. We did so to exclude features that amounted to less than 10\% of the total instance. Indeed,
44
- this would have excluded Good Fortune (see Figure \ref{fig:train_set_prefix_dist}). We include either way the feature to control for
45
- memorisation and counterbalance the Misfortune feature.
46
- In the present study, we focused on a subset of six HVDC features: Characters, Activities, Emotion, Friendliness, Misfortune, and Good Fortune.
47
  This selection was made to exclude features that represented less than 10\% of the total instances. Notably, Good Fortune would have been excluded under this
48
- criterion (refer to Figure \ref{fig:train_set_prefix_dist}), but we intentionally retained this feature to control against potential
49
  memorisation effects and to provide a counterbalance to the Misfortune feature. After filtering out instances whose annotation
50
- feature is not one of the six selected features, we are left with \~5.3K %5389
51
  dream reports. We then generate a random split of 80\%-20\% for the training (i.e., 4,311 reports) and testing (i.e. 1,078 reports) sets.
52
 
53
  ### Training
 
27
 
28
  ## Intended uses & limitations
29
 
30
+ This model is designed for research purposes. See the disclaimer for more details.
 
 
 
 
31
 
32
  ## Training procedure
33
 
34
+ The overall idea of our approach is to disentangle each dream report from its annotation as a whole and to create an augmented set of (dream report; single
35
  feature annotation). To make sure that, given the same report, the model would produce a specific HVDC feature, we simply append at
36
+ the beginning of each report a string of the form ``HVDC-Feature:'', in a manner that closely mimics T5 task-specific prefix fine-tuning.
37
 
38
+ After this procedure to the original dataset (\~1.8K) we obtain approximately 6.6K items. In the present study, we focused on a subset of six HVDC features:
39
+ Characters, Activities, Emotion, Friendliness, Misfortune, and Good Fortune.
 
 
 
40
  This selection was made to exclude features that represented less than 10\% of the total instances. Notably, Good Fortune would have been excluded under this
41
+ criterion, but we intentionally retained this feature to control against potential
42
  memorisation effects and to provide a counterbalance to the Misfortune feature. After filtering out instances whose annotation
43
+ feature is not one of the six selected features, we are left with \~5.3K
44
  dream reports. We then generate a random split of 80\%-20\% for the training (i.e., 4,311 reports) and testing (i.e. 1,078 reports) sets.
45
 
46
  ### Training