You can download the raw data directly via: https://doi.org/10.6084/m9.figshare.14452590.
Run the following commands:
- Models trained (one for each of the ten projects)
- Models predictions and weights extraction (one for each of the ten projects, follow the indication in our Ext. Summarizer repository version). You can download them directly via the link: https://doi.org/10.6084/m9.figshare.14414672 Note that there is a set of attention weights for each predicted tokens.
- Condense (using the average) the attention weights to have one set of attention weights for each single method. This step will also keep only methods relevant for which we have human annotations:
- Move the content of extracted_attention_weights folder (just downloaded) in data/raw_data/extreme_summarizer/ folder.
- To convert the raw data into machine attention weights run this from the scripts folder:
WARNING: this is a lengthy process since it will condense all the methods in the dataset. It takes around 9 hours on a laptop with i7 processor. So we suggest downloading the preprocessed data directly.python3 process_raw_human_logs.py prapareextsummarizer
- Models trained (one for each of the ten projects)
- Models predictions and weights extraction (one for each of the ten projects, follow the indication in our Ext. Summarizer repository version). You can download them directly via the link: https://doi.org/10.6084/m9.figshare.14431439 Note that here we already have a single set of attention weights for each method. And we already work only on the methods for which we have human annotations. The only step to perform is to condense the eight attention heads into one (using the averaging).
- Create a single file containing all the machine attention data:
- Move the content of tmp_single_project folder (just downloaded) in data/raw_data/transformer/ folder. Note that the big files with mdl and mdl.checkpoint extensions are not needed.
- To convert the raw data into machine attention weights run this from the scripts folder:
python3 process_raw_human_logs.py praparetransformer
When you download the repository you find the following content in the data folder: data
└─── README.md
│
└─── raw_data
| └─── empty folder
└─── datasets
| └─── methods_showed_to_original_participants
| └─── ... .json|
└─── precomputed_functions_info
└─── functions_sampled_for_the_experiment.json
You can download the preprocessed files ready for analysis directly from: https://doi.org/10.6084/m9.figshare.14462052
They contain:
data
└─── README.md
│
└─── processed_human_attention
| └─── attentions_processed_also_rejected_users.json
│ └─── users_processed_also_rejected_users.json
|
└─── user_info
| └─── users_evaluation.csv
│ └─── users_provenance.csv
│
└─── precomputed_model_prediction
└─── extreme_summarizer
| └─── machine_attention.json
└─── transformer
└─── machine_attention.json
The raw data collected during the experiment (in the same format in which they were stored in MongoDB) can be directly download via the following link:
To convert the raw data into human attention weights:
python3 process_raw_human_logs.py preparehuman --include_rejected_users True
In case you are using the setup for a new experiment you have to configure the config/setting.yaml file to point to the right mongodb database and then run the script to download the data from withing the script folder:
python3 process_raw_human_logs.py download
The output will be stored in the data/raw_data folder.