Good research practice requires that the data and methods described in a published article are accessible after the publication. There are several reasons to have a secure copy of all relevant data. With the model source, input files and model output, it is possible to prove that the data is genuine and not fabricated. It is also possible that years after the publication, the authors or other scientists would need part of the data or scripts for a follow-up study. Creating a secure archive can prevent data loss caused by hardware breakup or human error. Already when conducting research, it is a good idea to keep archiving in mind and document and organize data files and scripts appropriately. Placing data of a completed study in an archive also frees space from more expensive storage systems as /stornext/field/ on Voima
The exact list of objects to archive depends on each specific study, but some general guidelines can be given to archive data of a modeling study. This list is written with ECHAM-HAMMOZ in mind.
A folder /stornext/field/kuopiodata was created on Voima to store common data for the ISI modelling group. For the atmospheric and ocean modelling group (FMI Helsinki), a folder /stornext/field/hel551data was created.
However, files stored there are counted on file owner's quota. To create a secure archive that is accessible and easy to find for others, create a file named LastName_et_al_PublicationYear.txt on /stornext/field/kuopiodata/publication_archive. In the text file, write the full reference of the publication and the exact location of the data archive for that publication (see below for details).
There is also a tape drive available on stornext. It offers a cheaper alternative to store long-term data compared to /stornext/field. To get a folder on the tape drive, contact Lasse Jalava. The tape drive is not backed up. Therefore, in order to create a secure and doubled data archive, ask for two different locations on the tape drive. When using the tape drive, there are a few things to keep in mind:
Author: Antti-Ilari Partanen
This folder contains the archived data and scripts of the following publication:
Partanen, A.-I., Dunne, E. M., Bergman, T., Laakso, A., Kokkola, H., Ovadnevaite, J., Sogacheva, L., Baisnée, D., Sciare, J., Manders, A., O'Dowd, C., de Leeuw, G., and Korhonen, H.: Global modelling of direct and indirect effects of sea spray aerosol using a source function encapsulating wave state, Atmos. Chem. Phys., 14, 11731-11752, doi:10.5194/acp-14-11731-2014, 2014.
Monthly mean files of the ECHAM-HAMMOZ simulations are stored in files monthlymeans_EXP.tar where EXP is the name of each experiment. Temporal and spatial averages and other post-processed files are stored in files final_EXP.tar.
The name of the simulations in the data match the ones in the publication in the following way:
|Name in the data||Name in the publication|
Other files are explained below.
Anthropogenic aerosol emission files used in the simulations.
Contains tabular data of forcings, in-situ comparison, AOD comparison with Parasol data, and sea spray budget.
Matlab scripts used to process and visualize data, including scripts to make the publication figures.
The ‘prepare' folder contains the scripts used to initialize the working directories of all the runs (ie., generating running and post-processing scripts, and creating links for input files). The package contains also the scripts used to do initial processing of wave height and chlorophyll-a data. Further processing was done using the Matlab scripts in ossa-matlab-scripts.tar.gz .
Final and interpolated chlorophyll-a data used for model input. The dates on file names refer to the start date of each 8-day cycle.
In-situ and PARASOL data used in the study.
Scripts used to calculate monthly mean files from model output and do further post-processing such as calculate means over whole simulation time.
Working directories of each simulation. These contain run and post-processing logs, links to input files, model binary, namelists, and restart files to rerun the experiments from 1.1.2006.
Final and interpolated wave height data used as model input.