JASMIN Service Request Questionnaire
Before completing the JASMIN Service Request Questionnaire below, please read the JASMIN Service documentation at http://www.jasmin.ac.uk/services/.
Once you will have completed to the best of your knowledge the form below, please return it to the CEDA Helpdesk ().
NOTE: If you are in the process of submitting a grant proposal and you are thinking of using a JASMIN Group Work Space and/or a dedicated Virtual Machine to share and analyse data, then this request is free at the point of access for NERC grants, but space and processing power will be tensioned against other projects. Please flag this in the proposal’s Outlined Data Management Plan (ODMP) and specify the likely volumes needed. Thank you.
BACKGROUND INFORMATION
JASMIN has medium-term storage service called Group Workspaces (GWS). These are multi-terabyte storage areas created for & provided to projects for some finite period of time. Management of the disk space is the responsibility of a member of the project designated as the GWS Manager, and members of the project all have read/write access to the area, which is mounted where needed. There is no commitment on CEDA to curate the data in the long term. However, one could use a group workspace to gather data and prepare it for inclusion in the BADC archive. Please note that data in GWS are NOT BACKED UP, although the GWS Manager has access to the “Elastic Tape” service which enables him/her to create a secondary copy of the data on near-line tape.
The GWS are mounted on machines on JASMIN where one can login and do analysis or transfer to/from other institutions.
For analysis there are 3 options:
1. Shared analysis VMs: CEDA manage the analysis software stack on systems shared with other users. Your GWS would be mounted on these systems but only readable and writable by your group. You can request extra software to be installed.
2. Project VMs. CEDA can create VMs specific for your project which would have the GWS mounted on it (and the archive, read only, if needed).
3. LOTUS batch processing cluster: A PBS batch scheduling cluster for parallel or multi-job analysis.
Section 1: Your Project
Your name: Rhona Phipps
Your contact details: email :
CEDA username (address, Tel, email):
CEDA username: (to be supplied by Rhona)
Tel: 01183787356, 07711425138
Office: 2U04 Meteorology Building, University of Reading
Project name (include acronym): FIDUCIO
Funder/grant Ref: H2020-EO-2014, Grant Id: 638822
Project start date/duration: 1/3/2015 4 year duration
CEDA Contact (if known): Esther Conway
Section 2: Group Works Space (GWS) request.
Suggested short name or acronym for GWS: FIDUCIO
Designated Group Workspace Manager: Rhona Phipps
Description of the intended purpose of the GWS:
To hold a variety of input satellite data and processed output from the FIDUCIO project
Expected Data :
AVHRR - this data is already at CEMS as part of the ESA CCI SST project
and most has been copied to the general archive area as well I believe
HIRS - EUMETSAT Data Archive seem only to go back to NOAA-17, so we'll
definitely need data from the NOAA archives.
SSM/T2, AMSU-B and MHS - Need to ask Stefan (University of Hamburg) as
to data holdings. Met Office may also have record as they are
independently creating a new FCDR.
MVIRI - Meteosat First Generation. Raw data should be at EUMETSAT and
may be processed there as well, with only the final Level 1 (new format)
at CEMS. This is up to EUMETSAT and Rayference to decide and given the
possible data volume may make more sense to do it at EUMETSAT.
SEVIRI - Meteosat Second Generation. Ancillary data. Raw data is at
EUMETSAT and again where the processing is done will have to be decided
by EUMETSAT/Rayference. Used for validation only.
AIRS/IASI. Ancillary data. IASI data is at EUMETSAT and they are tasked
to do matches of IASI with HIRS/AVHRR which they may decide to do
internally - we're not creating new IASI/AIRS FCDRs. So these ancillary
datasets may not need to be at CEMS (and are very large...).
As to data volumes:
AVHRR : already hosted at CEMS to no extra storage
HIRS : probably ~ 15 Tbytes
Microwave : less than HIRS but have no idea really
MVIRI - no idea, but will be > AVHRR/HIRS holdings
SEVIRI - no idea, but could be huge depending on how many years (used
for validation so may not need than many years...).
AIRS/IASI - no idea but will be huge - but again ancillary data and
EUMETSAT should do this locally.
Requested Volume (Tb) 75Tb
Elastic Tape Quota (TB, by default the same as the requested disk volume):
Date required: 1/3/12015
Section 3: Dedicated Virtual Machine (VM) request.
Sometimes the software provided on the general purpose VMs is not suitable, or groups require exclusive access to a VM, so CEDA can set up dedicated project VMs accessible only to members of a project group. In some circumstances, upon signing a usage agreement, CEDA may be able to grant root access to trusted 3rd party root administrators on these machines where project groups need to be able to install their own software packages (via their nominated root person). However this will depend on the intended use of the system and requires further discussion.
Suggested short name or acronym for VM: FIDUCIO
Proposed VM Administrator:
Description of the intended purpose of the VM (e.g. will this VM be used for multi-threaded processing or mainly serial analysis?):
Large scale processing of satellite data involving multithreaded processing.
VM Basic Configuration options
CPU required:
1 (minimal) ; 2 (light) ; 4 (moderate/multi-user); 6 (Intermediate); 8 (heavy/multi-threaded processing)
8 (heavy/multi-threaded processing)
Note: Beyond 8CPUs, you would be best using the LOTUS batch queue system. This is a shared resource but offers plenty of cores for your needs. The system isn't suitable for interactive processing so you may want to have a VM as well. The whole system shares the same Group Work Spaces and home areas so you can work on both simultaneously.
Memory required (RAM) (GB): 1 (light); 2 (moderate); 4 (multiuser); 8 (multi threaded processing); 16 (Max Science workload)
System disk required (GB): 10 (light); 30 (moderate); 80 (large); 100 (database)
Disk space required (GB): 10; 30; 80; 100; (we recommend 10GB or 30GB for the OS)
OS required: rhel6; rhel5; sles11 (default is rhel6, please describe requirement for other choices here)
Any other information you feel would be of use?
Thank you for providing us with this valuable information. If you have any further questions, please email .
WHAT HAPPENS NEXT?
Your application will be discussed and passed to the relevant consortium manager, who has an allocation of resources on JASMIN to partition between all projects in that science area. You will be contacted by either the CEDA team or the consortium manager to discuss your requirements further and to let you know the next steps.
FOR INTERNAL USE ONLY:
Query associated number:
Consortium (if applicable): Atmos & Polar Science / Oceanography & Shelf Seas / Solid Earth & Mineral Physics / Genomics / Ecology & Hydrology / Earth Observation & Climate Services / Geology
Comments: