Access to Resources: Difference between revisions

From Lsdf
Jump to navigationJump to search
mNo edit summary
mNo edit summary
 
(18 intermediate revisions by 3 users not shown)
Line 1: Line 1:
In order to get access to the LSDF first read the [[LSDF_Usage|Usage Policy]].
For access to LSDF please contact Jos van Wezel <jos.vanwezel@kit.edu> at SCC. Needed is the following information:


Contact Jos van Wezel <jos.vanwezel@kit.edu> at SCC. In the end we need the following information:
- Name, email, institute


* Name, email, institute
- Other persons (name, email) involved in the intended activities who do already have access to Hadoop? (ie, do you belong to a group who requests access)
- Short description of the intended activities to be performed on the Hadoop infrastructure
* Other persons (name, email) involved in the intended activities who do already have access to DIS Storage or Hadoop? (ie, do you belong to a group who requests access)
* Short description of the intended activities
- Tentative timeline (start, end, any intermediate milestones?)
* Tentative timeline (start, end, any intermediate milestones?)
- Your SSH public key if you have one


For Hadoop usage please estimate:
For Hadoop usage please state:


* Expected nr. of jobs
* Expected nr. of jobs
Line 15: Line 15:
* Amount of RAM per job needed
* Amount of RAM per job needed
* Expected disk space usage in HDFS
* Expected disk space usage in HDFS

For use of Virtual Machines (Currently via Open Nebula) please estimate:

* Expected Nr. of VM's
* Nr. of CPUs per VM
* Disk space per VM
* RAM per VM

Latest revision as of 10:28, 29 July 2016

In order to get access to the LSDF first read the Usage Policy.

Contact Jos van Wezel <jos.vanwezel@kit.edu> at SCC. In the end we need the following information:

  • Name, email, institute
  • Other persons (name, email) involved in the intended activities who do already have access to DIS Storage or Hadoop? (ie, do you belong to a group who requests access)
  • Short description of the intended activities
  • Tentative timeline (start, end, any intermediate milestones?)

For Hadoop usage please state:

  • Expected nr. of jobs
  • Expected running time of average job
  • Nr. of map-tasks (and CPUs) per job foreseen
  • Amount of RAM per job needed
  • Expected disk space usage in HDFS