Services Usage

From Lsdf
Revision as of 11:12, 10 November 2011 by JillLundykwjyja (talk | contribs)
Jump to navigationJump to search

Following the usage recommendations means less support requests and better services for all the users.

The Hadoop cluster consists of 58 nodes with 464 physical cores, 36 GB of RAM and 2 TB of disk each. All nodes are however shared between the different Hadoop tasks and the OpenNebula virtual machines. Please don't assume that you will have access to all those resources at the same time :-)


Storage usage

  • Storing small files (say < 1MB) wastes resources unnecessarily and dramatically slows down the transfer speed in any protocol. Please try to use data store formats which result in smaller amounts of big files, instead of keeping thousands of small files around. Files of size 100MB or bigger will give the best results!


Hadoop cluster usage

  • The HDFS (Hadoop Filesystem) is accessible both natively (API, 'hadoop' command line tool) and via a Hadoop FUSE mount. This latter one is NOT production quality and should NOT be used for data processing. It is provided only for the sake of "user friendliness"!
  • You should strive to use the Hadoop Map-Reduce framework natively, as this is where the framework excels and where the performance gains will be maximal.
  • If you use the framework to call external executables (via the streaming-jar for instance), then you must make sure that the external executables do NOT start multiple threads.


Cloud/OpenNebula usage

  • Keeping unused virtual machines running wastes resources and complicates maintenance tasks. Please turn off unused instances after you have finished using them!