Services Usage

From Lsdf
Revision as of 15:10, 12 July 2011 by ariel (talk | contribs)

Following the usage recommendations means less support requests and better services for all the users.


ADD KEY DATA

Storage usage

  • Storing small files (say < 1MB) wastes resources unnecessarily and dramatically slows down the transfer speed in any protocol. Please try to use data store formats which result in smaller amounts of big files, instead of keeping thousands of small files around. Files of size 100MB or bigger will give the best results!


Hadoop cluster usage

  • The HDFS (Hadoop Filesystem) is accessible both natively (API, 'hadoop' command line tool) and via a Hadoop FUSE mount. This latter one is NOT production quality and should NOT be used for data processing. It is provided only for the sake of "user friendliness"!
  • You should strive to use the Hadoop Map-Reduce framework natively, as this is where the framework excels and where the performance gains will be maximal.
  • If you use the framework to call external executables (via the streaming-jar for instance), then you must make sure that the external executables do NOT start multiple threads.


Cloud/OpenNebula usage

  • Keeping unused virtual machines running wastes resources and complicates maintenance tasks. Please turn off unused instances after you have finished using them!