Services Usage
From Lsdf
Jump to navigationJump to search
Following these usage recommendations means less support requests and better services for all users. See also Services_overview for information about the infrastructure and usage.
Storage usage
- Storing small files wastes resources unnecessarily and dramatically slows down the transfer speed in any protocol. A small file is currently anything below 200 MB. Please try to use data store formats that return big files, instead of keeping thousands of small files around. e.g. Unix users have tar, cpio, zip, Windows can use similar tools.
Hadoop cluster usage
- The Hadoop cluster consists of 58 nodes with 464 physical cores, 36 GB of RAM and 2 TB of disk each. All nodes are however SHARED between the different Hadoop tasks and the OpenNebula virtual machines. Please don't assume that you will have access to all those resources at the same time :-)
- The HDFS (Hadoop Filesystem) is accessible both natively (API, 'hadoop' command line tool) and via a Hadoop FUSE mount. This latter one is NOT production quality and should NOT be used for data processing. It is provided only for the sake of "user friendliness"!
- You should strive to use the Hadoop Map-Reduce framework natively, as this is where the framework excels and where the performance gains will be maximal.
- If you use the framework to call external executables (via the streaming-jar for instance), then you must make sure that the external executables do NOT start multiple threads.
Cloud/OpenNebula usage
- Keeping unused virtual machines running wastes resources and complicates maintenance tasks. Please turn off unused instances after you have finished using them!