Difference between revisions of "Basic XRootD"
(→pdsh) |
(→pdsh) |
||
Line 67: | Line 67: | ||
</pre> | </pre> | ||
− | This of course does not involve any of the | + | This of course does not involve any of the advanced or useful features of pdsh yet - it's put here as a basic example. |
<!-- In Ubuntu, if you need to add a few applications to your desktop, this can be done as follows: | <!-- In Ubuntu, if you need to add a few applications to your desktop, this can be done as follows: |
Revision as of 01:53, 1 September 2016
Disclaimer
The information below is not meant to replace XRootD documentation. It may be helpful for experimentation with a small xrootd cluster and getting familiar with the basics of XRootD configuration.
Installing or Building XRootD
The XRootD team provides rpm packages for installation on Scientific Linux. If your OS is not supported in this manner, building from source is a workable option and is not difficult.
Download the archive and follow the instructions on the official XRootD site. Be sure to read the README file included in the archive. A recent version of CMake shall be required for the build. The build script may be finicky when it's testing the features of the C++ compiler found on the system so you may need to upgrade the compiler or install an alternative one and reconfigure PATH and some other elements of your setup correspondingly. It is to be expected that you'll need to take care of a few dependencies, for example install packages like zlib.
Running XRootD
Controlling your cluster
Basic experimentation with XRootD can start with utilizing a few machines which are idle or aren't heavily loaded by other applications. It is often helpful to have the "sudo" privilege (e.g. for installation in standard directories and configuring auxiliary services) but strictly speaking this is not necessary.
It is convenient to control a few machines from one screen. Typically ssh is used for this purpose, but if security is not a concern (e.g. then the network is strictly local) telnet can be also used as a quick solution.
ssh
You'll need to run the sshd service on every machine you want to connect to. On Linux, this is most frequently openssh-server and it can be trivially installed. To be used productively, private and public keys will need to be generated or imported as necessary. For example, public keys should be added to the file ".ssh/authorized_keys". A matching private key must be loaded to an identity managing service (e.g. ssh-agent in case of Linux) on the machine from which you are going to connect.
Restarting the service:
sudo systemctl restart ssh
Adding a key to the agent:
eval "$(ssh-agent -s)" ssh-add key_file
telnet
On Ubuntu one can install the software necessary to run the telnet service in the following manner:
sudo apt-get install xinetd telnetd
...and start the service as follows:
sudo /etc/init.d/xinetd start
pdsh
This is an advanced parallel shell designed for cluster management. It often uses ssh as the underlying protocol although there are other options as well. Configuration is defined by files residing in /etc/pdsh. For example, the file "machines" needs to contain the list of computers to be targeted by pdsh. Optionally, this is also the place for a file that can be sourced for convenience of setup, cf
# setup pdsh for cluster users export PDSH_RCMD_TYPE='ssh' export WCOLL='/etc/pdsh/machines'
This of course can be done from the command line anyway, cf
export PDSH_RCMD_TYPE=ssh
Using ssh as the underlying protocol for pdsh implies that you have set up private and public keys just like you normally would for ordinary ssh login. Once this is done, you should be able for example to do the following:
pdsh -w targetHost "ls"
This of course does not involve any of the advanced or useful features of pdsh yet - it's put here as a basic example.
Configuration and Log Files
In most cases you will want to run two processes on each machine, "cmsd" and "xrootd". Each of them will need a proper configuration file. See corresponding sections below. Be sure to include paths for log files in the configuration so the log information is readily available for debugging.
Starting a simple instance of xrootd service
There is more than way to start the xrootd service (see documentation). The simplest way is to start the requisite daemon processes from the command line. Starting the xrootd daemon by itself is enough to serve data from a single node (i.e. without creating a storage cluster).
xrootd -c configFile.cfg /path/to/data &
The "/path/to/data" denotes the designated directory, from which it is allowed to serve the data. In xrootd terminology, this path is "exported".
In the above example the file configFile.cfg contains the necessary configuration. Without it present, some simple defaults will be assumed which will allow for testing, but the setup isn't likely to be very useful for real applications. The path which is to be exported can also be defined in the configuration file (which is optimal), in which case it's not necessary to put it in the command line. See the corresponding section below.
The "-b" option will start the process in the background by default, and the "-l" option can be used to specify the path to the log file (otherwise stderr will be assumed). Examples:
cmsd -b -l /path/to/log/cmsd.log -c client.cfg xrootd -b -l /path/to/log/xrootd.log -c client.cfg
The "cmsd" is the clustering daemon which is explained in one of the following sections.
If the "path to data" is not explicitely defined, xrootd will default to /tmp which might work
for initial testing but isn't practical otherwise. Whether xrootd is running as expected can be
tested by using the xrdcp client from any machine from which the server is accessible, e.g.
xrdcp myFile.txt root://serverIP//path/to/data
Clustering
In a clustered environment, you also need to start the cluster manager daemon, e.g.
xrootd -c configFile.cfg /path/to/data & cmsd -c configFile.cfg /path/to/data &
Alternatively,
cmsd -b -l /path/to/log/cmsd.log -c client.cfg xrootd -b -l /path/to/log/xrootd.log -c client.cfg
...in which case the log files are explicitly defined on the command line (as opposed to the default stderr) and the processes are run as daemons.In this example the exported pat is defined in the config file so there is no need to put it in the command line.
The data in the cluster is exposed through the manager node, whose address is to be used in queries. Example:
xrdcp -f xroot://managerIP//my/path/foo local_foo
The file "foo" will be located and if it exists, will be copied to "local_foo" on the machine running the xrdcp client. Caveat: if multiple files exist in the system under the same path, the result (i.e. which one gets fetched) is random.
Configuration File
An example of a working configuration file suitable for a server node (not for the manager node):
all.role server all.export /path/to/data all.manager 192.168.0.191:3121 xrd.port 1094 acc.authdb /path/to/data/auth_file
In the example above the IP address for the manager needs to be set correctly, it's arbitrary in this sample.
authdb
The "authdb" bit is important, things mostly won't work without proper authorization (quite primitive in this case as it relies on a file with permissions). If all users are given access to all data, the content of the file can be as simple as
u * /path/to/data lr
Redirector
The redirector coordinates the function of the cluster. For example, it finds the data based on the path given by the clients such as xrdcp, without the client having to know which nodes contains this bit of data. A crude (but working) example of the redirector configuration (for cmsd):
all.manager managerIP:3121 all.role manager xrd.port 3121 all.export /path/to/data acc.authdb /path/to/data/auth_file
Note the port number. This is not the data port but the service port to used for communication inside the cluster (e.g. for metadata).
The configuration of the server (i.e. for xrootd) might look like this:
all.manager managerIP:3121 all.role manager xrd.port 1094 all.export /path/to/data acc.authdb /path/to/data/auth_file
The simplest way to initialize the redirector service on this node is as follows
xrootd -c server.cfg /path/to/data & cmsd -c redir.cfg /path/to/data &
xrdfs
File Info
Filesystem functionality. Example:
xrdfs managerIP ls -l /my/path xrdfs managerIP ls -u /my/path
In the above the first item performs similarly to "ls -l" in Linux shell, the second prints URLs of the files.
The following command locates the path, i.e. returns the address(es) of the server(s) which physically hold(s) the path - can be multiple machines:
xrdfs managerIP locate /my/path
Adding the "-r" option will force the server to refresh, i.e. to do a fresh query. Otherwise, a cached result will be used if it exists.
The "stat" command provides info similar to "stat":
xrdfs managerIP stat /my/path
The "rm" command does what the name suggest, with the usual caveat that if same path is present on a few machines, the result will be arbitrary - one of the files will be deleted at a time.
Host Info
xrdfs hostIP query config role
Checksum
XRootD hosts can report checksums for files, with a few checksum algorithms available. To enable this on a host a special line needs to be added to the configuration file, for example:
xrootd.chksum md5
As usual, it is only necessary to query the redirector in order to get this info by the xrdfs client:
xrdfs managerIP query checksum /my/path/to/file